code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' import math def _UpperCAmelCase ( _lowerCamelCase : int ) -> bool: assert isinstance(_lowerCamelCase , _lowerCamelCase ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False _lowerCAmelCase : Tuple = range(3 , int(math.sqrt(_lowerCamelCase ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def _UpperCAmelCase ( _lowerCamelCase : Dict , _lowerCamelCase : Tuple=1 , **_lowerCamelCase : Any ) -> str: _lowerCAmelCase : str = factor * value _lowerCAmelCase : List[Any] = value while not is_prime(_lowerCamelCase ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **_lowerCamelCase ) return value
309
'''simple docstring''' from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean UpperCamelCase_ = 0 UpperCamelCase_ = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] UpperCamelCase_ = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right UpperCamelCase_ = tuple[int, int] class a_ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): _lowerCAmelCase : Optional[int] = pos_x _lowerCAmelCase : List[str] = pos_y _lowerCAmelCase : Tuple = (pos_y, pos_x) _lowerCAmelCase : List[Any] = goal_x _lowerCAmelCase : int = goal_y _lowerCAmelCase : Union[str, Any] = g_cost _lowerCAmelCase : List[Any] = parent _lowerCAmelCase : List[Any] = self.calculate_heuristic() _lowerCAmelCase : Optional[int] = self.g_cost + self.h_cost def __UpperCamelCase ( self ): _lowerCAmelCase : List[str] = self.pos_x - self.goal_x _lowerCAmelCase : Optional[int] = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(snake_case_ ) + abs(snake_case_ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , snake_case_ ): return self.f_cost < other.f_cost class a_ : def __init__( self , snake_case_ , snake_case_ ): _lowerCAmelCase : Optional[Any] = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , snake_case_ ) _lowerCAmelCase : Tuple = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9_9_9_9 , snake_case_ ) _lowerCAmelCase : List[str] = [self.start] _lowerCAmelCase : list[Node] = [] _lowerCAmelCase : List[str] = False def __UpperCamelCase ( self ): while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() _lowerCAmelCase : Optional[int] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(snake_case_ ) self.closed_nodes.append(snake_case_ ) _lowerCAmelCase : Optional[int] = self.get_successors(snake_case_ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(snake_case_ ) else: # retrieve the best current path _lowerCAmelCase : Optional[Any] = self.open_nodes.pop(self.open_nodes.index(snake_case_ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(snake_case_ ) else: self.open_nodes.append(snake_case_ ) return [self.start.pos] def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : Union[str, Any] = [] for action in delta: _lowerCAmelCase : Union[str, Any] = parent.pos_x + action[1] _lowerCAmelCase : Dict = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(snake_case_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( snake_case_ , snake_case_ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , snake_case_ , ) ) return successors def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : List[Any] = node _lowerCAmelCase : Optional[Any] = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) _lowerCAmelCase : Optional[int] = current_node.parent path.reverse() return path class a_ : def __init__( self , snake_case_ , snake_case_ ): _lowerCAmelCase : List[str] = AStar(snake_case_ , snake_case_ ) _lowerCAmelCase : int = AStar(snake_case_ , snake_case_ ) _lowerCAmelCase : Optional[int] = False def __UpperCamelCase ( self ): while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() _lowerCAmelCase : Tuple = self.fwd_astar.open_nodes.pop(0 ) _lowerCAmelCase : Optional[Any] = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( snake_case_ , snake_case_ ) self.fwd_astar.closed_nodes.append(snake_case_ ) self.bwd_astar.closed_nodes.append(snake_case_ ) _lowerCAmelCase : List[str] = current_bwd_node _lowerCAmelCase : Dict = current_fwd_node _lowerCAmelCase : Any = { self.fwd_astar: self.fwd_astar.get_successors(snake_case_ ), self.bwd_astar: self.bwd_astar.get_successors(snake_case_ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(snake_case_ ) else: # retrieve the best current path _lowerCAmelCase : List[Any] = astar.open_nodes.pop( astar.open_nodes.index(snake_case_ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(snake_case_ ) else: astar.open_nodes.append(snake_case_ ) return [self.fwd_astar.start.pos] def __UpperCamelCase ( self , snake_case_ , snake_case_ ): _lowerCAmelCase : int = self.fwd_astar.retrace_path(snake_case_ ) _lowerCAmelCase : Optional[Any] = self.bwd_astar.retrace_path(snake_case_ ) bwd_path.pop() bwd_path.reverse() _lowerCAmelCase : Dict = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] UpperCamelCase_ = (0, 0) UpperCamelCase_ = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) UpperCamelCase_ = time.time() UpperCamelCase_ = AStar(init, goal) UpperCamelCase_ = a_star.search() UpperCamelCase_ = time.time() - start_time print(F'AStar execution time = {end_time:f} seconds') UpperCamelCase_ = time.time() UpperCamelCase_ = BidirectionalAStar(init, goal) UpperCamelCase_ = time.time() - bd_start_time print(F'BidirectionalAStar execution time = {bd_end_time:f} seconds')
309
1
import pprint import requests snake_case : List[Any] = '''https://zenquotes.io/api''' def __lowercase ( ): return requests.get(API_ENDPOINT_URL + '/today' ).json() def __lowercase ( ): return requests.get(API_ENDPOINT_URL + '/random' ).json() if __name__ == "__main__": snake_case : List[str] = random_quotes() pprint.pprint(response)
371
import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class snake_case_ (lowerCamelCase_ , unittest.TestCase ): UpperCAmelCase__ : int = MvpTokenizer UpperCAmelCase__ : List[Any] = MvpTokenizerFast UpperCAmelCase__ : Optional[int] = True UpperCAmelCase__ : Tuple = filter_roberta_detectors def lowerCamelCase__( self :Optional[Any] ) -> Union[str, Any]: super().setUp() a__ = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] a__ = dict(zip(__snake_case ,range(len(__snake_case ) ) ) ) a__ = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] a__ = {'unk_token': '<unk>'} a__ = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['vocab_file'] ) a__ = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file ,'w' ,encoding='utf-8' ) as fp: fp.write(json.dumps(__snake_case ) + '\n' ) with open(self.merges_file ,'w' ,encoding='utf-8' ) as fp: fp.write('\n'.join(__snake_case ) ) def lowerCamelCase__( self :Optional[Any] ,**__snake_case :Tuple ) -> int: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**__snake_case ) def lowerCamelCase__( self :Optional[Any] ,**__snake_case :Dict ) -> Dict: kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**__snake_case ) def lowerCamelCase__( self :Any ,__snake_case :Union[str, Any] ) -> List[Any]: return "lower newer", "lower newer" @cached_property def lowerCamelCase__( self :Tuple ) -> Optional[int]: return MvpTokenizer.from_pretrained('RUCAIBox/mvp' ) @cached_property def lowerCamelCase__( self :Optional[int] ) -> str: return MvpTokenizerFast.from_pretrained('RUCAIBox/mvp' ) @require_torch def lowerCamelCase__( self :Optional[int] ) -> str: a__ = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] a__ = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a__ = tokenizer(__snake_case ,max_length=len(__snake_case ) ,padding=__snake_case ,return_tensors='pt' ) self.assertIsInstance(__snake_case ,__snake_case ) self.assertEqual((2, 9) ,batch.input_ids.shape ) self.assertEqual((2, 9) ,batch.attention_mask.shape ) a__ = batch.input_ids.tolist()[0] self.assertListEqual(__snake_case ,__snake_case ) # Test that special tokens are reset @require_torch def lowerCamelCase__( self :Union[str, Any] ) -> List[Any]: a__ = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a__ = tokenizer(__snake_case ,padding=__snake_case ,return_tensors='pt' ) # check if input_ids are returned and no labels self.assertIn('input_ids' ,__snake_case ) self.assertIn('attention_mask' ,__snake_case ) self.assertNotIn('labels' ,__snake_case ) self.assertNotIn('decoder_attention_mask' ,__snake_case ) @require_torch def lowerCamelCase__( self :Any ) -> int: a__ = [ 'Summary of the text.', 'Another summary.', ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a__ = tokenizer(text_target=__snake_case ,max_length=32 ,padding='max_length' ,return_tensors='pt' ) self.assertEqual(32 ,targets['input_ids'].shape[1] ) @require_torch def lowerCamelCase__( self :int ) -> Union[str, Any]: for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a__ = tokenizer( ['I am a small frog' * 10_24, 'I am a small frog'] ,padding=__snake_case ,truncation=__snake_case ,return_tensors='pt' ) self.assertIsInstance(__snake_case ,__snake_case ) self.assertEqual(batch.input_ids.shape ,(2, 10_24) ) @require_torch def lowerCamelCase__( self :List[Any] ) -> Any: a__ = ['A long paragraph for summarization.'] a__ = [ 'Summary of the text.', ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: a__ = tokenizer(__snake_case ,text_target=__snake_case ,return_tensors='pt' ) a__ = inputs['input_ids'] a__ = inputs['labels'] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) def lowerCamelCase__( self :Union[str, Any] ) -> Tuple: pass def lowerCamelCase__( self :Tuple ) -> Optional[Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): a__ = self.rust_tokenizer_class.from_pretrained(__snake_case ,**__snake_case ) a__ = self.tokenizer_class.from_pretrained(__snake_case ,**__snake_case ) a__ = 'A, <mask> AllenNLP sentence.' a__ = tokenizer_r.encode_plus(__snake_case ,add_special_tokens=__snake_case ,return_token_type_ids=__snake_case ) a__ = tokenizer_p.encode_plus(__snake_case ,add_special_tokens=__snake_case ,return_token_type_ids=__snake_case ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['token_type_ids'] ) ,sum(tokens_p['token_type_ids'] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['attention_mask'] ) / len(tokens_r['attention_mask'] ) ,sum(tokens_p['attention_mask'] ) / len(tokens_p['attention_mask'] ) ,) a__ = tokenizer_r.convert_ids_to_tokens(tokens_r['input_ids'] ) a__ = tokenizer_p.convert_ids_to_tokens(tokens_p['input_ids'] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['input_ids'] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r['input_ids'] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( __snake_case ,['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] ) self.assertSequenceEqual( __snake_case ,['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] )
109
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class A_ ( snake_case__ ): _lowercase : str = 'openai/whisper-base' _lowercase : Union[str, Any] = ( 'This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the ' 'transcribed text.' ) _lowercase : Any = 'transcriber' _lowercase : Dict = WhisperProcessor _lowercase : List[Any] = WhisperForConditionalGeneration _lowercase : List[str] = ['audio'] _lowercase : List[str] = ['text'] def UpperCAmelCase ( self : str , UpperCAmelCase : Optional[int] ) -> Dict: return self.pre_processor(UpperCAmelCase , return_tensors='pt' ).input_features def UpperCAmelCase ( self : Tuple , UpperCAmelCase : List[str] ) -> Any: return self.model.generate(inputs=UpperCAmelCase ) def UpperCAmelCase ( self : int , UpperCAmelCase : Union[str, Any] ) -> int: return self.pre_processor.batch_decode(UpperCAmelCase , skip_special_tokens=UpperCAmelCase )[0]
322
def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[list[int]] ) -> int: """simple docstring""" def update_area_of_max_square(SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 __lowerCAmelCase: Union[str, Any] = update_area_of_max_square(SCREAMING_SNAKE_CASE , col + 1 ) __lowerCAmelCase: Tuple = update_area_of_max_square(row + 1 , col + 1 ) __lowerCAmelCase: int = update_area_of_max_square(row + 1 , SCREAMING_SNAKE_CASE ) if mat[row][col]: __lowerCAmelCase: List[str] = 1 + min([right, diagonal, down] ) __lowerCAmelCase: List[str] = max(largest_square_area[0] , SCREAMING_SNAKE_CASE ) return sub_problem_sol else: return 0 __lowerCAmelCase: List[str] = [0] update_area_of_max_square(0 , 0 ) return largest_square_area[0] def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[list[int]] ) -> int: """simple docstring""" def update_area_of_max_square_using_dp_array( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[list[int]] ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] __lowerCAmelCase: List[Any] = update_area_of_max_square_using_dp_array(SCREAMING_SNAKE_CASE , col + 1 , SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Union[str, Any] = update_area_of_max_square_using_dp_array(row + 1 , col + 1 , SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Any = update_area_of_max_square_using_dp_array(row + 1 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if mat[row][col]: __lowerCAmelCase: int = 1 + min([right, diagonal, down] ) __lowerCAmelCase: Union[str, Any] = max(largest_square_area[0] , SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Optional[int] = sub_problem_sol return sub_problem_sol else: return 0 __lowerCAmelCase: int = [0] __lowerCAmelCase: int = [[-1] * cols for _ in range(SCREAMING_SNAKE_CASE )] update_area_of_max_square_using_dp_array(0 , 0 , SCREAMING_SNAKE_CASE ) return largest_square_area[0] def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[list[int]] ) -> int: """simple docstring""" __lowerCAmelCase: int = [[0] * (cols + 1) for _ in range(rows + 1 )] __lowerCAmelCase: Optional[Any] = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): __lowerCAmelCase: Union[str, Any] = dp_array[row][col + 1] __lowerCAmelCase: str = dp_array[row + 1][col + 1] __lowerCAmelCase: Optional[int] = dp_array[row + 1][col] if mat[row][col] == 1: __lowerCAmelCase: Optional[Any] = 1 + min(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __lowerCAmelCase: str = max(dp_array[row][col] , SCREAMING_SNAKE_CASE ) else: __lowerCAmelCase: Dict = 0 return largest_square_area def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[list[int]] ) -> int: """simple docstring""" __lowerCAmelCase: Tuple = [0] * (cols + 1) __lowerCAmelCase: Optional[int] = [0] * (cols + 1) __lowerCAmelCase: str = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): __lowerCAmelCase: int = current_row[col + 1] __lowerCAmelCase: Union[str, Any] = next_row[col + 1] __lowerCAmelCase: Any = next_row[col] if mat[row][col] == 1: __lowerCAmelCase: str = 1 + min(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __lowerCAmelCase: str = max(current_row[col] , SCREAMING_SNAKE_CASE ) else: __lowerCAmelCase: Optional[Any] = 0 __lowerCAmelCase: int = current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
322
1
'''simple docstring''' import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) _A : str ={ '''sample_size''': 32, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': 1_000, '''block_out_channels''': [32, 64], '''attention_head_dim''': 8, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } _A : Union[str, Any] ={ '''sample_size''': 64, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 3, '''num_class_embeds''': 1_000, '''block_out_channels''': [192, 192 * 2, 192 * 3, 192 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } _A : Dict ={ '''sample_size''': 256, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': None, '''block_out_channels''': [256, 256, 256 * 2, 256 * 2, 256 * 4, 256 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''default''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } _A : Dict ={ '''num_train_timesteps''': 40, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } _A : str ={ '''num_train_timesteps''': 201, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } _A : int ={ '''num_train_timesteps''': 151, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Dict: if isinstance(UpperCamelCase , UpperCamelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("""boolean value expected""" ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase=False ) -> Any: lowerCamelCase__ : Any = checkpoint[f'''{old_prefix}.in_layers.0.weight'''] lowerCamelCase__ : int = checkpoint[f'''{old_prefix}.in_layers.0.bias'''] lowerCamelCase__ : Any = checkpoint[f'''{old_prefix}.in_layers.2.weight'''] lowerCamelCase__ : Any = checkpoint[f'''{old_prefix}.in_layers.2.bias'''] lowerCamelCase__ : Optional[Any] = checkpoint[f'''{old_prefix}.emb_layers.1.weight'''] lowerCamelCase__ : Optional[int] = checkpoint[f'''{old_prefix}.emb_layers.1.bias'''] lowerCamelCase__ : Dict = checkpoint[f'''{old_prefix}.out_layers.0.weight'''] lowerCamelCase__ : Tuple = checkpoint[f'''{old_prefix}.out_layers.0.bias'''] lowerCamelCase__ : str = checkpoint[f'''{old_prefix}.out_layers.3.weight'''] lowerCamelCase__ : int = checkpoint[f'''{old_prefix}.out_layers.3.bias'''] if has_skip: lowerCamelCase__ : Tuple = checkpoint[f'''{old_prefix}.skip_connection.weight'''] lowerCamelCase__ : List[Any] = checkpoint[f'''{old_prefix}.skip_connection.bias'''] return new_checkpoint def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase=None ) -> str: lowerCamelCase__ : Dict = checkpoint[f'''{old_prefix}.qkv.weight'''].chunk(3 , dim=0 ) lowerCamelCase__ : str = checkpoint[f'''{old_prefix}.qkv.bias'''].chunk(3 , dim=0 ) lowerCamelCase__ : Any = checkpoint[f'''{old_prefix}.norm.weight'''] lowerCamelCase__ : Optional[int] = checkpoint[f'''{old_prefix}.norm.bias'''] lowerCamelCase__ : List[Any] = weight_q.squeeze(-1 ).squeeze(-1 ) lowerCamelCase__ : List[Any] = bias_q.squeeze(-1 ).squeeze(-1 ) lowerCamelCase__ : Any = weight_k.squeeze(-1 ).squeeze(-1 ) lowerCamelCase__ : Optional[Any] = bias_k.squeeze(-1 ).squeeze(-1 ) lowerCamelCase__ : Dict = weight_v.squeeze(-1 ).squeeze(-1 ) lowerCamelCase__ : Union[str, Any] = bias_v.squeeze(-1 ).squeeze(-1 ) lowerCamelCase__ : Optional[Any] = ( checkpoint[f'''{old_prefix}.proj_out.weight'''].squeeze(-1 ).squeeze(-1 ) ) lowerCamelCase__ : Dict = checkpoint[f'''{old_prefix}.proj_out.bias'''].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> List[Any]: lowerCamelCase__ : str = torch.load(UpperCamelCase , map_location="""cpu""" ) lowerCamelCase__ : Optional[int] = {} lowerCamelCase__ : Optional[int] = checkpoint["""time_embed.0.weight"""] lowerCamelCase__ : List[Any] = checkpoint["""time_embed.0.bias"""] lowerCamelCase__ : int = checkpoint["""time_embed.2.weight"""] lowerCamelCase__ : Optional[Any] = checkpoint["""time_embed.2.bias"""] if unet_config["num_class_embeds"] is not None: lowerCamelCase__ : Optional[Any] = checkpoint["""label_emb.weight"""] lowerCamelCase__ : Tuple = checkpoint["""input_blocks.0.0.weight"""] lowerCamelCase__ : List[str] = checkpoint["""input_blocks.0.0.bias"""] lowerCamelCase__ : Optional[Any] = unet_config["""down_block_types"""] lowerCamelCase__ : Any = unet_config["""layers_per_block"""] lowerCamelCase__ : Any = unet_config["""attention_head_dim"""] lowerCamelCase__ : List[Any] = unet_config["""block_out_channels"""] lowerCamelCase__ : str = 1 lowerCamelCase__ : str = channels_list[0] for i, layer_type in enumerate(UpperCamelCase ): lowerCamelCase__ : List[Any] = channels_list[i] lowerCamelCase__ : List[Any] = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(UpperCamelCase ): lowerCamelCase__ : int = f'''down_blocks.{i}.resnets.{j}''' lowerCamelCase__ : Dict = f'''input_blocks.{current_layer}.0''' lowerCamelCase__ : Tuple = True if j == 0 and downsample_block_has_skip else False lowerCamelCase__ : List[Any] = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , has_skip=UpperCamelCase ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(UpperCamelCase ): lowerCamelCase__ : Tuple = f'''down_blocks.{i}.resnets.{j}''' lowerCamelCase__ : Optional[Any] = f'''input_blocks.{current_layer}.0''' lowerCamelCase__ : str = True if j == 0 and downsample_block_has_skip else False lowerCamelCase__ : Union[str, Any] = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , has_skip=UpperCamelCase ) lowerCamelCase__ : Any = f'''down_blocks.{i}.attentions.{j}''' lowerCamelCase__ : Dict = f'''input_blocks.{current_layer}.1''' lowerCamelCase__ : Tuple = convert_attention( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) current_layer += 1 if i != len(UpperCamelCase ) - 1: lowerCamelCase__ : Tuple = f'''down_blocks.{i}.downsamplers.0''' lowerCamelCase__ : str = f'''input_blocks.{current_layer}.0''' lowerCamelCase__ : Union[str, Any] = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) current_layer += 1 lowerCamelCase__ : Union[str, Any] = current_channels # hardcoded the mid-block for now lowerCamelCase__ : Any = """mid_block.resnets.0""" lowerCamelCase__ : Optional[Any] = """middle_block.0""" lowerCamelCase__ : int = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : List[Any] = """mid_block.attentions.0""" lowerCamelCase__ : Dict = """middle_block.1""" lowerCamelCase__ : int = convert_attention(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Any = """mid_block.resnets.1""" lowerCamelCase__ : Tuple = """middle_block.2""" lowerCamelCase__ : int = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = 0 lowerCamelCase__ : Any = unet_config["""up_block_types"""] for i, layer_type in enumerate(UpperCamelCase ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): lowerCamelCase__ : int = f'''up_blocks.{i}.resnets.{j}''' lowerCamelCase__ : Optional[Any] = f'''output_blocks.{current_layer}.0''' lowerCamelCase__ : Any = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , has_skip=UpperCamelCase ) current_layer += 1 if i != len(UpperCamelCase ) - 1: lowerCamelCase__ : Dict = f'''up_blocks.{i}.upsamplers.0''' lowerCamelCase__ : List[str] = f'''output_blocks.{current_layer-1}.1''' lowerCamelCase__ : Optional[Any] = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): lowerCamelCase__ : str = f'''up_blocks.{i}.resnets.{j}''' lowerCamelCase__ : List[Any] = f'''output_blocks.{current_layer}.0''' lowerCamelCase__ : Optional[Any] = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , has_skip=UpperCamelCase ) lowerCamelCase__ : Optional[Any] = f'''up_blocks.{i}.attentions.{j}''' lowerCamelCase__ : Any = f'''output_blocks.{current_layer}.1''' lowerCamelCase__ : Optional[int] = convert_attention( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) current_layer += 1 if i != len(UpperCamelCase ) - 1: lowerCamelCase__ : Tuple = f'''up_blocks.{i}.upsamplers.0''' lowerCamelCase__ : Tuple = f'''output_blocks.{current_layer-1}.2''' lowerCamelCase__ : List[str] = convert_resnet(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Dict = checkpoint["""out.0.weight"""] lowerCamelCase__ : Dict = checkpoint["""out.0.bias"""] lowerCamelCase__ : Dict = checkpoint["""out.2.weight"""] lowerCamelCase__ : Tuple = checkpoint["""out.2.bias"""] return new_checkpoint if __name__ == "__main__": _A : Tuple =argparse.ArgumentParser() parser.add_argument('''--unet_path''', default=None, type=str, required=True, help='''Path to the unet.pt to convert.''') parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output the converted UNet model.''' ) parser.add_argument('''--class_cond''', default=True, type=str, help='''Whether the model is class-conditional.''') _A : Tuple =parser.parse_args() _A : Optional[int] =strabool(args.class_cond) _A : List[str] =os.path.basename(args.unet_path) print(F'Checkpoint: {ckpt_name}') # Get U-Net config if "imagenet64" in ckpt_name: _A : int =IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): _A : Tuple =LSUN_256_UNET_CONFIG elif "test" in ckpt_name: _A : Any =TEST_UNET_CONFIG else: raise ValueError(F'Checkpoint type {ckpt_name} is not currently supported.') if not args.class_cond: _A : str =None _A : Optional[int] =con_pt_to_diffuser(args.unet_path, unet_config) _A : Optional[int] =UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: _A : Tuple =CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: _A : int =CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): _A : Union[str, Any] =CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(F'Checkpoint type {ckpt_name} is not currently supported.') _A : str =CMStochasticIterativeScheduler(**scheduler_config) _A : Optional[Any] =ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
353
'''simple docstring''' import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient _A : str =WebClient(token=os.environ['''CI_SLACK_BOT_TOKEN''']) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Tuple: lowerCamelCase__ : int = test_results.split(""" """ ) lowerCamelCase__ : Optional[int] = 0 lowerCamelCase__ : Any = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. lowerCamelCase__ : Union[str, Any] = expressions[-2] if """=""" in expressions[-1] else expressions[-1] for i, expression in enumerate(UpperCamelCase ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Tuple: lowerCamelCase__ : Optional[Any] = {} lowerCamelCase__ : int = None lowerCamelCase__ : Optional[int] = False for line in failures_short_lines.split("""\n""" ): if re.search(r"""_ \[doctest\]""" , UpperCamelCase ): lowerCamelCase__ : Dict = True lowerCamelCase__ : Optional[int] = line.split(""" """ )[2] elif in_error and not line.split(""" """ )[0].isdigit(): lowerCamelCase__ : List[str] = line lowerCamelCase__ : int = False return failures class _lowercase : def __init__( self: Tuple , UpperCamelCase__: str , UpperCamelCase__: Dict ): lowerCamelCase__ : Union[str, Any] = title lowerCamelCase__ : Tuple = doc_test_results["""time_spent"""].split(""",""" )[0] lowerCamelCase__ : Union[str, Any] = doc_test_results["""success"""] lowerCamelCase__ : List[Any] = doc_test_results["""failures"""] lowerCamelCase__ : List[str] = self.n_success + self.n_failures # Failures and success of the modeling tests lowerCamelCase__ : str = doc_test_results @property def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Union[str, Any] = [self._time_spent] lowerCamelCase__ : Tuple = 0 for time in time_spent: lowerCamelCase__ : Tuple = time.split(""":""" ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(UpperCamelCase__ ) == 1: lowerCamelCase__ : Tuple = [0, 0, time_parts[0]] lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : List[Any] = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 3_600 + minutes * 60 + seconds lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : List[Any] = total_secs // 3_600, (total_secs % 3_600) // 60, total_secs % 60 return F'''{int(UpperCamelCase__ )}h{int(UpperCamelCase__ )}m{int(UpperCamelCase__ )}s''' @property def lowerCamelCase_ ( self: Dict ): return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def lowerCamelCase_ ( self: Any ): return { "type": "section", "text": { "type": "plain_text", "text": F'''🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.''', "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } @property def lowerCamelCase_ ( self: Any ): return { "type": "section", "text": { "type": "plain_text", "text": ( F'''There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in''' F''' {self.time}.''' ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } @property def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Any = 40 lowerCamelCase__ : List[str] = {k: v["""failed"""] for k, v in doc_test_results.items() if isinstance(UpperCamelCase__ , UpperCamelCase__ )} lowerCamelCase__ : List[Any] = """""" for category, failures in category_failures.items(): if len(UpperCamelCase__ ) == 0: continue if report != "": report += "\n\n" report += F'''*{category} failures*:'''.ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(UpperCamelCase__ ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": F'''The following examples had failures:\n\n\n{report}\n''', }, } @property def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Union[str, Any] = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(UpperCamelCase__ ) @staticmethod def lowerCamelCase_ ( ): lowerCamelCase__ : List[Any] = [ { """type""": """section""", """text""": { """type""": """plain_text""", """text""": """There was an issue running the tests.""", }, """accessory""": { """type""": """button""", """text""": {"""type""": """plain_text""", """text""": """Check Action results""", """emoji""": True}, """url""": F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } ] print("""Sending the following payload""" ) print(json.dumps({"""blocks""": json.loads(UpperCamelCase__ )} ) ) client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , text="""There was an issue running the tests.""" , blocks=UpperCamelCase__ , ) def lowerCamelCase_ ( self: Any ): print("""Sending the following payload""" ) print(json.dumps({"""blocks""": json.loads(self.payload )} ) ) lowerCamelCase__ : Any = F'''{self.n_failures} failures out of {self.n_tests} tests,''' if self.n_failures else """All tests passed.""" lowerCamelCase__ : List[str] = client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , blocks=self.payload , text=UpperCamelCase__ , ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: str , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = """""" for key, value in failures.items(): lowerCamelCase__ : int = value[:200] + """ [Truncated]""" if len(UpperCamelCase__ ) > 250 else value failures_text += F'''*{key}*\n_{value}_\n\n''' lowerCamelCase__ : Tuple = job_name lowerCamelCase__ : Union[str, Any] = {"""type""": """section""", """text""": {"""type""": """mrkdwn""", """text""": text}} if job_link is not None: lowerCamelCase__ : Union[str, Any] = { """type""": """button""", """text""": {"""type""": """plain_text""", """text""": """GitHub Action job""", """emoji""": True}, """url""": job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def lowerCamelCase_ ( self: Tuple ): if self.thread_ts is None: raise ValueError("""Can only post reply if a post has been made.""" ) lowerCamelCase__ : int = self.doc_test_results.pop("""job_link""" ) self.doc_test_results.pop("""failures""" ) self.doc_test_results.pop("""success""" ) self.doc_test_results.pop("""time_spent""" ) lowerCamelCase__ : List[Any] = sorted(self.doc_test_results.items() , key=lambda UpperCamelCase__ : t[0] ) for job, job_result in sorted_dict: if len(job_result["""failures"""] ): lowerCamelCase__ : Union[str, Any] = F'''*Num failures* :{len(job_result['failed'] )} \n''' lowerCamelCase__ : Union[str, Any] = job_result["""failures"""] lowerCamelCase__ : int = self.get_reply_blocks(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , text=UpperCamelCase__ ) print("""Sending the following reply""" ) print(json.dumps({"""blocks""": blocks} ) ) client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , text=F'''Results for {job}''' , blocks=UpperCamelCase__ , thread_ts=self.thread_ts["""ts"""] , ) time.sleep(1 ) def SCREAMING_SNAKE_CASE_ () -> Tuple: lowerCamelCase__ : Any = os.environ["""GITHUB_RUN_ID"""] lowerCamelCase__ : List[Any] = f'''https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100''' lowerCamelCase__ : Optional[int] = requests.get(UpperCamelCase ).json() lowerCamelCase__ : List[Any] = {} try: jobs.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) lowerCamelCase__ : Any = math.ceil((result["""total_count"""] - 100) / 100 ) for i in range(UpperCamelCase ): lowerCamelCase__ : List[Any] = requests.get(url + f'''&page={i + 2}''' ).json() jobs.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) return jobs except Exception as e: print("""Unknown error, could not fetch links.""" , UpperCamelCase ) return {} def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ : int = {} if os.path.exists(UpperCamelCase ): lowerCamelCase__ : List[str] = os.listdir(UpperCamelCase ) for file in files: try: with open(os.path.join(UpperCamelCase , UpperCamelCase ) , encoding="""utf-8""" ) as f: lowerCamelCase__ : List[Any] = f.read() except UnicodeDecodeError as e: raise ValueError(f'''Could not open {os.path.join(UpperCamelCase , UpperCamelCase )}.''' ) from e return _artifact def SCREAMING_SNAKE_CASE_ () -> Optional[Any]: class _lowercase : def __init__( self: Tuple , UpperCamelCase__: str ): lowerCamelCase__ : Any = name lowerCamelCase__ : Union[str, Any] = [] def __str__( self: int ): return self.name def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: str ): self.paths.append({"""name""": self.name, """path""": path} ) lowerCamelCase__ : Dict[str, Artifact] = {} lowerCamelCase__ : List[str] = filter(os.path.isdir , os.listdir() ) for directory in directories: lowerCamelCase__ : Union[str, Any] = directory if artifact_name not in _available_artifacts: lowerCamelCase__ : Optional[int] = Artifact(UpperCamelCase ) _available_artifacts[artifact_name].add_path(UpperCamelCase ) return _available_artifacts if __name__ == "__main__": _A : Any =get_job_links() _A : str =retrieve_available_artifacts() _A : int =collections.OrderedDict( [ ('''*.py''', '''API Examples'''), ('''*.md''', '''MD Examples'''), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' _A : Union[str, Any] ={ v: { '''failed''': [], '''failures''': {}, } for v in docs.values() } # Link to the GitHub Action job _A : Union[str, Any] =github_actions_job_links.get('''run_doctests''') _A : Any =available_artifacts['''doc_tests_gpu_test_reports'''].paths[0] _A : Dict =retrieve_artifact(artifact_path['''name''']) if "stats" in artifact: _A , _A , _A : Optional[int] =handle_test_results(artifact['''stats''']) _A : Union[str, Any] =failed _A : int =success _A : Optional[int] =time_spent[1:-1] + ''', ''' _A : Any =extract_first_line_failure(artifact['''failures_short''']) for line in artifact["summary_short"].split('''\n'''): if re.search('''FAILED''', line): _A : List[Any] =line.replace('''FAILED ''', '''''') _A : Any =line.split()[0].replace('''\n''', '''''') if "::" in line: _A , _A : Any =line.split('''::''') else: _A , _A : Tuple =line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): _A : str =docs[file_regex] doc_test_results[category]["failed"].append(test) _A : str =all_failures[test] if test in all_failures else '''N/A''' _A : Tuple =failure break _A : Union[str, Any] =Message('''🤗 Results of the doc tests.''', doc_test_results) message.post() message.post_reply()
129
0
import os import string import sys lowercase__ : List[str] = 1 << 8 lowercase__ : Tuple = { "tab": ord("\t"), "newline": ord("\r"), "esc": 27, "up": 65 + ARROW_KEY_FLAG, "down": 66 + ARROW_KEY_FLAG, "right": 67 + ARROW_KEY_FLAG, "left": 68 + ARROW_KEY_FLAG, "mod_int": 91, "undefined": sys.maxsize, "interrupt": 3, "insert": 50, "delete": 51, "pg_up": 53, "pg_down": 54, } lowercase__ : Any = KEYMAP["up"] lowercase__ : Tuple = KEYMAP["left"] if sys.platform == "win32": lowercase__ : Any = [] lowercase__ : List[Any] = { b"\xe0H": KEYMAP["up"] - ARROW_KEY_FLAG, b"\x00H": KEYMAP["up"] - ARROW_KEY_FLAG, b"\xe0P": KEYMAP["down"] - ARROW_KEY_FLAG, b"\x00P": KEYMAP["down"] - ARROW_KEY_FLAG, b"\xe0M": KEYMAP["right"] - ARROW_KEY_FLAG, b"\x00M": KEYMAP["right"] - ARROW_KEY_FLAG, b"\xe0K": KEYMAP["left"] - ARROW_KEY_FLAG, b"\x00K": KEYMAP["left"] - ARROW_KEY_FLAG, } for i in range(10): lowercase__ : str = ord(str(i)) def lowerCamelCase__ ( ): '''simple docstring''' if os.name == "nt": import msvcrt snake_case_ = "mbcs" # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(UpperCamelCase_ ) == 0: # Read the keystroke snake_case_ = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): snake_case_ = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: snake_case_ = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP["mod_int"] ) ) WIN_CH_BUFFER.append(UpperCamelCase_ ) if ord(UpperCamelCase_ ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(126 ) ) snake_case_ = chr(KEYMAP["esc"] ) except KeyError: snake_case_ = cha[1] else: snake_case_ = ch.decode(UpperCamelCase_ ) else: snake_case_ = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty snake_case_ = sys.stdin.fileno() snake_case_ = termios.tcgetattr(UpperCamelCase_ ) try: tty.setraw(UpperCamelCase_ ) snake_case_ = sys.stdin.read(1 ) finally: termios.tcsetattr(UpperCamelCase_ , termios.TCSADRAIN , UpperCamelCase_ ) return ch def lowerCamelCase__ ( ): '''simple docstring''' snake_case_ = get_raw_chars() if ord(UpperCamelCase_ ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(UpperCamelCase_ ) == KEYMAP["esc"]: snake_case_ = get_raw_chars() if ord(UpperCamelCase_ ) == KEYMAP["mod_int"]: snake_case_ = get_raw_chars() if ord(UpperCamelCase_ ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(UpperCamelCase_ ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(UpperCamelCase_ ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
187
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __snake_case = logging.get_logger(__name__) __snake_case = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class lowercase__ ( _UpperCAmelCase ): def __init__( self : str , UpperCAmelCase_ : int=None , UpperCAmelCase_ : List[str]=None , *UpperCAmelCase_ : Union[str, Any] , **UpperCAmelCase_ : Optional[Any] ): super().__init__(*UpperCAmelCase_ , **UpperCAmelCase_ ) if config is None: assert isinstance(self.model , UpperCAmelCase_ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F' {self.model.__class__}' ) SCREAMING_SNAKE_CASE__ = self.model.config else: SCREAMING_SNAKE_CASE__ = config SCREAMING_SNAKE_CASE__ = data_args SCREAMING_SNAKE_CASE__ = self.config.tgt_vocab_size if isinstance(self.config , UpperCAmelCase_ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F'The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for' ' padding..' ) if self.args.label_smoothing == 0: SCREAMING_SNAKE_CASE__ = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss SCREAMING_SNAKE_CASE__ = label_smoothed_nll_loss def A_ ( self : Tuple , UpperCAmelCase_ : int ): if self.optimizer is None: SCREAMING_SNAKE_CASE__ = ['bias', 'LayerNorm.weight'] SCREAMING_SNAKE_CASE__ = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] SCREAMING_SNAKE_CASE__ = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: SCREAMING_SNAKE_CASE__ = Adafactor SCREAMING_SNAKE_CASE__ = {'scale_parameter': False, 'relative_step': False} else: SCREAMING_SNAKE_CASE__ = AdamW SCREAMING_SNAKE_CASE__ = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } SCREAMING_SNAKE_CASE__ = self.args.learning_rate if self.sharded_ddp: SCREAMING_SNAKE_CASE__ = OSS( params=UpperCAmelCase_ , optim=UpperCAmelCase_ , **UpperCAmelCase_ , ) else: SCREAMING_SNAKE_CASE__ = optimizer_cls(UpperCAmelCase_ , **UpperCAmelCase_ ) if self.lr_scheduler is None: SCREAMING_SNAKE_CASE__ = self._get_lr_scheduler(UpperCAmelCase_ ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def A_ ( self : str , UpperCAmelCase_ : List[Any] ): SCREAMING_SNAKE_CASE__ = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": SCREAMING_SNAKE_CASE__ = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": SCREAMING_SNAKE_CASE__ = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: SCREAMING_SNAKE_CASE__ = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=UpperCAmelCase_ ) return scheduler def A_ ( self : List[str] ): if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def A_ ( self : Any , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict ): if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token SCREAMING_SNAKE_CASE__ = model(**UpperCAmelCase_ , use_cache=UpperCAmelCase_ )[0] SCREAMING_SNAKE_CASE__ = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = model(**UpperCAmelCase_ , labels=UpperCAmelCase_ , use_cache=UpperCAmelCase_ )[:2] else: # compute label smoothed loss SCREAMING_SNAKE_CASE__ = model(**UpperCAmelCase_ , use_cache=UpperCAmelCase_ )[0] SCREAMING_SNAKE_CASE__ = torch.nn.functional.log_softmax(UpperCAmelCase_ , dim=-1 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.loss_fn(UpperCAmelCase_ , UpperCAmelCase_ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def A_ ( self : Union[str, Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : List[Any] ): SCREAMING_SNAKE_CASE__ = inputs.pop('labels' ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self._compute_loss(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return loss def A_ ( self : List[str] , UpperCAmelCase_ : nn.Module , UpperCAmelCase_ : Dict[str, Union[torch.Tensor, Any]] , UpperCAmelCase_ : bool , UpperCAmelCase_ : Optional[List[str]] = None , ): SCREAMING_SNAKE_CASE__ = self._prepare_inputs(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: SCREAMING_SNAKE_CASE__ = self.model.generate( inputs['input_ids'] , attention_mask=inputs['attention_mask'] , **UpperCAmelCase_ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: SCREAMING_SNAKE_CASE__ = self._pad_tensors_to_max_len(UpperCAmelCase_ , gen_kwargs['max_length'] ) SCREAMING_SNAKE_CASE__ = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self._compute_loss(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) SCREAMING_SNAKE_CASE__ = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: SCREAMING_SNAKE_CASE__ = self._pad_tensors_to_max_len(UpperCAmelCase_ , gen_kwargs['max_length'] ) return (loss, logits, labels) def A_ ( self : Union[str, Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[str] ): # If PAD token is not defined at least EOS token has to be defined SCREAMING_SNAKE_CASE__ = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F' padded to `max_length`={max_length}' ) SCREAMING_SNAKE_CASE__ = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) SCREAMING_SNAKE_CASE__ = tensor return padded_tensor
176
0
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): lowerCAmelCase_ = "pt" elif is_tf_available(): lowerCAmelCase_ = "tf" else: lowerCAmelCase_ = "jax" class lowerCamelCase ( __lowerCAmelCase , unittest.TestCase ): snake_case_ = PerceiverTokenizer snake_case_ = False def _lowerCamelCase ( self ) -> Any: super().setUp() snake_case = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _lowerCamelCase ( self ) -> Dict: return PerceiverTokenizer.from_pretrained('deepmind/language-perceiver' ) def _lowerCamelCase ( self, **lowercase_ ) -> PerceiverTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname, **lowercase_ ) def _lowerCamelCase ( self, lowercase_, lowercase_=False, lowercase_=20, lowercase_=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. snake_case = [] for i in range(len(lowercase_ ) ): try: snake_case = tokenizer.decode([i], clean_up_tokenization_spaces=lowercase_ ) except UnicodeDecodeError: pass toks.append((i, tok) ) snake_case = list(filter(lambda lowercase_ : re.match(r'^[ a-zA-Z]+$', t[1] ), lowercase_ ) ) snake_case = list(filter(lambda lowercase_ : [t[0]] == tokenizer.encode(t[1], add_special_tokens=lowercase_ ), lowercase_ ) ) if max_length is not None and len(lowercase_ ) > max_length: snake_case = toks[:max_length] if min_length is not None and len(lowercase_ ) < min_length and len(lowercase_ ) > 0: while len(lowercase_ ) < min_length: snake_case = toks + toks # toks_str = [t[1] for t in toks] snake_case = [t[0] for t in toks] # Ensure consistency snake_case = tokenizer.decode(lowercase_, clean_up_tokenization_spaces=lowercase_ ) if " " not in output_txt and len(lowercase_ ) > 1: snake_case = ( tokenizer.decode([toks_ids[0]], clean_up_tokenization_spaces=lowercase_ ) + ' ' + tokenizer.decode(toks_ids[1:], clean_up_tokenization_spaces=lowercase_ ) ) if with_prefix_space: snake_case = ' ' + output_txt snake_case = tokenizer.encode(lowercase_, add_special_tokens=lowercase_ ) return output_txt, output_ids def _lowerCamelCase ( self ) -> List[str]: snake_case = self.perceiver_tokenizer snake_case = 'Unicode €.' snake_case = tokenizer(lowercase_ ) snake_case = [4, 91, 116, 111, 105, 117, 106, 107, 38, 232, 136, 178, 52, 5] self.assertEqual(encoded['input_ids'], lowercase_ ) # decoding snake_case = tokenizer.decode(lowercase_ ) self.assertEqual(lowercase_, '[CLS]Unicode €.[SEP]' ) snake_case = tokenizer('e è é ê ë' ) snake_case = [4, 107, 38, 201, 174, 38, 201, 175, 38, 201, 176, 38, 201, 177, 5] self.assertEqual(encoded['input_ids'], lowercase_ ) # decoding snake_case = tokenizer.decode(lowercase_ ) self.assertEqual(lowercase_, '[CLS]e è é ê ë[SEP]' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('e è é ê ë' ) ), '[CLS]e è é ê ë[SEP]' ) def _lowerCamelCase ( self ) -> List[Any]: snake_case = self.perceiver_tokenizer snake_case = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] # fmt: off snake_case = [4, 71, 38, 114, 117, 116, 109, 38, 118, 103, 120, 103, 109, 120, 103, 118, 110, 38, 108, 117, 120, 38, 121, 123, 115, 115, 103, 120, 111, 128, 103, 122, 111, 117, 116, 52, 5, 0] # fmt: on snake_case = tokenizer(lowercase_, padding=lowercase_, return_tensors=lowercase_ ) self.assertIsInstance(lowercase_, lowercase_ ) if FRAMEWORK != "jax": snake_case = list(batch.input_ids.numpy()[0] ) else: snake_case = list(batch.input_ids.tolist()[0] ) self.assertListEqual(lowercase_, lowercase_ ) self.assertEqual((2, 38), batch.input_ids.shape ) self.assertEqual((2, 38), batch.attention_mask.shape ) def _lowerCamelCase ( self ) -> str: snake_case = self.perceiver_tokenizer snake_case = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] snake_case = tokenizer(lowercase_, padding=lowercase_, return_tensors=lowercase_ ) # check if input_ids are returned and no decoder_input_ids self.assertIn('input_ids', lowercase_ ) self.assertIn('attention_mask', lowercase_ ) self.assertNotIn('decoder_input_ids', lowercase_ ) self.assertNotIn('decoder_attention_mask', lowercase_ ) def _lowerCamelCase ( self ) -> Optional[Any]: snake_case = self.perceiver_tokenizer snake_case = [ 'Summary of the text.', 'Another summary.', ] snake_case = tokenizer( text_target=lowercase_, max_length=32, padding='max_length', truncation=lowercase_, return_tensors=lowercase_ ) self.assertEqual(32, targets['input_ids'].shape[1] ) def _lowerCamelCase ( self ) -> str: # safety check on max_len default value so we are sure the test works snake_case = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length, 42 ) # Now let's start the test snake_case = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc snake_case = tempfile.mkdtemp() snake_case = ' He is very happy, UNwant\u00E9d,running' snake_case = tokenizer.encode(lowercase_, add_special_tokens=lowercase_ ) tokenizer.save_pretrained(lowercase_ ) snake_case = tokenizer.__class__.from_pretrained(lowercase_ ) snake_case = after_tokenizer.encode(lowercase_, add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_, lowercase_ ) shutil.rmtree(lowercase_ ) snake_case = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc snake_case = tempfile.mkdtemp() snake_case = ' He is very happy, UNwant\u00E9d,running' tokenizer.add_tokens(['bim', 'bambam'] ) snake_case = tokenizer.additional_special_tokens additional_special_tokens.append('new_additional_special_token' ) tokenizer.add_special_tokens({'additional_special_tokens': additional_special_tokens} ) snake_case = tokenizer.encode(lowercase_, add_special_tokens=lowercase_ ) tokenizer.save_pretrained(lowercase_ ) snake_case = tokenizer.__class__.from_pretrained(lowercase_ ) snake_case = after_tokenizer.encode(lowercase_, add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_, lowercase_ ) self.assertIn('new_additional_special_token', after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length, 42 ) snake_case = tokenizer.__class__.from_pretrained(lowercase_, model_max_length=43 ) self.assertEqual(tokenizer.model_max_length, 43 ) shutil.rmtree(lowercase_ ) def _lowerCamelCase ( self ) -> List[Any]: snake_case = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_, 'special_tokens_map.json' ), encoding='utf-8' ) as json_file: snake_case = json.load(lowercase_ ) with open(os.path.join(lowercase_, 'tokenizer_config.json' ), encoding='utf-8' ) as json_file: snake_case = json.load(lowercase_ ) snake_case = [F'''<extra_id_{i}>''' for i in range(125 )] snake_case = added_tokens_extra_ids + [ 'an_additional_special_token' ] snake_case = added_tokens_extra_ids + [ 'an_additional_special_token' ] with open(os.path.join(lowercase_, 'special_tokens_map.json' ), 'w', encoding='utf-8' ) as outfile: json.dump(lowercase_, lowercase_ ) with open(os.path.join(lowercase_, 'tokenizer_config.json' ), 'w', encoding='utf-8' ) as outfile: json.dump(lowercase_, lowercase_ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files snake_case = tokenizer_class.from_pretrained( lowercase_, ) self.assertIn( 'an_additional_special_token', tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['an_additional_special_token'], tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['an_additional_special_token'] ) ), ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained snake_case = added_tokens_extra_ids + [AddedToken('a_new_additional_special_token', lstrip=lowercase_ )] snake_case = tokenizer_class.from_pretrained( lowercase_, additional_special_tokens=lowercase_, ) self.assertIn('a_new_additional_special_token', tokenizer.additional_special_tokens ) self.assertEqual( ['a_new_additional_special_token'], tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['a_new_additional_special_token'] ) ), ) def _lowerCamelCase ( self ) -> Any: snake_case = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([178] ), '�' ) def _lowerCamelCase ( self ) -> str: pass def _lowerCamelCase ( self ) -> Optional[Any]: pass def _lowerCamelCase ( self ) -> List[Any]: pass def _lowerCamelCase ( self ) -> Optional[Any]: pass def _lowerCamelCase ( self ) -> Optional[Any]: # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens snake_case = self.get_tokenizers(fast=lowercase_, do_lower_case=lowercase_ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case = ['[CLS]', 't', 'h', 'i', 's', ' ', 'i', 's', ' ', 'a', ' ', 't', 'e', 's', 't', '[SEP]'] snake_case = tokenizer.convert_tokens_to_string(lowercase_ ) self.assertIsInstance(lowercase_, lowercase_ )
357
'''simple docstring''' def __magic_name__ ( A ) -> float: return 1_0 - x * x def __magic_name__ ( A , A ) -> float: # Bolzano theory in order to find if there is a root between a and b if equation(A ) * equation(A ) >= 0: raise ValueError('Wrong space!' ) snake_case = a while (b - a) >= 0.01: # Find middle point snake_case = (a + b) / 2 # Check if middle point is root if equation(A ) == 0.0: break # Decide the side to repeat the steps if equation(A ) * equation(A ) < 0: snake_case = c else: snake_case = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
332
0
"""simple docstring""" import unittest from knapsack import greedy_knapsack as kp class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self ): __a = [10, 20, 30, 40, 50, 60] __a = [2, 4, 6, 8, 10, 12] __a = 100 self.assertEqual(kp.calc_profit(_a , _a , _a ) , 210 ) def __UpperCAmelCase ( self ): self.assertRaisesRegex(_a , '''max_weight must greater than zero.''' ) def __UpperCAmelCase ( self ): self.assertRaisesRegex(_a , '''Weight can not be negative.''' ) def __UpperCAmelCase ( self ): self.assertRaisesRegex(_a , '''Profit can not be negative.''' ) def __UpperCAmelCase ( self ): self.assertRaisesRegex(_a , '''max_weight must greater than zero.''' ) def __UpperCAmelCase ( self ): self.assertRaisesRegex( _a , '''The length of profit and weight must be same.''' ) if __name__ == "__main__": unittest.main()
45
from __future__ import annotations import math def A_ ( snake_case : int ) -> bool: '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True lowercase__ : int = [num for num in range(3, 1_0_0_0_0_1, 2) if not is_prime(num)] def A_ ( snake_case : int ) -> list[int]: '''simple docstring''' if not isinstance(snake_case , snake_case ): raise ValueError('''n must be an integer''' ) if n <= 0: raise ValueError('''n must be >= 0''' ) __UpperCamelCase = [] for num in range(len(snake_case ) ): __UpperCamelCase = 0 while 2 * i * i <= odd_composites[num]: __UpperCamelCase = odd_composites[num] - 2 * i * i if is_prime(snake_case ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(snake_case ) == n: return list_nums return [] def A_ ( ) -> int: '''simple docstring''' return compute_nums(1 )[0] if __name__ == "__main__": print(F"{solution() = }")
328
0
import argparse import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_dummies.py _SCREAMING_SNAKE_CASE = 'src/diffusers' # Matches is_xxx_available() _SCREAMING_SNAKE_CASE = re.compile(R'is\_([a-z_]*)_available\(\)') # Matches from xxx import bla _SCREAMING_SNAKE_CASE = re.compile(R'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') _SCREAMING_SNAKE_CASE = '\n{0} = None\n' _SCREAMING_SNAKE_CASE = '\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n\n @classmethod\n def from_config(cls, *args, **kwargs):\n requires_backends(cls, {1})\n\n @classmethod\n def from_pretrained(cls, *args, **kwargs):\n requires_backends(cls, {1})\n' _SCREAMING_SNAKE_CASE = '\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n' def snake_case ( snake_case__ :Optional[Any]) -> str: _A = _re_backend.findall(snake_case__) if len(snake_case__) == 0: return None return "_and_".join(snake_case__) def snake_case ( ) -> Tuple: with open(os.path.join(snake_case__ , """__init__.py""") , """r""" , encoding="""utf-8""" , newline="""\n""") as f: _A = f.readlines() # Get to the point we do the actual imports for type checking _A = 0 _A = {} # Go through the end of the file while line_index < len(snake_case__): # If the line contains is_backend_available, we grab all objects associated with the `else` block _A = find_backend(lines[line_index]) if backend is not None: while not lines[line_index].startswith("""else:"""): line_index += 1 line_index += 1 _A = [] # Until we unindent, add backend objects to the list while line_index < len(snake_case__) and len(lines[line_index]) > 1: _A = lines[line_index] _A = _re_single_line_import.search(snake_case__) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(""", """)) elif line.startswith(""" """ * 8): objects.append(line[8:-2]) line_index += 1 if len(snake_case__) > 0: _A = objects else: line_index += 1 return backend_specific_objects def snake_case ( snake_case__ :Union[str, Any] , snake_case__ :Optional[Any]) -> int: if name.isupper(): return DUMMY_CONSTANT.format(snake_case__) elif name.islower(): return DUMMY_FUNCTION.format(snake_case__ , snake_case__) else: return DUMMY_CLASS.format(snake_case__ , snake_case__) def snake_case ( snake_case__ :Dict=None) -> Optional[Any]: if backend_specific_objects is None: _A = read_init() # For special correspondence backend to module name as used in the function requires_modulename _A = {} for backend, objects in backend_specific_objects.items(): _A = """[""" + """, """.join(F'''"{b}"''' for b in backend.split("""_and_""")) + """]""" _A = """# This file is autogenerated by the command `make fix-copies`, do not edit.\n""" dummy_file += "from ..utils import DummyObject, requires_backends\n\n" dummy_file += "\n".join([create_dummy_object(snake_case__ , snake_case__) for o in objects]) _A = dummy_file return dummy_files def snake_case ( snake_case__ :Any=False) -> Tuple: _A = create_dummy_files() # For special correspondence backend to shortcut as used in utils/dummy_xxx_objects.py _A = {"""torch""": """pt"""} # Locate actual dummy modules and read their content. _A = os.path.join(snake_case__ , """utils""") _A = { backend: os.path.join(snake_case__ , F'''dummy_{short_names.get(snake_case__ , snake_case__)}_objects.py''') for backend in dummy_files.keys() } _A = {} for backend, file_path in dummy_file_paths.items(): if os.path.isfile(snake_case__): with open(snake_case__ , """r""" , encoding="""utf-8""" , newline="""\n""") as f: _A = f.read() else: _A = """""" for backend in dummy_files.keys(): if dummy_files[backend] != actual_dummies[backend]: if overwrite: print( F'''Updating diffusers.utils.dummy_{short_names.get(snake_case__ , snake_case__)}_objects.py as the main ''' """__init__ has new objects.""") with open(dummy_file_paths[backend] , """w""" , encoding="""utf-8""" , newline="""\n""") as f: f.write(dummy_files[backend]) else: raise ValueError( """The main __init__ has objects that are not present in """ F'''diffusers.utils.dummy_{short_names.get(snake_case__ , snake_case__)}_objects.py. Run `make fix-copies` ''' """to fix this.""") if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') _SCREAMING_SNAKE_CASE = parser.parse_args() check_dummies(args.fix_and_overwrite)
81
def snake_case ( snake_case__ :str , snake_case__ :str) -> list: _A = len(snake_case__) _A = [] for i in range(len(snake_case__) - pat_len + 1): _A = True for j in range(snake_case__): if s[i + j] != pattern[j]: _A = False break if match_found: position.append(snake_case__) return position if __name__ == "__main__": assert naive_pattern_search('ABCDEFG', 'DE') == [3] print(naive_pattern_search('ABAAABCDBBABCDDEBCABC', 'ABC'))
81
1
import functools def lowerCamelCase__ ( __lowerCAmelCase : list[int] , __lowerCAmelCase : list[int] ): """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not all(isinstance(__lowerCAmelCase , __lowerCAmelCase ) for day in days ): raise ValueError("The parameter days should be a list of integers" ) if len(__lowerCAmelCase ) != 3 or not all(isinstance(__lowerCAmelCase , __lowerCAmelCase ) for cost in costs ): raise ValueError("The parameter costs should be a list of three integers" ) if len(__lowerCAmelCase ) == 0: return 0 if min(__lowerCAmelCase ) <= 0: raise ValueError("All days elements should be greater than 0" ) if max(__lowerCAmelCase ) >= 366: raise ValueError("All days elements should be less than 366" ) lowerCAmelCase_ = set(__lowerCAmelCase ) @functools.cache def dynamic_programming(__lowerCAmelCase : int ) -> int: if index > 365: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) , costs[1] + dynamic_programming(index + 7 ) , costs[2] + dynamic_programming(index + 30 ) , ) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
231
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _A = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys _A = _LazyModule(__name__, globals()["__file__"], _import_structure)
231
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case = { '''configuration_time_series_transformer''': [ '''TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimeSeriesTransformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ '''TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimeSeriesTransformerForPrediction''', '''TimeSeriesTransformerModel''', '''TimeSeriesTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys __snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
363
"""simple docstring""" import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def A_ ( _lowerCAmelCase : Union[str, Any], _lowerCAmelCase : Tuple, _lowerCAmelCase : Optional[Any], _lowerCAmelCase : Optional[Any], _lowerCAmelCase : Optional[int] ): """simple docstring""" _a = StableDiffusionPipeline.from_pretrained(_lowerCAmelCase, torch_dtype=torch.floataa ) # load LoRA weight from .safetensors _a = load_file(_lowerCAmelCase ) _a = [] # directly update weight in diffusers model for key in state_dict: # it is suggested to print out the key, it usually will be something like below # "lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight" # as we have set the alpha beforehand, so just skip if ".alpha" in key or key in visited: continue if "text" in key: _a = key.split('''.''' )[0].split(LORA_PREFIX_TEXT_ENCODER + '''_''' )[-1].split('''_''' ) _a = pipeline.text_encoder else: _a = key.split('''.''' )[0].split(LORA_PREFIX_UNET + '''_''' )[-1].split('''_''' ) _a = pipeline.unet # find the target layer _a = layer_infos.pop(0 ) while len(_lowerCAmelCase ) > -1: try: _a = curr_layer.__getattr__(_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: _a = layer_infos.pop(0 ) elif len(_lowerCAmelCase ) == 0: break except Exception: if len(_lowerCAmelCase ) > 0: temp_name += "_" + layer_infos.pop(0 ) else: _a = layer_infos.pop(0 ) _a = [] if "lora_down" in key: pair_keys.append(key.replace('''lora_down''', '''lora_up''' ) ) pair_keys.append(_lowerCAmelCase ) else: pair_keys.append(_lowerCAmelCase ) pair_keys.append(key.replace('''lora_up''', '''lora_down''' ) ) # update weight if len(state_dict[pair_keys[0]].shape ) == 4: _a = state_dict[pair_keys[0]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) _a = state_dict[pair_keys[1]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(_lowerCAmelCase, _lowerCAmelCase ).unsqueeze(2 ).unsqueeze(3 ) else: _a = state_dict[pair_keys[0]].to(torch.floataa ) _a = state_dict[pair_keys[1]].to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(_lowerCAmelCase, _lowerCAmelCase ) # update visited list for item in pair_keys: visited.append(_lowerCAmelCase ) return pipeline if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument( '''--base_model_path''', default=None, type=str, required=True, help='''Path to the base model in diffusers format.''' ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--lora_prefix_unet''', default='''lora_unet''', type=str, help='''The prefix of UNet weight in safetensors''' ) parser.add_argument( '''--lora_prefix_text_encoder''', default='''lora_te''', type=str, help='''The prefix of text encoder weight in safetensors''', ) parser.add_argument('''--alpha''', default=0.75, type=float, help='''The merging ratio in W = W0 + alpha * deltaW''') parser.add_argument( '''--to_safetensors''', action='''store_true''', help='''Whether to store pipeline in safetensors format or not.''' ) parser.add_argument('''--device''', type=str, help='''Device to use (e.g. cpu, cuda:0, cuda:1, etc.)''') __snake_case = parser.parse_args() __snake_case = args.base_model_path __snake_case = args.checkpoint_path __snake_case = args.dump_path __snake_case = args.lora_prefix_unet __snake_case = args.lora_prefix_text_encoder __snake_case = args.alpha __snake_case = convert(base_model_path, checkpoint_path, lora_prefix_unet, lora_prefix_text_encoder, alpha) __snake_case = pipe.to(args.device) pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
153
0
import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class SCREAMING_SNAKE_CASE__ (unittest.TestCase ): @require_torch def snake_case_ ( self): lowercase__ : List[str] = pipeline( task='zero-shot-audio-classification' , model='hf-internal-testing/tiny-clap-htsat-unfused') lowercase__ : Union[str, Any] = load_dataset('ashraq/esc50') lowercase__ : Union[str, Any] = dataset['train']['audio'][-1]['array'] lowercase__ : Optional[int] = audio_classifier(a , candidate_labels=['Sound of a dog', 'Sound of vaccum cleaner']) self.assertEqual( nested_simplify(a) , [{'score': 0.501, 'label': 'Sound of a dog'}, {'score': 0.499, 'label': 'Sound of vaccum cleaner'}] , ) @unittest.skip('No models are available in TF') def snake_case_ ( self): pass @slow @require_torch def snake_case_ ( self): lowercase__ : str = pipeline( task='zero-shot-audio-classification' , model='laion/clap-htsat-unfused' , ) # This is an audio of a dog lowercase__ : Optional[Any] = load_dataset('ashraq/esc50') lowercase__ : Union[str, Any] = dataset['train']['audio'][-1]['array'] lowercase__ : int = audio_classifier(a , candidate_labels=['Sound of a dog', 'Sound of vaccum cleaner']) self.assertEqual( nested_simplify(a) , [ {'score': 0.999, 'label': 'Sound of a dog'}, {'score': 0.001, 'label': 'Sound of vaccum cleaner'}, ] , ) lowercase__ : str = audio_classifier([audio] * 5 , candidate_labels=['Sound of a dog', 'Sound of vaccum cleaner']) self.assertEqual( nested_simplify(a) , [ [ {'score': 0.999, 'label': 'Sound of a dog'}, {'score': 0.001, 'label': 'Sound of vaccum cleaner'}, ], ] * 5 , ) lowercase__ : Optional[int] = audio_classifier( [audio] * 5 , candidate_labels=['Sound of a dog', 'Sound of vaccum cleaner'] , batch_size=5) self.assertEqual( nested_simplify(a) , [ [ {'score': 0.999, 'label': 'Sound of a dog'}, {'score': 0.001, 'label': 'Sound of vaccum cleaner'}, ], ] * 5 , ) @unittest.skip('No models are available in TF') def snake_case_ ( self): pass
214
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case_ = { '''configuration_instructblip''': [ '''INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InstructBlipConfig''', '''InstructBlipQFormerConfig''', '''InstructBlipVisionConfig''', ], '''processing_instructblip''': ['''InstructBlipProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ '''INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InstructBlipQFormerModel''', '''InstructBlipPreTrainedModel''', '''InstructBlipForConditionalGeneration''', '''InstructBlipVisionModel''', ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys snake_case_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
214
1
import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'vocab_file': 'vocab.json', 'tokenizer_config_file': 'tokenizer_config.json', 'merges_file': 'merges.txt', } __UpperCAmelCase = { 'vocab_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json' ), }, 'tokenizer_config_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json' ), }, 'merges_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt' ), }, } __UpperCAmelCase = '</w>' __UpperCAmelCase = '@@ ' def lowercase__ ( __snake_case : str ): '''simple docstring''' UpperCAmelCase_ : List[Any] = set() UpperCAmelCase_ : Any = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase_ : Optional[Any] = char return pairs # Speech2Text2 has no max input length __UpperCAmelCase = {'facebook/s2t-wav2vec2-large-en-de': 1024} class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : str = VOCAB_FILES_NAMES _snake_case : Tuple = PRETRAINED_VOCAB_FILES_MAP _snake_case : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : Optional[int] = ['''input_ids''', '''attention_mask'''] def __init__( self , _UpperCamelCase , _UpperCamelCase="<s>" , _UpperCamelCase="<pad>" , _UpperCamelCase="</s>" , _UpperCamelCase="<unk>" , _UpperCamelCase=False , _UpperCamelCase=None , **_UpperCamelCase , ) -> Dict: super().__init__( unk_token=_SCREAMING_SNAKE_CASE , bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , do_lower_case=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) UpperCAmelCase_ : str = do_lower_case with open(_SCREAMING_SNAKE_CASE , encoding='utf-8' ) as vocab_handle: UpperCAmelCase_ : List[Any] = json.load(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"No merges files provided. {self.__class__.__name__} can only be used for decoding." ) UpperCAmelCase_ : List[Any] = None UpperCAmelCase_ : Optional[Any] = None else: with open(_SCREAMING_SNAKE_CASE , encoding='utf-8' ) as merges_handle: UpperCAmelCase_ : Union[str, Any] = merges_handle.read().split('\n' )[:-1] UpperCAmelCase_ : List[str] = [tuple(merge.split()[:2] ) for merge in merges] UpperCAmelCase_ : Union[str, Any] = dict(zip(_SCREAMING_SNAKE_CASE , range(len(_SCREAMING_SNAKE_CASE ) ) ) ) UpperCAmelCase_ : List[str] = {} @property def __UpperCAmelCase ( self ) -> int: return len(self.decoder ) def __UpperCAmelCase ( self ) -> Dict: return dict(self.encoder , **self.added_tokens_encoder ) def __UpperCAmelCase ( self , _UpperCamelCase ) -> Optional[Any]: UpperCAmelCase_ : Any = tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] UpperCAmelCase_ : str = get_pairs(_SCREAMING_SNAKE_CASE ) if not pairs: return token while True: UpperCAmelCase_ : Union[str, Any] = min(_SCREAMING_SNAKE_CASE , key=lambda _UpperCamelCase : self.bpe_ranks.get(_SCREAMING_SNAKE_CASE , float('inf' ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase_ : Dict = bigram UpperCAmelCase_ : str = [] UpperCAmelCase_ : List[Any] = 0 while i < len(_SCREAMING_SNAKE_CASE ): try: UpperCAmelCase_ : List[Any] = word.index(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase_ : Union[str, Any] = j if word[i] == first and i < len(_SCREAMING_SNAKE_CASE ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase_ : int = tuple(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = new_word if len(_SCREAMING_SNAKE_CASE ) == 1: break else: UpperCAmelCase_ : Tuple = get_pairs(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Tuple = ''' '''.join(_SCREAMING_SNAKE_CASE ) if word == "\n " + BPE_TOKEN_MERGES: UpperCAmelCase_ : Tuple = '''\n''' + BPE_TOKEN_MERGES if word.endswith(_SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : List[Any] = word.replace(_SCREAMING_SNAKE_CASE , '' ) UpperCAmelCase_ : Any = word.replace(' ' , _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : str = word return word def __UpperCAmelCase ( self , _UpperCamelCase ) -> Optional[int]: if self.bpe_ranks is None: raise ValueError( 'This tokenizer was instantiated without a `merges.txt` file, so' ' that it can only be used for decoding, not for encoding.' 'Make sure to provide `merges.txt` file at instantiation to enable ' 'encoding.' ) if self.do_lower_case: UpperCAmelCase_ : List[str] = text.lower() UpperCAmelCase_ : List[Any] = text.split() UpperCAmelCase_ : Union[str, Any] = [] for token in text: if token: split_tokens.extend(list(self.bpe(_SCREAMING_SNAKE_CASE ).split(' ' ) ) ) return split_tokens def __UpperCAmelCase ( self , _UpperCamelCase ) -> int: return self.encoder.get(_SCREAMING_SNAKE_CASE , self.encoder.get(self.unk_token ) ) def __UpperCAmelCase ( self , _UpperCamelCase ) -> str: UpperCAmelCase_ : str = self.decoder.get(_SCREAMING_SNAKE_CASE , self.unk_token ) return result def __UpperCAmelCase ( self , _UpperCamelCase ) -> str: UpperCAmelCase_ : List[Any] = ''' '''.join(_SCREAMING_SNAKE_CASE ) # make sure @@ tokens are concatenated UpperCAmelCase_ : List[Any] = ''''''.join(string.split(_SCREAMING_SNAKE_CASE ) ) return string def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase = None ) -> Tuple[str]: if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return UpperCAmelCase_ : List[Any] = os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase_ : Dict = os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(_SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_SCREAMING_SNAKE_CASE , ensure_ascii=_SCREAMING_SNAKE_CASE ) + '\n' ) UpperCAmelCase_ : Optional[Any] = 0 if self.bpe_ranks is None: return (vocab_file,) with open(_SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _UpperCamelCase : kv[1] ): if index != token_index: logger.warning( f"Saving vocabulary to {merges_file}: BPE merge indices are not consecutive." ' Please check that the tokenizer is not corrupted!' ) UpperCAmelCase_ : Optional[Any] = token_index writer.write(' '.join(_SCREAMING_SNAKE_CASE ) + '\n' ) index += 1 return (vocab_file, merges_file)
371
from collections import defaultdict from math import ceil, sqrt def lowercase__ ( __snake_case : int = 1_000_000 , __snake_case : int = 10 ): '''simple docstring''' UpperCAmelCase_ : defaultdict = defaultdict(__snake_case ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_limit: UpperCAmelCase_ : Union[str, Any] = max( ceil(sqrt(outer_width * outer_width - t_limit ) ) , 1 ) else: UpperCAmelCase_ : int = 1 hole_width_lower_bound += (outer_width - hole_width_lower_bound) % 2 for hole_width in range(__snake_case , outer_width - 1 , 2 ): count[outer_width * outer_width - hole_width * hole_width] += 1 return sum(1 for n in count.values() if 1 <= n <= 10 ) if __name__ == "__main__": print(F'{solution() = }')
145
0
import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() A__ = logging.get_logger(__name__) A__ = {name: getattr(transformers, name + '''Fast''') for name in SLOW_TO_FAST_CONVERTERS} def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: """simple docstring""" if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(f"""Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.""" ) if tokenizer_name is None: snake_case__ : Optional[Any] = TOKENIZER_CLASSES else: snake_case__ : Tuple = {tokenizer_name: getattr(__lowerCAmelCase , tokenizer_name + '''Fast''' )} logger.info(f"""Loading tokenizer classes: {tokenizer_names}""" ) for tokenizer_name in tokenizer_names: snake_case__ : Dict = TOKENIZER_CLASSES[tokenizer_name] snake_case__ : str = True if checkpoint_name is None: snake_case__ : str = list(tokenizer_class.max_model_input_sizes.keys() ) else: snake_case__ : Union[str, Any] = [checkpoint_name] logger.info(f"""For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}""" ) for checkpoint in checkpoint_names: logger.info(f"""Loading {tokenizer_class.__class__.__name__} {checkpoint}""" ) # Load tokenizer snake_case__ : List[Any] = tokenizer_class.from_pretrained(__lowerCAmelCase , force_download=__lowerCAmelCase ) # Save fast tokenizer logger.info(f"""Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}""" ) # For organization names we create sub-directories if "/" in checkpoint: snake_case__ , snake_case__ : str = checkpoint.split('''/''' ) snake_case__ : int = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) elif add_prefix: snake_case__ : List[Any] = checkpoint snake_case__ : Any = dump_path else: snake_case__ : List[str] = None snake_case__ : int = dump_path logger.info(f"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: snake_case__ : Tuple = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] snake_case__ : Optional[Any] = file_path.split(__lowerCAmelCase )[-1][0] if next_char == "/": snake_case__ : Any = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : str = None logger.info(f"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) snake_case__ : Optional[int] = tokenizer.save_pretrained( __lowerCAmelCase , legacy_format=__lowerCAmelCase , filename_prefix=__lowerCAmelCase ) logger.info(f"""=> File names {file_names}""" ) for file_name in file_names: if not file_name.endswith('''tokenizer.json''' ): os.remove(__lowerCAmelCase ) logger.info(f"""=> removing {file_name}""" ) if __name__ == "__main__": A__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output generated fast tokenizer files.''' ) parser.add_argument( '''--tokenizer_name''', default=None, type=str, help=( f"""Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will """ '''download and convert all the checkpoints from AWS.''' ), ) parser.add_argument( '''--checkpoint_name''', default=None, type=str, help='''Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.''', ) parser.add_argument( '''--force_download''', action='''store_true''', help='''Re-download checkpoints.''', ) A__ = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
230
import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from utils_ner import Split, TokenClassificationDataset, TokenClassificationTask import transformers from transformers import ( AutoConfig, AutoModelForTokenClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process A__ = logging.getLogger(__name__) @dataclass class a : __lowerCAmelCase : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __lowerCAmelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __lowerCAmelCase : Optional[str] = field( default="""NER""" , metadata={"""help""": """Task type to fine tune in training (e.g. NER, POS, etc)"""} ) __lowerCAmelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __lowerCAmelCase : bool = field(default=__lowerCamelCase , metadata={"""help""": """Set this flag to use fast tokenization."""} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. __lowerCAmelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class a : __lowerCAmelCase : str = field( metadata={"""help""": """The input data dir. Should contain the .txt files for a CoNLL-2003-formatted task."""} ) __lowerCAmelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Path to a file containing all labels. If not specified, CoNLL-2003 labels are used."""} , ) __lowerCAmelCase : int = field( default=1_28 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __lowerCAmelCase : bool = field( default=__lowerCamelCase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def _lowerCAmelCase ( ) -> Dict: """simple docstring""" snake_case__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case__ , snake_case__ , snake_case__ : str = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case__ , snake_case__ , snake_case__ : List[Any] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ''' --overwrite_output_dir to overcome.''' ) snake_case__ : int = import_module('''tasks''' ) try: snake_case__ : Optional[int] = getattr(__lowerCAmelCase , model_args.task_type ) snake_case__ : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( f"""Task {model_args.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. """ f"""Available tasks classes are: {TokenClassificationTask.__subclasses__()}""" ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , __lowerCAmelCase ) # Set seed set_seed(training_args.seed ) # Prepare CONLL-2003 task snake_case__ : Optional[int] = token_classification_task.get_labels(data_args.labels ) snake_case__ : Dict[int, str] = dict(enumerate(__lowerCAmelCase ) ) snake_case__ : Optional[Any] = len(__lowerCAmelCase ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case__ : List[str] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__lowerCAmelCase , idalabel=__lowerCAmelCase , labelaid={label: i for i, label in enumerate(__lowerCAmelCase )} , cache_dir=model_args.cache_dir , ) snake_case__ : List[str] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast , ) snake_case__ : Tuple = AutoModelForTokenClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , ) # Get datasets snake_case__ : Dict = ( TokenClassificationDataset( token_classification_task=__lowerCAmelCase , data_dir=data_args.data_dir , tokenizer=__lowerCAmelCase , labels=__lowerCAmelCase , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case__ : Optional[Any] = ( TokenClassificationDataset( token_classification_task=__lowerCAmelCase , data_dir=data_args.data_dir , tokenizer=__lowerCAmelCase , labels=__lowerCAmelCase , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def align_predictions(__lowerCAmelCase , __lowerCAmelCase ) -> Tuple[List[int], List[int]]: snake_case__ : Any = np.argmax(__lowerCAmelCase , axis=2 ) snake_case__ , snake_case__ : List[Any] = preds.shape snake_case__ : List[Any] = [[] for _ in range(__lowerCAmelCase )] snake_case__ : int = [[] for _ in range(__lowerCAmelCase )] for i in range(__lowerCAmelCase ): for j in range(__lowerCAmelCase ): if label_ids[i, j] != nn.CrossEntropyLoss().ignore_index: out_label_list[i].append(label_map[label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) return preds_list, out_label_list def compute_metrics(__lowerCAmelCase ) -> Dict: snake_case__ , snake_case__ : List[str] = align_predictions(p.predictions , p.label_ids ) return { "accuracy_score": accuracy_score(__lowerCAmelCase , __lowerCAmelCase ), "precision": precision_score(__lowerCAmelCase , __lowerCAmelCase ), "recall": recall_score(__lowerCAmelCase , __lowerCAmelCase ), "f1": fa_score(__lowerCAmelCase , __lowerCAmelCase ), } # Data collator snake_case__ : Any = DataCollatorWithPadding(__lowerCAmelCase , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case__ : Tuple = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=__lowerCAmelCase , eval_dataset=__lowerCAmelCase , compute_metrics=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_process_zero(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case__ : Any = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case__ : Tuple = trainer.evaluate() snake_case__ : Optional[Any] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_process_zero(): with open(__lowerCAmelCase , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , __lowerCAmelCase , __lowerCAmelCase ) writer.write('''%s = %s\n''' % (key, value) ) results.update(__lowerCAmelCase ) # Predict if training_args.do_predict: snake_case__ : Optional[Any] = TokenClassificationDataset( token_classification_task=__lowerCAmelCase , data_dir=data_args.data_dir , tokenizer=__lowerCAmelCase , labels=__lowerCAmelCase , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.test , ) snake_case__ , snake_case__ , snake_case__ : List[str] = trainer.predict(__lowerCAmelCase ) snake_case__ , snake_case__ : int = align_predictions(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : Optional[int] = os.path.join(training_args.output_dir , '''test_results.txt''' ) if trainer.is_world_process_zero(): with open(__lowerCAmelCase , '''w''' ) as writer: for key, value in metrics.items(): logger.info(''' %s = %s''' , __lowerCAmelCase , __lowerCAmelCase ) writer.write('''%s = %s\n''' % (key, value) ) # Save predictions snake_case__ : Any = os.path.join(training_args.output_dir , '''test_predictions.txt''' ) if trainer.is_world_process_zero(): with open(__lowerCAmelCase , '''w''' ) as writer: with open(os.path.join(data_args.data_dir , '''test.txt''' ) , '''r''' ) as f: token_classification_task.write_predictions_to_file(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) return results def _lowerCAmelCase ( __lowerCAmelCase ) -> Union[str, Any]: """simple docstring""" main() if __name__ == "__main__": main()
230
1
from __future__ import annotations from math import pi def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): if (inductance, frequency, reactance).count(0) != 1: raise ValueError('One and only one argument must be 0') if inductance < 0: raise ValueError('Inductance cannot be negative') if frequency < 0: raise ValueError('Frequency cannot be negative') if reactance < 0: raise ValueError('Inductive reactance cannot be negative') if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError('Exactly one argument must be 0') if __name__ == "__main__": import doctest doctest.testmod()
327
import logging import os import quant_trainer import torch from torch.utils.data import DataLoader from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput a_ : Dict = logging.getLogger(__name__) if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class _snake_case ( A__ ): def __init__( self , *a , a=None , a=None , a=None , **a) -> List[Any]: super().__init__(*a , **a) SCREAMING_SNAKE_CASE = eval_examples SCREAMING_SNAKE_CASE = post_process_function SCREAMING_SNAKE_CASE = quant_trainer_args SCREAMING_SNAKE_CASE = 128 # default number of calibration samples def SCREAMING_SNAKE_CASE__ ( self , a=None) -> Union[str, Any]: if calib_dataset is None and self.calib_dataset is None: raise ValueError('Trainer: calibration requires an calib_dataset.') SCREAMING_SNAKE_CASE = calib_dataset if calib_dataset is not None else self.calib_dataset SCREAMING_SNAKE_CASE = self._remove_unused_columns(a , description='Calibration') return DataLoader( a , batch_size=self.args.eval_batch_size , collate_fn=self.data_collator , drop_last=self.args.dataloader_drop_last , num_workers=self.args.dataloader_num_workers , pin_memory=self.args.dataloader_pin_memory , shuffle=a , ) def SCREAMING_SNAKE_CASE__ ( self , a=None) -> Optional[Any]: SCREAMING_SNAKE_CASE = self.train_dataset if calib_dataset is None else calib_dataset SCREAMING_SNAKE_CASE = self.get_calib_dataloader(a) SCREAMING_SNAKE_CASE = self.model quant_trainer.configure_model(a , self.quant_trainer_args , calib=a) model.eval() quant_trainer.enable_calibration(a) logger.info('***** Running calibration *****') logger.info(f''' Num examples = {self.calib_num}''') logger.info(f''' Batch size = {calib_dataloader.batch_size}''') for step, inputs in enumerate(a): # Prediction step SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.prediction_step(a , a , prediction_loss_only=a) if (step + 1) * calib_dataloader.batch_size >= self.calib_num: break quant_trainer.finish_calibration(a , self.quant_trainer_args) SCREAMING_SNAKE_CASE = model def SCREAMING_SNAKE_CASE__ ( self , a=None , a=None , a=None , a = "eval") -> str: SCREAMING_SNAKE_CASE = self.eval_dataset if eval_dataset is None else eval_dataset SCREAMING_SNAKE_CASE = self.get_eval_dataloader(a) SCREAMING_SNAKE_CASE = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. SCREAMING_SNAKE_CASE = self.compute_metrics SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: SCREAMING_SNAKE_CASE = eval_loop( a , description='Evaluation' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=a , ) finally: SCREAMING_SNAKE_CASE = compute_metrics if self.post_process_function is not None and self.compute_metrics is not None: SCREAMING_SNAKE_CASE = self.post_process_function(a , a , output.predictions) SCREAMING_SNAKE_CASE = self.compute_metrics(a) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'''{metric_key_prefix}_'''): SCREAMING_SNAKE_CASE = metrics.pop(a) self.log(a) else: SCREAMING_SNAKE_CASE = {} if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report()) SCREAMING_SNAKE_CASE = self.callback_handler.on_evaluate(self.args , self.state , self.control , a) return metrics def SCREAMING_SNAKE_CASE__ ( self , a , a , a=None , a = "test") -> Optional[Any]: SCREAMING_SNAKE_CASE = self.get_test_dataloader(a) # Temporarily disable metric computation, we will do it in the loop here. SCREAMING_SNAKE_CASE = self.compute_metrics SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: SCREAMING_SNAKE_CASE = eval_loop( a , description='Prediction' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=a , ) finally: SCREAMING_SNAKE_CASE = compute_metrics if self.post_process_function is None or self.compute_metrics is None: return output SCREAMING_SNAKE_CASE = self.post_process_function(a , a , output.predictions , 'predict') SCREAMING_SNAKE_CASE = self.compute_metrics(a) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'''{metric_key_prefix}_'''): SCREAMING_SNAKE_CASE = metrics.pop(a) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=a) def SCREAMING_SNAKE_CASE__ ( self , a="./") -> List[Any]: SCREAMING_SNAKE_CASE = self.eval_dataset SCREAMING_SNAKE_CASE = self.get_eval_dataloader(a) SCREAMING_SNAKE_CASE = next(iter(a)) # saving device - to make it consistent SCREAMING_SNAKE_CASE = torch.device('cuda' if torch.cuda.is_available() else 'cpu') # convert to tuple SCREAMING_SNAKE_CASE = tuple(v.to(a) for k, v in batch.items()) logger.info('Converting model to be onnx compatible') from pytorch_quantization.nn import TensorQuantizer SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = self.model.to(a) model.eval() model.float() SCREAMING_SNAKE_CASE = model.module if hasattr(a , 'module') else model quant_trainer.configure_model(a , self.quant_trainer_args) SCREAMING_SNAKE_CASE = os.path.join(a , 'model.onnx') logger.info(f'''exporting model to {output_model_file}''') SCREAMING_SNAKE_CASE = {0: 'batch_size', 1: 'seq_len'} torch.onnx.export( a , a , a , export_params=a , opset_version=13 , do_constant_folding=a , input_names=['input_ids', 'attention_mask', 'token_type_ids'] , output_names=['output_start_logits', 'output_end_logits'] , dynamic_axes={ 'input_ids': axes, 'attention_mask': axes, 'token_type_ids': axes, 'output_start_logits': axes, 'output_end_logits': axes, } , verbose=a , ) logger.info('onnx export finished')
327
1
from __future__ import annotations class __snake_case : def __init__( self : Tuple , _snake_case : Dict=None): """simple docstring""" UpperCAmelCase_ = data UpperCAmelCase_ = None def __repr__( self : str): """simple docstring""" UpperCAmelCase_ = [] UpperCAmelCase_ = self while temp: string_rep.append(F"""{temp.data}""") UpperCAmelCase_ = temp.next return "->".join(_snake_case) def A (__A : list ) -> List[str]: """simple docstring""" if not elements_list: raise Exception('''The Elements List is empty''' ) UpperCAmelCase_ = UpperCAmelCase_ = Node(elements_list[0] ) for i in range(1 , len(__A ) ): UpperCAmelCase_ = Node(elements_list[i] ) UpperCAmelCase_ = current.next return head def A (__A : Node ) -> None: """simple docstring""" if head_node is not None and isinstance(__A , __A ): print_reverse(head_node.next ) print(head_node.data ) def A () -> List[str]: """simple docstring""" from doctest import testmod testmod() UpperCAmelCase_ = make_linked_list([14, 52, 14, 12, 43] ) print('''Linked List:''' ) print(__A ) print('''Elements in Reverse:''' ) print_reverse(__A ) if __name__ == "__main__": main()
51
# This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class __snake_case ( a , a , a , unittest.TestCase ): UpperCAmelCase__ : List[Any] = StableDiffusionControlNetImgaImgPipeline UpperCAmelCase__ : Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} UpperCAmelCase__ : Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCAmelCase__ : Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({'''control_image'''} ) UpperCAmelCase__ : Optional[Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase ( self : int): """simple docstring""" torch.manual_seed(0) UpperCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) torch.manual_seed(0) UpperCAmelCase_ = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0) UpperCAmelCase_ = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='''scaled_linear''' , clip_sample=_snake_case , set_alpha_to_one=_snake_case , ) torch.manual_seed(0) UpperCAmelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0) UpperCAmelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) UpperCAmelCase_ = CLIPTextModel(_snake_case) UpperCAmelCase_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') UpperCAmelCase_ = { '''unet''': unet, '''controlnet''': controlnet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def lowerCamelCase ( self : Union[str, Any] , _snake_case : Any , _snake_case : Dict=0): """simple docstring""" if str(_snake_case).startswith('''mps'''): UpperCAmelCase_ = torch.manual_seed(_snake_case) else: UpperCAmelCase_ = torch.Generator(device=_snake_case).manual_seed(_snake_case) UpperCAmelCase_ = 2 UpperCAmelCase_ = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=_snake_case , device=torch.device(_snake_case) , ) UpperCAmelCase_ = floats_tensor(control_image.shape , rng=random.Random(_snake_case)).to(_snake_case) UpperCAmelCase_ = image.cpu().permute(0 , 2 , 3 , 1)[0] UpperCAmelCase_ = Image.fromarray(np.uinta(_snake_case)).convert('''RGB''').resize((64, 64)) UpperCAmelCase_ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', '''image''': image, '''control_image''': control_image, } return inputs def lowerCamelCase ( self : Any): """simple docstring""" return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def lowerCamelCase ( self : Any): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3) def lowerCamelCase ( self : Optional[Any]): """simple docstring""" self._test_inference_batch_single_identical(expected_max_diff=2e-3) class __snake_case ( a , a , unittest.TestCase ): UpperCAmelCase__ : str = StableDiffusionControlNetImgaImgPipeline UpperCAmelCase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} UpperCAmelCase__ : Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCAmelCase__ : str = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def lowerCamelCase ( self : str): """simple docstring""" torch.manual_seed(0) UpperCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) torch.manual_seed(0) def init_weights(_snake_case : Optional[int]): if isinstance(_snake_case , torch.nn.Convad): torch.nn.init.normal(m.weight) m.bias.data.fill_(1.0) UpperCAmelCase_ = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(_snake_case) torch.manual_seed(0) UpperCAmelCase_ = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(_snake_case) torch.manual_seed(0) UpperCAmelCase_ = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='''scaled_linear''' , clip_sample=_snake_case , set_alpha_to_one=_snake_case , ) torch.manual_seed(0) UpperCAmelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0) UpperCAmelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) UpperCAmelCase_ = CLIPTextModel(_snake_case) UpperCAmelCase_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') UpperCAmelCase_ = MultiControlNetModel([controlneta, controlneta]) UpperCAmelCase_ = { '''unet''': unet, '''controlnet''': controlnet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def lowerCamelCase ( self : int , _snake_case : Union[str, Any] , _snake_case : str=0): """simple docstring""" if str(_snake_case).startswith('''mps'''): UpperCAmelCase_ = torch.manual_seed(_snake_case) else: UpperCAmelCase_ = torch.Generator(device=_snake_case).manual_seed(_snake_case) UpperCAmelCase_ = 2 UpperCAmelCase_ = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=_snake_case , device=torch.device(_snake_case) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=_snake_case , device=torch.device(_snake_case) , ), ] UpperCAmelCase_ = floats_tensor(control_image[0].shape , rng=random.Random(_snake_case)).to(_snake_case) UpperCAmelCase_ = image.cpu().permute(0 , 2 , 3 , 1)[0] UpperCAmelCase_ = Image.fromarray(np.uinta(_snake_case)).convert('''RGB''').resize((64, 64)) UpperCAmelCase_ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', '''image''': image, '''control_image''': control_image, } return inputs def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ = self.get_dummy_components() UpperCAmelCase_ = self.pipeline_class(**_snake_case) pipe.to(_snake_case) UpperCAmelCase_ = 1_0.0 UpperCAmelCase_ = 4 UpperCAmelCase_ = self.get_dummy_inputs(_snake_case) UpperCAmelCase_ = steps UpperCAmelCase_ = scale UpperCAmelCase_ = pipe(**_snake_case)[0] UpperCAmelCase_ = self.get_dummy_inputs(_snake_case) UpperCAmelCase_ = steps UpperCAmelCase_ = scale UpperCAmelCase_ = pipe(**_snake_case , control_guidance_start=0.1 , control_guidance_end=0.2)[0] UpperCAmelCase_ = self.get_dummy_inputs(_snake_case) UpperCAmelCase_ = steps UpperCAmelCase_ = scale UpperCAmelCase_ = pipe(**_snake_case , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7])[0] UpperCAmelCase_ = self.get_dummy_inputs(_snake_case) UpperCAmelCase_ = steps UpperCAmelCase_ = scale UpperCAmelCase_ = pipe(**_snake_case , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8])[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a)) > 1e-3 assert np.sum(np.abs(output_a - output_a)) > 1e-3 assert np.sum(np.abs(output_a - output_a)) > 1e-3 def lowerCamelCase ( self : Dict): """simple docstring""" return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def lowerCamelCase ( self : int): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3) def lowerCamelCase ( self : int): """simple docstring""" self._test_inference_batch_single_identical(expected_max_diff=2e-3) def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = self.get_dummy_components() UpperCAmelCase_ = self.pipeline_class(**_snake_case) pipe.to(_snake_case) pipe.set_progress_bar_config(disable=_snake_case) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(_snake_case) except NotImplementedError: pass @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase ( self : Optional[int]): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = ControlNetModel.from_pretrained('''lllyasviel/sd-controlnet-canny''') UpperCAmelCase_ = StableDiffusionControlNetImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , safety_checker=_snake_case , controlnet=_snake_case) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=_snake_case) UpperCAmelCase_ = torch.Generator(device='''cpu''').manual_seed(0) UpperCAmelCase_ = '''evil space-punk bird''' UpperCAmelCase_ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png''').resize((512, 512)) UpperCAmelCase_ = load_image( '''https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png''').resize((512, 512)) UpperCAmelCase_ = pipe( _snake_case , _snake_case , control_image=_snake_case , generator=_snake_case , output_type='''np''' , num_inference_steps=50 , strength=0.6 , ) UpperCAmelCase_ = output.images[0] assert image.shape == (512, 512, 3) UpperCAmelCase_ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy''') assert np.abs(expected_image - image).max() < 9e-2
51
1
import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType _SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) class UpperCAmelCase__ ( A__ ): """simple docstring""" a = "vision-encoder-decoder" a = True def __init__( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: super().__init__(**__lowerCamelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f'''A configuraton of type {self.model_type} cannot be instantiated because ''' f'''not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}''' ) SCREAMING_SNAKE_CASE__ = kwargs.pop('''encoder''' ) SCREAMING_SNAKE_CASE__ = encoder_config.pop('''model_type''' ) SCREAMING_SNAKE_CASE__ = kwargs.pop('''decoder''' ) SCREAMING_SNAKE_CASE__ = decoder_config.pop('''model_type''' ) SCREAMING_SNAKE_CASE__ = AutoConfig.for_model(__lowerCamelCase , **__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = AutoConfig.for_model(__lowerCamelCase , **__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = True @classmethod def lowercase_ ( cls : Union[str, Any] , __lowerCamelCase : PretrainedConfig , __lowerCamelCase : PretrainedConfig , **__lowerCamelCase : Optional[int] ) -> PretrainedConfig: logger.info('''Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config''' ) SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **__lowerCamelCase ) def lowercase_ ( self : Tuple ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.encoder.to_dict() SCREAMING_SNAKE_CASE__ = self.decoder.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output class UpperCAmelCase__ ( A__ ): """simple docstring""" a = version.parse("1.11" ) @property def lowercase_ ( self : Any ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowercase_ ( self : List[str] ) -> float: return 1e-4 @property def lowercase_ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({'''last_hidden_state''': {0: '''batch''', 1: '''encoder_sequence'''}} ) class UpperCAmelCase__ ( A__ ): """simple docstring""" @property def lowercase_ ( self : Optional[Any] ) -> Mapping[str, Mapping[int, str]]: SCREAMING_SNAKE_CASE__ = OrderedDict() SCREAMING_SNAKE_CASE__ = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} SCREAMING_SNAKE_CASE__ = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} SCREAMING_SNAKE_CASE__ = {0: '''batch''', 1: '''encoder_sequence'''} return common_inputs def lowercase_ ( self : List[Any] , __lowerCamelCase : "PreTrainedTokenizerBase" , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional["TensorType"] = None , ) -> Mapping[str, Any]: import torch SCREAMING_SNAKE_CASE__ = OrderedDict() SCREAMING_SNAKE_CASE__ = super().generate_dummy_inputs( __lowerCamelCase , batch_size=__lowerCamelCase , seq_length=__lowerCamelCase , is_pair=__lowerCamelCase , framework=__lowerCamelCase ) SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = dummy_input['''input_ids'''].shape SCREAMING_SNAKE_CASE__ = (batch, encoder_sequence, self._config.encoder_hidden_size) SCREAMING_SNAKE_CASE__ = dummy_input.pop('''input_ids''' ) SCREAMING_SNAKE_CASE__ = dummy_input.pop('''attention_mask''' ) SCREAMING_SNAKE_CASE__ = torch.zeros(__lowerCamelCase ) return common_inputs class UpperCAmelCase__ ( A__ ): """simple docstring""" @property def lowercase_ ( self : Any ) -> None: pass def lowercase_ ( self : Optional[Any] , __lowerCamelCase : PretrainedConfig ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(__lowerCamelCase ) def lowercase_ ( self : Union[str, Any] , __lowerCamelCase : PretrainedConfig , __lowerCamelCase : PretrainedConfig , __lowerCamelCase : str = "default" ) -> OnnxConfig: SCREAMING_SNAKE_CASE__ = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(__lowerCamelCase , __lowerCamelCase )
365
from functools import reduce _SCREAMING_SNAKE_CASE : Any = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCAmelCase_ ( _A = N ): '''simple docstring''' return max( # mypy cannot properly interpret reduce int(reduce(lambda _A , _A : str(int(_A ) * int(_A ) ) , n[i : i + 13] ) ) for i in range(len(_A ) - 12 ) ) if __name__ == "__main__": print(F"{solution() = }")
218
0
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowercase_ = logging.get_logger(__name__) lowercase_ = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} lowercase_ = { 'vocab_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/vocab.json', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/vocab.json', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/vocab.json', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/vocab.json', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/vocab.json', }, 'merges_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/merges.txt', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/merges.txt', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/merges.txt', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/merges.txt', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/merges.txt', }, 'tokenizer_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/tokenizer.json', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/tokenizer.json', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/tokenizer.json', }, } lowercase_ = { 'gpt2': 1_024, 'gpt2-medium': 1_024, 'gpt2-large': 1_024, 'gpt2-xl': 1_024, 'distilgpt2': 1_024, } class __UpperCamelCase ( UpperCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = VOCAB_FILES_NAMES lowerCAmelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ = ["input_ids", "attention_mask"] lowerCAmelCase_ = GPTaTokenizer def __init__( self : List[str] , _A : Any=None , _A : List[Any]=None , _A : Union[str, Any]=None , _A : List[str]="<|endoftext|>" , _A : str="<|endoftext|>" , _A : Optional[int]="<|endoftext|>" , _A : str=False , **_A : Dict , ): """simple docstring""" super().__init__( snake_case__ , snake_case__ , tokenizer_file=snake_case__ , unk_token=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , add_prefix_space=snake_case__ , **snake_case__ , ) __SCREAMING_SNAKE_CASE : int = kwargs.pop('''add_bos_token''' , snake_case__ ) __SCREAMING_SNAKE_CASE : List[str] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , snake_case__ ) != add_prefix_space: __SCREAMING_SNAKE_CASE : int = getattr(snake_case__ , pre_tok_state.pop('''type''' ) ) __SCREAMING_SNAKE_CASE : List[str] = add_prefix_space __SCREAMING_SNAKE_CASE : Tuple = pre_tok_class(**snake_case__ ) __SCREAMING_SNAKE_CASE : Optional[Any] = add_prefix_space def UpperCAmelCase__ ( self : Union[str, Any] , *_A : int , **_A : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Tuple = kwargs.get('''is_split_into_words''' , snake_case__ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*snake_case__ , **snake_case__ ) def UpperCAmelCase__ ( self : int , *_A : Optional[Any] , **_A : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = kwargs.get('''is_split_into_words''' , snake_case__ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*snake_case__ , **snake_case__ ) def UpperCAmelCase__ ( self : Tuple , _A : str , _A : Optional[str] = None ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = self._tokenizer.model.save(snake_case__ , name=snake_case__ ) return tuple(snake_case__ ) def UpperCAmelCase__ ( self : Dict , _A : "Conversation" ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(snake_case__ , add_special_tokens=snake_case__ ) + [self.eos_token_id] ) if len(snake_case__ ) > self.model_max_length: __SCREAMING_SNAKE_CASE : List[Any] = input_ids[-self.model_max_length :] return input_ids
303
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowercase_ : str = logging.get_logger(__name__) lowercase_ : Union[str, Any] = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} lowercase_ : Optional[Any] = { 'vocab_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/vocab.json', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/vocab.json', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/vocab.json', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/vocab.json', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/vocab.json', }, 'merges_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/merges.txt', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/merges.txt', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/merges.txt', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/merges.txt', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/merges.txt', }, 'tokenizer_file': { 'gpt2': 'https://huggingface.co/gpt2/resolve/main/tokenizer.json', 'gpt2-medium': 'https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json', 'gpt2-large': 'https://huggingface.co/gpt2-large/resolve/main/tokenizer.json', 'gpt2-xl': 'https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json', 'distilgpt2': 'https://huggingface.co/distilgpt2/resolve/main/tokenizer.json', }, } lowercase_ : int = { 'gpt2': 10_24, 'gpt2-medium': 10_24, 'gpt2-large': 10_24, 'gpt2-xl': 10_24, 'distilgpt2': 10_24, } class __lowerCAmelCase ( UpperCAmelCase__ ): snake_case_ : Optional[int] = VOCAB_FILES_NAMES snake_case_ : List[str] = PRETRAINED_VOCAB_FILES_MAP snake_case_ : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ : Tuple = ["input_ids", "attention_mask"] snake_case_ : str = GPTaTokenizer def __init__( self : List[str] , snake_case__ : Any=None , snake_case__ : List[Any]=None , snake_case__ : Union[str, Any]=None , snake_case__ : List[str]="<|endoftext|>" , snake_case__ : str="<|endoftext|>" , snake_case__ : Optional[int]="<|endoftext|>" , snake_case__ : str=False , **snake_case__ : Dict , ): """simple docstring""" super().__init__( snake_case__ , snake_case__ , tokenizer_file=snake_case__ , unk_token=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , add_prefix_space=snake_case__ , **snake_case__ , ) _UpperCAmelCase = kwargs.pop("add_bos_token" , snake_case__ ) _UpperCAmelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , snake_case__ ) != add_prefix_space: _UpperCAmelCase = getattr(snake_case__ , pre_tok_state.pop("type" ) ) _UpperCAmelCase = add_prefix_space _UpperCAmelCase = pre_tok_class(**snake_case__ ) _UpperCAmelCase = add_prefix_space def UpperCamelCase ( self : Union[str, Any] , *snake_case__ : int , **snake_case__ : Union[str, Any] ): """simple docstring""" _UpperCAmelCase = kwargs.get("is_split_into_words" , snake_case__ ) assert self.add_prefix_space or not is_split_into_words, ( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*snake_case__ , **snake_case__ ) def UpperCamelCase ( self : int , *snake_case__ : Optional[Any] , **snake_case__ : List[str] ): """simple docstring""" _UpperCAmelCase = kwargs.get("is_split_into_words" , snake_case__ ) assert self.add_prefix_space or not is_split_into_words, ( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._encode_plus(*snake_case__ , **snake_case__ ) def UpperCamelCase ( self : Tuple , snake_case__ : str , snake_case__ : Optional[str] = None ): """simple docstring""" _UpperCAmelCase = self._tokenizer.model.save(snake_case__ , name=snake_case__ ) return tuple(snake_case__ ) def UpperCamelCase ( self : Dict , snake_case__ : "Conversation" ): """simple docstring""" _UpperCAmelCase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(snake_case__ , add_special_tokens=snake_case__ ) + [self.eos_token_id] ) if len(snake_case__ ) > self.model_max_length: _UpperCAmelCase = input_ids[-self.model_max_length :] return input_ids
133
0
"""simple docstring""" import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( __lowerCAmelCase , unittest.TestCase ): """simple docstring""" a : List[str] =FunnelTokenizer a : List[str] =FunnelTokenizerFast a : Optional[Any] =True a : Dict =True def lowercase__ ( self ): """simple docstring""" super().setUp() lowerCAmelCase : Tuple = [ '''<unk>''', '''<cls>''', '''<sep>''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] lowerCAmelCase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def lowercase__ ( self , **snake_case__ ): """simple docstring""" return FunnelTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def lowercase__ ( self , **snake_case__ ): """simple docstring""" return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : List[str] = '''UNwant\u00E9d,running''' lowerCAmelCase : Dict = '''unwanted, running''' return input_text, output_text def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : str = self.tokenizer_class(self.vocab_file ) lowerCAmelCase : Dict = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(lowerCamelCase__ , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [7, 4, 5, 10, 8, 9] ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : int = self.get_tokenizers(do_lower_case=lowerCamelCase__ ) for tokenizer in tokenizers: lowerCAmelCase : Union[str, Any] = tokenizer("UNwant\u00E9d,running" ) lowerCAmelCase : Optional[int] = len(inputs["input_ids"] ) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len ) lowerCAmelCase : Any = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running" ) self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len )
357
"""simple docstring""" from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" a : Dict ="autoformer" a : Dict ={ "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", "num_hidden_layers": "encoder_layers", } def __init__( self , snake_case__ = None , snake_case__ = None , snake_case__ = "student_t" , snake_case__ = "nll" , snake_case__ = 1 , snake_case__ = [1, 2, 3, 4, 5, 6, 7] , snake_case__ = True , snake_case__ = 0 , snake_case__ = 0 , snake_case__ = 0 , snake_case__ = 0 , snake_case__ = None , snake_case__ = None , snake_case__ = 64 , snake_case__ = 2 , snake_case__ = 2 , snake_case__ = 2 , snake_case__ = 2 , snake_case__ = 32 , snake_case__ = 32 , snake_case__ = "gelu" , snake_case__ = 0.1 , snake_case__ = 0.1 , snake_case__ = 0.1 , snake_case__ = 0.1 , snake_case__ = 0.1 , snake_case__ = 100 , snake_case__ = 0.02 , snake_case__ = True , snake_case__=True , snake_case__ = 10 , snake_case__ = 25 , snake_case__ = 3 , **snake_case__ , ): """simple docstring""" lowerCAmelCase : Any = prediction_length lowerCAmelCase : Dict = context_length if context_length is not None else prediction_length lowerCAmelCase : Tuple = distribution_output lowerCAmelCase : List[Any] = loss lowerCAmelCase : int = input_size lowerCAmelCase : str = num_time_features lowerCAmelCase : str = lags_sequence lowerCAmelCase : List[str] = scaling lowerCAmelCase : List[Any] = num_dynamic_real_features lowerCAmelCase : Tuple = num_static_real_features lowerCAmelCase : Dict = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(snake_case__ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) lowerCAmelCase : Any = cardinality else: lowerCAmelCase : Union[str, Any] = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(snake_case__ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) lowerCAmelCase : Tuple = embedding_dimension else: lowerCAmelCase : Any = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] lowerCAmelCase : Any = num_parallel_samples # Transformer architecture configuration lowerCAmelCase : str = input_size * len(self.lags_sequence ) + self._number_of_features lowerCAmelCase : Any = d_model lowerCAmelCase : List[str] = encoder_attention_heads lowerCAmelCase : Union[str, Any] = decoder_attention_heads lowerCAmelCase : Optional[int] = encoder_ffn_dim lowerCAmelCase : Optional[Any] = decoder_ffn_dim lowerCAmelCase : int = encoder_layers lowerCAmelCase : int = decoder_layers lowerCAmelCase : List[Any] = dropout lowerCAmelCase : Optional[int] = attention_dropout lowerCAmelCase : Union[str, Any] = activation_dropout lowerCAmelCase : Optional[int] = encoder_layerdrop lowerCAmelCase : Dict = decoder_layerdrop lowerCAmelCase : Tuple = activation_function lowerCAmelCase : Optional[Any] = init_std lowerCAmelCase : List[Any] = use_cache # Autoformer lowerCAmelCase : Any = label_length lowerCAmelCase : Any = moving_average lowerCAmelCase : Optional[Any] = autocorrelation_factor super().__init__(is_encoder_decoder=snake_case__ , **snake_case__ ) @property def lowercase__ ( self ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
133
0
import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) _A = [ 'cross_validation.py', 'gradient_accumulation.py', 'local_sgd.py', 'multi_process_metrics.py', 'memory.py', 'automatic_gradient_accumulation.py', 'fsdp_with_peak_mem_tracking.py', 'deepspeed_with_config_support.py', 'megatron_lm_gpt_pretraining.py', ] class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def _a ( self , A_ , A_ , A_ = None , A_ = None ) -> Optional[Any]: __UpperCamelCase =None __UpperCamelCase =os.path.abspath(os.path.join('examples' , 'by_feature' ) ) __UpperCamelCase =os.path.abspath('examples' ) for item in os.listdir(A_ ): if item not in EXCLUDE_EXAMPLES: __UpperCamelCase =os.path.join(A_ , A_ ) if os.path.isfile(A_ ) and ".py" in item_path: with self.subTest( tested_script=A_ , feature_script=A_ , tested_section='main()' if parser_only else 'training_function()' , ): __UpperCamelCase =compare_against_test( os.path.join(A_ , A_ ) , A_ , A_ , A_ ) __UpperCamelCase ='\n'.join(A_ ) if special_strings is not None: for string in special_strings: __UpperCamelCase =diff.replace(A_ , '' ) self.assertEqual(A_ , '' ) def _a ( self ) -> Dict: self.one_complete_example('complete_nlp_example.py' , A_ ) self.one_complete_example('complete_nlp_example.py' , A_ ) def _a ( self ) -> Dict: __UpperCamelCase =os.path.abspath(os.path.join('examples' , 'cv_example.py' ) ) __UpperCamelCase =[ ' ' * 16 + '{\n\n', ' ' * 20 + '"accuracy": eval_metric["accuracy"],\n\n', ' ' * 20 + '"f1": eval_metric["f1"],\n\n', ' ' * 20 + '"train_loss": total_loss.item() / len(train_dataloader),\n\n', ' ' * 20 + '"epoch": epoch,\n\n', ' ' * 16 + '},\n\n', ' ' * 16 + 'step=epoch,\n', ' ' * 12, ' ' * 8 + 'for step, batch in enumerate(active_dataloader):\n', ] self.one_complete_example('complete_cv_example.py' , A_ , A_ , A_ ) self.one_complete_example('complete_cv_example.py' , A_ , A_ , A_ ) @mock.patch.dict(os.environ , {"TESTING_MOCKED_DATALOADERS": "1"} ) class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Tuple = False @classmethod def _a ( cls ) -> Union[str, Any]: super().setUpClass() __UpperCamelCase =tempfile.mkdtemp() __UpperCamelCase =os.path.join(cls._tmpdir , 'default_config.yml' ) write_basic_config(save_location=cls.configPath ) __UpperCamelCase =['accelerate', 'launch', '--config_file', cls.configPath] @classmethod def _a ( cls ) -> Union[str, Any]: super().tearDownClass() shutil.rmtree(cls._tmpdir ) def _a ( self ) -> List[Any]: __UpperCamelCase =f'\n examples/by_feature/checkpointing.py\n --checkpointing_steps epoch\n --output_dir {self.tmpdir}\n '.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'epoch_0' ) ) ) def _a ( self ) -> Union[str, Any]: __UpperCamelCase =f'\n examples/by_feature/checkpointing.py\n --checkpointing_steps 1\n --output_dir {self.tmpdir}\n '.split() __UpperCamelCase =run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'step_2' ) ) ) def _a ( self ) -> Tuple: __UpperCamelCase =f'\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir , "epoch_0" )}\n '.split() __UpperCamelCase =run_command(self._launch_args + testargs , return_stdout=A_ ) self.assertNotIn('epoch 0:' , A_ ) self.assertIn('epoch 1:' , A_ ) def _a ( self ) -> int: __UpperCamelCase =f'\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir , "step_2" )}\n '.split() __UpperCamelCase =run_command(self._launch_args + testargs , return_stdout=A_ ) if torch.cuda.is_available(): __UpperCamelCase =torch.cuda.device_count() else: __UpperCamelCase =1 if num_processes > 1: self.assertNotIn('epoch 0:' , A_ ) self.assertIn('epoch 1:' , A_ ) else: self.assertIn('epoch 0:' , A_ ) self.assertIn('epoch 1:' , A_ ) @slow def _a ( self ) -> Optional[Any]: __UpperCamelCase ='\n examples/by_feature/cross_validation.py\n --num_folds 2\n '.split() with mock.patch.dict(os.environ , {'TESTING_MOCKED_DATALOADERS': '0'} ): __UpperCamelCase =run_command(self._launch_args + testargs , return_stdout=A_ ) __UpperCamelCase =re.findall('({.+})' , A_ ) __UpperCamelCase =[r for r in results if 'accuracy' in r][-1] __UpperCamelCase =ast.literal_eval(A_ ) self.assertGreaterEqual(results['accuracy'] , 0.75 ) def _a ( self ) -> str: __UpperCamelCase =['examples/by_feature/multi_process_metrics.py'] run_command(self._launch_args + testargs ) @require_trackers @mock.patch.dict(os.environ , {'WANDB_MODE': 'offline'} ) def _a ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdir: __UpperCamelCase =f'\n examples/by_feature/tracking.py\n --with_tracking\n --project_dir {tmpdir}\n '.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(A_ , 'tracking' ) ) ) def _a ( self ) -> Optional[int]: __UpperCamelCase =['examples/by_feature/gradient_accumulation.py'] run_command(self._launch_args + testargs ) def _a ( self ) -> List[Any]: __UpperCamelCase =['examples/by_feature/local_sgd.py'] run_command(self._launch_args + testargs )
62
"""simple docstring""" import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A: List[str] = logging.get_logger(__name__) A: Dict = { "microsoft/conditional-detr-resnet-50": ( "https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json" ), } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): __lowerCAmelCase : Any = 'conditional_detr' __lowerCAmelCase : Union[str, Any] = ['past_key_values'] __lowerCAmelCase : int = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=300 , _SCREAMING_SNAKE_CASE=6 , _SCREAMING_SNAKE_CASE=2048 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=6 , _SCREAMING_SNAKE_CASE=2048 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE="relu" , _SCREAMING_SNAKE_CASE=256 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=1.0 , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE="sine" , _SCREAMING_SNAKE_CASE="resnet50" , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.25 , **_SCREAMING_SNAKE_CASE , ) -> Tuple: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) UpperCAmelCase : Tuple = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCAmelCase : str = backbone_config.get("""model_type""" ) UpperCAmelCase : int = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase : Union[str, Any] = config_class.from_dict(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Union[str, Any] = use_timm_backbone UpperCAmelCase : Optional[int] = backbone_config UpperCAmelCase : List[str] = num_channels UpperCAmelCase : Any = num_queries UpperCAmelCase : Union[str, Any] = d_model UpperCAmelCase : List[str] = encoder_ffn_dim UpperCAmelCase : Optional[int] = encoder_layers UpperCAmelCase : Union[str, Any] = encoder_attention_heads UpperCAmelCase : Optional[Any] = decoder_ffn_dim UpperCAmelCase : Any = decoder_layers UpperCAmelCase : Optional[int] = decoder_attention_heads UpperCAmelCase : Optional[int] = dropout UpperCAmelCase : Dict = attention_dropout UpperCAmelCase : Dict = activation_dropout UpperCAmelCase : Any = activation_function UpperCAmelCase : Any = init_std UpperCAmelCase : Tuple = init_xavier_std UpperCAmelCase : Optional[int] = encoder_layerdrop UpperCAmelCase : Any = decoder_layerdrop UpperCAmelCase : Any = encoder_layers UpperCAmelCase : Optional[Any] = auxiliary_loss UpperCAmelCase : List[Any] = position_embedding_type UpperCAmelCase : Union[str, Any] = backbone UpperCAmelCase : List[Any] = use_pretrained_backbone UpperCAmelCase : Dict = dilation # Hungarian matcher UpperCAmelCase : Optional[int] = class_cost UpperCAmelCase : List[str] = bbox_cost UpperCAmelCase : List[str] = giou_cost # Loss coefficients UpperCAmelCase : List[Any] = mask_loss_coefficient UpperCAmelCase : List[str] = dice_loss_coefficient UpperCAmelCase : Optional[int] = cls_loss_coefficient UpperCAmelCase : Union[str, Any] = bbox_loss_coefficient UpperCAmelCase : Union[str, Any] = giou_loss_coefficient UpperCAmelCase : int = focal_alpha super().__init__(is_encoder_decoder=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) @property def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return self.encoder_attention_heads @property def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return self.d_model def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : List[Any] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: UpperCAmelCase : Union[str, Any] = self.backbone_config.to_dict() UpperCAmelCase : Dict = self.__class__.model_type return output class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): __lowerCAmelCase : Any = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def SCREAMING_SNAKE_CASE ( self ) -> float: '''simple docstring''' return 1E-5 @property def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return 12
109
0
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets _SCREAMING_SNAKE_CASE : List[Any] = datasets.logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Tuple = """\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel's Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = \"{COMET}: A Neural Framework for {MT} Evaluation\", author = \"Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon\", booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\", month = nov, year = \"2020\", address = \"Online\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\", pages = \"2685--2702\", } """ _SCREAMING_SNAKE_CASE : Union[str, Any] = """\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA's or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. """ _SCREAMING_SNAKE_CASE : Dict = """ COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric('comet') >>> # comet_metric = load_metric('comet', 'wmt20-comet-da') # you can also choose which model to use >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"] >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"] >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results[\"scores\"]]) [0.19, 0.92] """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class __a ( datasets.Metric ): """simple docstring""" def _lowerCAmelCase ( self : Tuple ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://unbabel.github.io/COMET/html/index.html''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''sources''': datasets.Value('''string''' , id='''sequence''' ), '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/Unbabel/COMET'''] , reference_urls=[ '''https://github.com/Unbabel/COMET''', '''https://www.aclweb.org/anthology/2020.emnlp-main.213/''', '''http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6''', ] , ) def _lowerCAmelCase ( self : int , lowercase_ : Optional[int] ): if self.config_name == "default": UpperCamelCase__ : Union[str, Any] =comet.load_from_checkpoint(comet.download_model('''wmt20-comet-da''' ) ) else: UpperCamelCase__ : Optional[int] =comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def _lowerCAmelCase ( self : int , lowercase_ : List[str] , lowercase_ : List[str] , lowercase_ : Tuple , lowercase_ : str=None , lowercase_ : Union[str, Any]=False ): if gpus is None: UpperCamelCase__ : Dict =1 if torch.cuda.is_available() else 0 UpperCamelCase__ : Dict ={'''src''': sources, '''mt''': predictions, '''ref''': references} UpperCamelCase__ : Tuple =[dict(zip(lowercase_ , lowercase_ ) ) for t in zip(*data.values() )] UpperCamelCase__ , UpperCamelCase__ : str =self.scorer.predict(lowercase_ , gpus=lowercase_ , progress_bar=lowercase_ ) return {"mean_score": mean_score, "scores": scores}
157
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device _SCREAMING_SNAKE_CASE : str = False class __a ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class __a ( unittest.TestCase ): """simple docstring""" def _lowerCAmelCase ( self : List[Any] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCAmelCase ( self : Optional[int] ): UpperCamelCase__ : Any =VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) UpperCamelCase__ : int =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) UpperCamelCase__ : Dict =torch.manual_seed(0 ) UpperCamelCase__ : Optional[int] =pipe.dual_guided( prompt='''first prompt''' , image=lowercase_ , text_to_image_strength=0.7_5 , generator=lowercase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowercase_ ) UpperCamelCase__ : str =VersatileDiffusionPipeline.from_pretrained(lowercase_ , torch_dtype=torch.floataa ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) UpperCamelCase__ : int =generator.manual_seed(0 ) UpperCamelCase__ : str =pipe.dual_guided( prompt='''first prompt''' , image=lowercase_ , text_to_image_strength=0.7_5 , generator=lowercase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def _lowerCAmelCase ( self : Optional[Any] ): UpperCamelCase__ : Dict =VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) UpperCamelCase__ : str ='''cyberpunk 2077''' UpperCamelCase__ : str =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) UpperCamelCase__ : int =torch.manual_seed(0 ) UpperCamelCase__ : int =pipe.dual_guided( prompt=lowercase_ , image=lowercase_ , text_to_image_strength=0.7_5 , generator=lowercase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images UpperCamelCase__ : List[str] =image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCamelCase__ : Dict =np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCamelCase__ : Dict ='''A painting of a squirrel eating a burger ''' UpperCamelCase__ : Optional[int] =torch.manual_seed(0 ) UpperCamelCase__ : str =pipe.text_to_image( prompt=lowercase_ , generator=lowercase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' ).images UpperCamelCase__ : str =image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCamelCase__ : List[Any] =np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 UpperCamelCase__ : Optional[Any] =pipe.image_variation(lowercase_ , generator=lowercase_ , output_type='''numpy''' ).images UpperCamelCase__ : str =image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCamelCase__ : Tuple =np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
157
1
"""simple docstring""" class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : List[str] = size _lowerCamelCase : Optional[int] = [0] * size _lowerCamelCase : Any = [0] * size @staticmethod def A_ ( lowercase ): return index | (index + 1) @staticmethod def A_ ( lowercase ): return (index & (index + 1)) - 1 def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[int] = value while index < self.size: _lowerCamelCase : List[Any] = self.get_prev(__lowerCamelCase ) + 1 if current_left_border == index: _lowerCamelCase : Tuple = value else: _lowerCamelCase : List[str] = max(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) _lowerCamelCase : Tuple = self.get_next(__lowerCamelCase ) def A_ ( self , lowercase , lowercase ): right -= 1 # Because of right is exclusive _lowerCamelCase : int = 0 while left <= right: _lowerCamelCase : Union[str, Any] = self.get_prev(__lowerCamelCase ) if left <= current_left: _lowerCamelCase : List[Any] = max(__lowerCamelCase , self.tree[right] ) _lowerCamelCase : Dict = current_left else: _lowerCamelCase : Union[str, Any] = max(__lowerCamelCase , self.arr[right] ) right -= 1 return result if __name__ == "__main__": import doctest doctest.testmod()
96
import math def lowerCAmelCase__ ( lowerCamelCase_ : int): '''simple docstring''' if not isinstance(lowerCamelCase_ ,lowerCamelCase_): lowerCAmelCase__ : Union[str, Any] = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowerCamelCase_) if number < 1: lowerCAmelCase__ : Dict = f"""Input value of [number={number}] must be > 0""" raise ValueError(lowerCamelCase_) elif number == 1: return 3 elif number == 2: return 5 else: lowerCAmelCase__ : Optional[Any] = int(math.log(number // 3 ,2)) + 2 lowerCAmelCase__ : Optional[Any] = [3, 5] lowerCAmelCase__ : List[Any] = 2 lowerCAmelCase__ : Tuple = 3 for block in range(1 ,lowerCamelCase_): for _ in range(lowerCamelCase_): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1]) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(1_1): __snake_case : Optional[int] =0 try: __snake_case : List[Any] =proth(number) except ValueError: print(f"""ValueError: there is no {number}th Proth number""") continue print(f"""The {number}th Proth number: {value}""")
129
0
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class A : __snake_case = None __snake_case = False __snake_case = False __snake_case = False __snake_case = None __snake_case = None __snake_case = False __snake_case = False __snake_case = False __snake_case = True __snake_case = None __snake_case = 1 __snake_case = None __snake_case = False __snake_case = None __snake_case = None def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" return self.__class__(**{k: copy.deepcopy(_snake_case ) for k, v in self.__dict__.items()} )
357
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class A ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = tempfile.mkdtemp() # fmt: off lowerCAmelCase_ = ['''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on lowerCAmelCase_ = dict(zip(UpperCamelCase__, range(len(UpperCamelCase__ ) ) ) ) lowerCAmelCase_ = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>''', ''''''] lowerCAmelCase_ = {'''unk_token''': '''<unk>'''} lowerCAmelCase_ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase_ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file, '''w''', encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCamelCase__ ) + '''\n''' ) with open(self.merges_file, '''w''', encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCamelCase__ ) ) lowerCAmelCase_ = { '''do_resize''': True, '''size''': 20, '''do_center_crop''': True, '''crop_size''': 18, '''do_normalize''': True, '''image_mean''': [0.48_145_466, 0.4_578_275, 0.40_821_073], '''image_std''': [0.26_862_954, 0.26_130_258, 0.27_577_711], } lowerCAmelCase_ = os.path.join(self.tmpdirname, UpperCamelCase__ ) with open(self.image_processor_file, '''w''', encoding='''utf-8''' ) as fp: json.dump(UpperCamelCase__, UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self, **UpperCamelCase__ ): """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname, **UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self, **UpperCamelCase__ ): """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname, **UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self, **UpperCamelCase__ ): """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname, **UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = [np.random.randint(255, size=(3, 30, 400), dtype=np.uinta )] lowerCAmelCase_ = [Image.fromarray(np.moveaxis(UpperCamelCase__, 0, -1 ) ) for x in image_inputs] return image_inputs def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = self.get_tokenizer() lowerCAmelCase_ = self.get_rust_tokenizer() lowerCAmelCase_ = self.get_image_processor() lowerCAmelCase_ = CLIPSegProcessor(tokenizer=UpperCamelCase__, image_processor=UpperCamelCase__ ) processor_slow.save_pretrained(self.tmpdirname ) lowerCAmelCase_ = CLIPSegProcessor.from_pretrained(self.tmpdirname, use_fast=UpperCamelCase__ ) lowerCAmelCase_ = CLIPSegProcessor(tokenizer=UpperCamelCase__, image_processor=UpperCamelCase__ ) processor_fast.save_pretrained(self.tmpdirname ) lowerCAmelCase_ = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab(), tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab(), tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab(), tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer, UpperCamelCase__ ) self.assertIsInstance(processor_fast.tokenizer, UpperCamelCase__ ) self.assertEqual(processor_slow.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor, UpperCamelCase__ ) self.assertIsInstance(processor_fast.image_processor, UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = CLIPSegProcessor(tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCAmelCase_ = self.get_tokenizer(bos_token='''(BOS)''', eos_token='''(EOS)''' ) lowerCAmelCase_ = self.get_image_processor(do_normalize=UpperCamelCase__, padding_value=1.0 ) lowerCAmelCase_ = CLIPSegProcessor.from_pretrained( self.tmpdirname, bos_token='''(BOS)''', eos_token='''(EOS)''', do_normalize=UpperCamelCase__, padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, UpperCamelCase__ ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = self.get_image_processor() lowerCAmelCase_ = self.get_tokenizer() lowerCAmelCase_ = CLIPSegProcessor(tokenizer=UpperCamelCase__, image_processor=UpperCamelCase__ ) lowerCAmelCase_ = self.prepare_image_inputs() lowerCAmelCase_ = image_processor(UpperCamelCase__, return_tensors='''np''' ) lowerCAmelCase_ = processor(images=UpperCamelCase__, return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1E-2 ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = self.get_image_processor() lowerCAmelCase_ = self.get_tokenizer() lowerCAmelCase_ = CLIPSegProcessor(tokenizer=UpperCamelCase__, image_processor=UpperCamelCase__ ) lowerCAmelCase_ = '''lower newer''' lowerCAmelCase_ = processor(text=UpperCamelCase__ ) lowerCAmelCase_ = tokenizer(UpperCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key], encoded_processor[key] ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = self.get_image_processor() lowerCAmelCase_ = self.get_tokenizer() lowerCAmelCase_ = CLIPSegProcessor(tokenizer=UpperCamelCase__, image_processor=UpperCamelCase__ ) lowerCAmelCase_ = '''lower newer''' lowerCAmelCase_ = self.prepare_image_inputs() lowerCAmelCase_ = processor(text=UpperCamelCase__, images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ), ['''input_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(UpperCamelCase__ ): processor() def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = self.get_image_processor() lowerCAmelCase_ = self.get_tokenizer() lowerCAmelCase_ = CLIPSegProcessor(tokenizer=UpperCamelCase__, image_processor=UpperCamelCase__ ) lowerCAmelCase_ = self.prepare_image_inputs() lowerCAmelCase_ = self.prepare_image_inputs() lowerCAmelCase_ = processor(images=UpperCamelCase__, visual_prompt=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ), ['''pixel_values''', '''conditional_pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(UpperCamelCase__ ): processor() def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = self.get_image_processor() lowerCAmelCase_ = self.get_tokenizer() lowerCAmelCase_ = CLIPSegProcessor(tokenizer=UpperCamelCase__, image_processor=UpperCamelCase__ ) lowerCAmelCase_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCAmelCase_ = processor.batch_decode(UpperCamelCase__ ) lowerCAmelCase_ = tokenizer.batch_decode(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__, UpperCamelCase__ )
167
0
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class __a ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=99 , lowerCAmelCase__=32 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=0.0_2 , lowerCAmelCase__=4 , ) -> int: '''simple docstring''' lowercase__: List[str] = parent lowercase__: List[Any] = batch_size lowercase__: Optional[int] = seq_length lowercase__: Optional[int] = is_training lowercase__: List[str] = use_attention_mask lowercase__: List[str] = use_token_type_ids lowercase__: Any = use_labels lowercase__: Any = vocab_size lowercase__: List[Any] = hidden_size lowercase__: List[str] = num_hidden_layers lowercase__: List[Any] = num_attention_heads lowercase__: List[Any] = intermediate_size lowercase__: str = hidden_act lowercase__: Optional[Any] = hidden_dropout_prob lowercase__: Optional[Any] = attention_probs_dropout_prob lowercase__: int = max_position_embeddings lowercase__: Dict = type_vocab_size lowercase__: Tuple = type_sequence_label_size lowercase__: Tuple = initializer_range lowercase__: int = num_choices def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' lowercase__: List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__: Tuple = None if self.use_attention_mask: lowercase__: Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__: List[str] = None if self.use_token_type_ids: lowercase__: Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase__: List[str] = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' lowercase__: List[str] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ , lowercase__: Optional[Any] = config_and_inputs lowercase__: List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class __a ( _lowerCAmelCase , unittest.TestCase ): __lowercase : Union[str, Any] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' lowercase__: str = FlaxAlbertModelTester(self ) @slow def SCREAMING_SNAKE_CASE__ ( self ) -> int: '''simple docstring''' for model_class_name in self.all_model_classes: lowercase__: Dict = model_class_name.from_pretrained('albert-base-v2' ) lowercase__: List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(_lowercase ) @require_flax class __a ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[Any]: '''simple docstring''' lowercase__: Any = FlaxAlbertModel.from_pretrained('albert-base-v2' ) lowercase__: List[str] = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) lowercase__: Any = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) lowercase__: str = model(_lowercase , attention_mask=_lowercase )[0] lowercase__: Optional[Any] = (1, 11, 768) self.assertEqual(output.shape , _lowercase ) lowercase__: str = np.array( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , _lowercase , atol=1E-4 ) )
196
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _lowercase : Union[str, Any] = { 'configuration_resnet': ['RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ResNetConfig', 'ResNetOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : int = [ 'RESNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'ResNetForImageClassification', 'ResNetModel', 'ResNetPreTrainedModel', 'ResNetBackbone', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Union[str, Any] = [ 'TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFResNetForImageClassification', 'TFResNetModel', 'TFResNetPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Optional[int] = [ 'FlaxResNetForImageClassification', 'FlaxResNetModel', 'FlaxResNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys _lowercase : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure)
332
0
import math def lowerCamelCase__ ( A__ : int ): '''simple docstring''' assert isinstance(A__ , A__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False __lowerCamelCase = range(3 , int(math.sqrt(A__ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def lowerCamelCase__ ( A__ : Optional[int] , A__ : Union[str, Any]=1 , **A__ : Dict ): '''simple docstring''' __lowerCamelCase = factor * value __lowerCamelCase = value while not is_prime(A__ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **A__ ) return value
29
from __future__ import annotations UpperCAmelCase_ = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } class lowerCamelCase__: def __init__( self: Tuple , UpperCamelCase_: dict[str, list[str]] , UpperCamelCase_: str ): __lowerCamelCase = graph # mapping node to its parent in resulting breadth first tree __lowerCamelCase = {} __lowerCamelCase = source_vertex def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = {self.source_vertex} __lowerCamelCase = None __lowerCamelCase = [self.source_vertex] # first in first out queue while queue: __lowerCamelCase = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(UpperCamelCase_ ) __lowerCamelCase = vertex queue.append(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: str ): if target_vertex == self.source_vertex: return self.source_vertex __lowerCamelCase = self.parent.get(UpperCamelCase_ ) if target_vertex_parent is None: __lowerCamelCase = ( F'No path from vertex: {self.source_vertex} to vertex: {target_vertex}' ) raise ValueError(UpperCamelCase_ ) return self.shortest_path(UpperCamelCase_ ) + F'->{target_vertex}' if __name__ == "__main__": UpperCAmelCase_ = Graph(graph, 'G') g.breath_first_search() print(g.shortest_path('D')) print(g.shortest_path('G')) print(g.shortest_path('Foo'))
29
1
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def _A ( *lowercase ): """simple docstring""" with open(lowercase , '''r''' ) as fh: fcntl.flock(lowercase , fcntl.LOCK_EX ) try: print(*lowercase ) finally: fcntl.flock(lowercase , fcntl.LOCK_UN ) lowerCamelCase_ : Tuple = int(os.environ["""LOCAL_RANK"""]) torch.cuda.set_device(local_rank) lowerCamelCase_ : str = torch.device("""cuda""", local_rank) lowerCamelCase_ : List[str] = socket.gethostname() lowerCamelCase_ : List[Any] = F'[{hostname}-{local_rank}]' try: # test distributed dist.init_process_group("""nccl""") dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank lowerCamelCase_ : int = dist.get_rank() lowerCamelCase_ : int = dist.get_world_size() printflock(F'{gpu} is OK (global rank: {rank}/{world_size})') dist.barrier() if rank == 0: printflock(F'pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}') except Exception: printflock(F'{gpu} is broken') raise
81
"""simple docstring""" def _A ( lowercase , lowercase ): """simple docstring""" while second != 0: a =first & second first ^= second a =c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase_ : Dict = int(input("""Enter the first number: """).strip()) lowerCamelCase_ : List[Any] = int(input("""Enter the second number: """).strip()) print(F'{add(first, second) = }')
81
1
'''simple docstring''' import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() _A : Optional[int] =logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: lowerCamelCase__ : Union[str, Any] = SwinConfig.from_pretrained( """microsoft/swin-tiny-patch4-window7-224""" , out_features=["""stage1""", """stage2""", """stage3""", """stage4"""] ) lowerCamelCase__ : Dict = MaskFormerConfig(backbone_config=UpperCamelCase ) lowerCamelCase__ : str = """huggingface/label-files""" if "ade20k-full" in model_name: # this should be ok lowerCamelCase__ : List[Any] = 847 lowerCamelCase__ : int = """maskformer-ade20k-full-id2label.json""" elif "ade" in model_name: # this should be ok lowerCamelCase__ : Optional[Any] = 150 lowerCamelCase__ : Tuple = """ade20k-id2label.json""" elif "coco-stuff" in model_name: # this should be ok lowerCamelCase__ : Tuple = 171 lowerCamelCase__ : List[Any] = """maskformer-coco-stuff-id2label.json""" elif "coco" in model_name: # TODO lowerCamelCase__ : Optional[Any] = 133 lowerCamelCase__ : List[Any] = """coco-panoptic-id2label.json""" elif "cityscapes" in model_name: # this should be ok lowerCamelCase__ : int = 19 lowerCamelCase__ : List[Any] = """cityscapes-id2label.json""" elif "vistas" in model_name: # this should be ok lowerCamelCase__ : Tuple = 65 lowerCamelCase__ : Optional[int] = """mapillary-vistas-id2label.json""" lowerCamelCase__ : Dict = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Tuple = {int(UpperCamelCase ): v for k, v in idalabel.items()} return config def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> List[str]: lowerCamelCase__ : Dict = [] # stem # fmt: off rename_keys.append(("""backbone.patch_embed.proj.weight""", """model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight""") ) rename_keys.append(("""backbone.patch_embed.proj.bias""", """model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias""") ) rename_keys.append(("""backbone.patch_embed.norm.weight""", """model.pixel_level_module.encoder.model.embeddings.norm.weight""") ) rename_keys.append(("""backbone.patch_embed.norm.bias""", """model.pixel_level_module.encoder.model.embeddings.norm.bias""") ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.norm1.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.norm1.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.attn.relative_position_index''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.attn.proj.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.attn.proj.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.norm2.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.norm2.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.mlp.fc1.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.mlp.fc1.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.mlp.fc2.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.mlp.fc2.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((f'''backbone.layers.{i}.downsample.reduction.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((f'''backbone.layers.{i}.downsample.norm.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((f'''backbone.layers.{i}.downsample.norm.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append((f'''backbone.norm{i}.weight''', f'''model.pixel_level_module.encoder.hidden_states_norms.{i}.weight''') ) rename_keys.append((f'''backbone.norm{i}.bias''', f'''model.pixel_level_module.encoder.hidden_states_norms.{i}.bias''') ) # FPN rename_keys.append(("""sem_seg_head.layer_4.weight""", """model.pixel_level_module.decoder.fpn.stem.0.weight""") ) rename_keys.append(("""sem_seg_head.layer_4.norm.weight""", """model.pixel_level_module.decoder.fpn.stem.1.weight""") ) rename_keys.append(("""sem_seg_head.layer_4.norm.bias""", """model.pixel_level_module.decoder.fpn.stem.1.bias""") ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((f'''sem_seg_head.adapter_{source_index}.weight''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight''') ) rename_keys.append((f'''sem_seg_head.adapter_{source_index}.norm.weight''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight''') ) rename_keys.append((f'''sem_seg_head.adapter_{source_index}.norm.bias''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias''') ) rename_keys.append((f'''sem_seg_head.layer_{source_index}.weight''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight''') ) rename_keys.append((f'''sem_seg_head.layer_{source_index}.norm.weight''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight''') ) rename_keys.append((f'''sem_seg_head.layer_{source_index}.norm.bias''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias''') ) rename_keys.append(("""sem_seg_head.mask_features.weight""", """model.pixel_level_module.decoder.mask_projection.weight""") ) rename_keys.append(("""sem_seg_head.mask_features.bias""", """model.pixel_level_module.decoder.mask_projection.bias""") ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight''', f'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias''', f'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias''') ) # cross-attention out projection rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight''', f'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias''', f'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias''') ) # MLP 1 rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight''', f'''model.transformer_module.decoder.layers.{idx}.fc1.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias''', f'''model.transformer_module.decoder.layers.{idx}.fc1.bias''') ) # MLP 2 rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight''', f'''model.transformer_module.decoder.layers.{idx}.fc2.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias''', f'''model.transformer_module.decoder.layers.{idx}.fc2.bias''') ) # layernorm 1 (self-attention layernorm) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight''', f'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias''', f'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias''') ) # layernorm 2 (cross-attention layernorm) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight''', f'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias''', f'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias''') ) # layernorm 3 (final layernorm) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight''', f'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias''', f'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias''') ) rename_keys.append(("""sem_seg_head.predictor.transformer.decoder.norm.weight""", """model.transformer_module.decoder.layernorm.weight""") ) rename_keys.append(("""sem_seg_head.predictor.transformer.decoder.norm.bias""", """model.transformer_module.decoder.layernorm.bias""") ) # heads on top rename_keys.append(("""sem_seg_head.predictor.query_embed.weight""", """model.transformer_module.queries_embedder.weight""") ) rename_keys.append(("""sem_seg_head.predictor.input_proj.weight""", """model.transformer_module.input_projection.weight""") ) rename_keys.append(("""sem_seg_head.predictor.input_proj.bias""", """model.transformer_module.input_projection.bias""") ) rename_keys.append(("""sem_seg_head.predictor.class_embed.weight""", """class_predictor.weight""") ) rename_keys.append(("""sem_seg_head.predictor.class_embed.bias""", """class_predictor.bias""") ) for i in range(3 ): rename_keys.append((f'''sem_seg_head.predictor.mask_embed.layers.{i}.weight''', f'''mask_embedder.{i}.0.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.mask_embed.layers.{i}.bias''', f'''mask_embedder.{i}.0.bias''') ) # fmt: on return rename_keys def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Any: lowerCamelCase__ : Tuple = dct.pop(UpperCamelCase ) lowerCamelCase__ : Optional[int] = val def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Any = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): lowerCamelCase__ : str = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) lowerCamelCase__ : str = state_dict.pop(f'''backbone.layers.{i}.blocks.{j}.attn.qkv.weight''' ) lowerCamelCase__ : Any = state_dict.pop(f'''backbone.layers.{i}.blocks.{j}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Union[str, Any] = in_proj_weight[:dim, :] lowerCamelCase__ : int = in_proj_bias[: dim] lowerCamelCase__ : int = in_proj_weight[ dim : dim * 2, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[ dim : dim * 2 ] lowerCamelCase__ : Optional[Any] = in_proj_weight[ -dim :, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[-dim :] # fmt: on def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> str: # fmt: off lowerCamelCase__ : List[Any] = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) lowerCamelCase__ : List[str] = state_dict.pop(f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight''' ) lowerCamelCase__ : Optional[Any] = state_dict.pop(f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Tuple = in_proj_weight[: hidden_size, :] lowerCamelCase__ : Dict = in_proj_bias[:config.hidden_size] lowerCamelCase__ : Any = in_proj_weight[hidden_size : hidden_size * 2, :] lowerCamelCase__ : Any = in_proj_bias[hidden_size : hidden_size * 2] lowerCamelCase__ : Any = in_proj_weight[-hidden_size :, :] lowerCamelCase__ : int = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) lowerCamelCase__ : str = state_dict.pop(f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight''' ) lowerCamelCase__ : int = state_dict.pop(f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Any = in_proj_weight[: hidden_size, :] lowerCamelCase__ : Optional[Any] = in_proj_bias[:config.hidden_size] lowerCamelCase__ : str = in_proj_weight[hidden_size : hidden_size * 2, :] lowerCamelCase__ : List[str] = in_proj_bias[hidden_size : hidden_size * 2] lowerCamelCase__ : List[str] = in_proj_weight[-hidden_size :, :] lowerCamelCase__ : Optional[Any] = in_proj_bias[-hidden_size :] # fmt: on def SCREAMING_SNAKE_CASE_ () -> torch.Tensor: lowerCamelCase__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Union[str, Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = False ) -> int: lowerCamelCase__ : List[str] = get_maskformer_config(UpperCamelCase ) # load original state_dict with open(UpperCamelCase , """rb""" ) as f: lowerCamelCase__ : int = pickle.load(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = data["""model"""] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys lowerCamelCase__ : Any = create_rename_keys(UpperCamelCase ) for src, dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) read_in_swin_q_k_v(UpperCamelCase , config.backbone_config ) read_in_decoder_q_k_v(UpperCamelCase , UpperCamelCase ) # update to torch tensors for key, value in state_dict.items(): lowerCamelCase__ : Optional[Any] = torch.from_numpy(UpperCamelCase ) # load 🤗 model lowerCamelCase__ : str = MaskFormerForInstanceSegmentation(UpperCamelCase ) model.eval() for name, param in model.named_parameters(): print(UpperCamelCase , param.shape ) lowerCamelCase__ , lowerCamelCase__ : Any = model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(UpperCamelCase ) == 0, f'''Unexpected keys: {unexpected_keys}''' # verify results lowerCamelCase__ : Optional[Any] = prepare_img() if "vistas" in model_name: lowerCamelCase__ : Tuple = 65 elif "cityscapes" in model_name: lowerCamelCase__ : List[Any] = 65535 else: lowerCamelCase__ : str = 255 lowerCamelCase__ : Any = True if """ade""" in model_name else False lowerCamelCase__ : Optional[int] = MaskFormerImageProcessor(ignore_index=UpperCamelCase , reduce_labels=UpperCamelCase ) lowerCamelCase__ : str = image_processor(UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Any = model(**UpperCamelCase ) print("""Logits:""" , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": lowerCamelCase__ : int = torch.tensor( [[3.6353, -4.4770, -2.6065], [0.5081, -4.2394, -3.5343], [2.1909, -5.0353, -1.9323]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCamelCase , atol=1E-4 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(f'''Saving model and image processor to {pytorch_dump_folder_path}''' ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) image_processor.save_pretrained(UpperCamelCase ) if push_to_hub: print("""Pushing model and image processor to the hub...""" ) model.push_to_hub(f'''nielsr/{model_name}''' ) image_processor.push_to_hub(f'''nielsr/{model_name}''' ) if __name__ == "__main__": _A : List[Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''maskformer-swin-tiny-ade''', type=str, help=('''Name of the MaskFormer model you\'d like to convert''',), ) parser.add_argument( '''--checkpoint_path''', default='''/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl''', type=str, help='''Path to the original state dict (.pth file).''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _A : Dict =parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
129
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _A : List[str] =logging.get_logger(__name__) class _lowercase ( _lowercase ): a = ["""input_features""", """attention_mask"""] def __init__( self: Optional[int] , UpperCamelCase__: str=80 , UpperCamelCase__: Union[str, Any]=16_000 , UpperCamelCase__: str=0.0 , UpperCamelCase__: Tuple=10 , UpperCamelCase__: Dict=25 , UpperCamelCase__: List[Any]="hamming_window" , UpperCamelCase__: Tuple=32_768.0 , UpperCamelCase__: List[Any]=0.97 , UpperCamelCase__: Optional[int]=1.0 , UpperCamelCase__: str=True , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Optional[int]=False , **UpperCamelCase__: Optional[int] , ): super().__init__(feature_size=UpperCamelCase__ , sampling_rate=UpperCamelCase__ , padding_value=UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase__ : Dict = feature_size lowerCamelCase__ : Tuple = sampling_rate lowerCamelCase__ : Union[str, Any] = padding_value lowerCamelCase__ : Union[str, Any] = hop_length lowerCamelCase__ : Any = win_length lowerCamelCase__ : Tuple = frame_signal_scale lowerCamelCase__ : Dict = preemphasis_coeff lowerCamelCase__ : Optional[Any] = mel_floor lowerCamelCase__ : Any = normalize_means lowerCamelCase__ : Optional[int] = normalize_vars lowerCamelCase__ : Dict = win_function lowerCamelCase__ : Dict = return_attention_mask lowerCamelCase__ : Optional[Any] = win_length * sampling_rate // 1_000 lowerCamelCase__ : Tuple = hop_length * sampling_rate // 1_000 lowerCamelCase__ : List[Any] = optimal_fft_length(self.sample_size ) lowerCamelCase__ : List[str] = (self.n_fft // 2) + 1 def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: np.array ): if self.win_function == "hamming_window": lowerCamelCase__ : Dict = window_function(window_length=self.sample_size , name=self.win_function , periodic=UpperCamelCase__ ) else: lowerCamelCase__ : Union[str, Any] = window_function(window_length=self.sample_size , name=self.win_function ) lowerCamelCase__ : List[Any] = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) lowerCamelCase__ : Tuple = spectrogram( one_waveform * self.frame_signal_scale , window=UpperCamelCase__ , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=UpperCamelCase__ , preemphasis=self.preemphasis_coeff , mel_filters=UpperCamelCase__ , mel_floor=self.mel_floor , log_mel="""log""" , ) return msfc_features.T def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: List[str] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): # make sure we normalize float32 arrays if self.normalize_means: lowerCamelCase__ : List[str] = x[:input_length].mean(axis=0 ) lowerCamelCase__ : List[str] = np.subtract(UpperCamelCase__ , UpperCamelCase__ ) if self.normalize_vars: lowerCamelCase__ : Union[str, Any] = x[:input_length].std(axis=0 ) lowerCamelCase__ : str = np.divide(UpperCamelCase__ , UpperCamelCase__ ) if input_length < x.shape[0]: lowerCamelCase__ : List[Any] = padding_value # make sure array is in float32 lowerCamelCase__ : str = x.astype(np.floataa ) return x def lowerCamelCase_ ( self: str , UpperCamelCase__: List[np.ndarray] , UpperCamelCase__: Optional[np.ndarray] = None ): lowerCamelCase__ : Optional[int] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(UpperCamelCase__ , UpperCamelCase__ , self.padding_value ) for x, n in zip(UpperCamelCase__ , UpperCamelCase__ )] def __call__( self: List[str] , UpperCamelCase__: Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase__: Union[bool, str, PaddingStrategy] = False , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: bool = False , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: Optional[bool] = None , UpperCamelCase__: Optional[Union[str, TensorType]] = None , UpperCamelCase__: Optional[int] = None , **UpperCamelCase__: Tuple , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) lowerCamelCase__ : List[str] = isinstance(UpperCamelCase__ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) lowerCamelCase__ : List[str] = is_batched_numpy or ( isinstance(UpperCamelCase__ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCamelCase__ : str = [np.asarray(UpperCamelCase__ , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase__ , np.ndarray ): lowerCamelCase__ : List[Any] = np.asarray(UpperCamelCase__ , dtype=np.floataa ) elif isinstance(UpperCamelCase__ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCamelCase__ : Optional[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCamelCase__ : Optional[Any] = [raw_speech] # extract fbank features lowerCamelCase__ : str = [self._extract_mfsc_features(UpperCamelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCamelCase__ : str = BatchFeature({"""input_features""": features} ) lowerCamelCase__ : str = self.pad( UpperCamelCase__ , padding=UpperCamelCase__ , max_length=UpperCamelCase__ , truncation=UpperCamelCase__ , pad_to_multiple_of=UpperCamelCase__ , return_attention_mask=UpperCamelCase__ , **UpperCamelCase__ , ) # make sure list is in array format lowerCamelCase__ : str = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] , UpperCamelCase__ ): lowerCamelCase__ : Dict = [np.asarray(UpperCamelCase__ , dtype=np.floataa ) for feature in input_features] lowerCamelCase__ : List[str] = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: lowerCamelCase__ : Any = [np.asarray(UpperCamelCase__ , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCamelCase__ : str = ( np.array(UpperCamelCase__ , dtype=np.intaa ) if self._get_padding_strategies(UpperCamelCase__ , max_length=UpperCamelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCamelCase__ : Union[str, Any] = self.normalize( padded_inputs["""input_features"""] , attention_mask=UpperCamelCase__ ) if return_tensors is not None: lowerCamelCase__ : Dict = padded_inputs.convert_to_tensors(UpperCamelCase__ ) return padded_inputs
129
1
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm lowerCAmelCase__ : int =logging.get_logger(__name__) @dataclass class UpperCAmelCase_ ( _lowercase ): '''simple docstring''' UpperCamelCase__ : Dict = [ '''no_inference''', '''no_cuda''', '''no_tpu''', '''no_speed''', '''no_memory''', '''no_env_print''', '''no_multi_process''', ] def __init__( self , **_A ): '''simple docstring''' for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __SCREAMING_SNAKE_CASE = deprecated_arg[3:] setattr(self , __a , not kwargs.pop(__a ) ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) __SCREAMING_SNAKE_CASE = kwargs.pop('torchscript' , self.torchscript ) __SCREAMING_SNAKE_CASE = kwargs.pop('torch_xla_tpu_print_metrics' , self.torch_xla_tpu_print_metrics ) __SCREAMING_SNAKE_CASE = kwargs.pop('fp16_opt_level' , self.fpaa_opt_level ) super().__init__(**__a ) UpperCamelCase__ : Any = field(default=_lowercase , metadata={'''help''': '''Trace the models using torchscript'''} ) UpperCamelCase__ : Optional[int] = field(default=_lowercase , metadata={'''help''': '''Print Xla/PyTorch tpu metrics'''} ) UpperCamelCase__ : Dict = field( default='''O1''' , metadata={ '''help''': ( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ''' '''See details at https://nvidia.github.io/apex/amp.html''' ) } , ) @cached_property def _A ( self ): '''simple docstring''' requires_backends(self , ['torch'] ) logger.info('PyTorch: setting up devices' ) if not self.cuda: __SCREAMING_SNAKE_CASE = torch.device('cpu' ) __SCREAMING_SNAKE_CASE = 0 elif is_torch_tpu_available(): __SCREAMING_SNAKE_CASE = xm.xla_device() __SCREAMING_SNAKE_CASE = 0 else: __SCREAMING_SNAKE_CASE = torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) __SCREAMING_SNAKE_CASE = torch.cuda.device_count() return device, n_gpu @property def _A ( self ): '''simple docstring''' return is_torch_tpu_available() and self.tpu @property def _A ( self ): '''simple docstring''' requires_backends(self , ['torch'] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def _A ( self ): '''simple docstring''' requires_backends(self , ['torch'] ) return self._setup_devices[0] @property def _A ( self ): '''simple docstring''' requires_backends(self , ['torch'] ) return self._setup_devices[1] @property def _A ( self ): '''simple docstring''' return self.n_gpu > 0
257
"""simple docstring""" import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets lowerCAmelCase__ = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' lowerCAmelCase__ = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' lowerCAmelCase__ = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCamelCase ( datasets.Metric ): def snake_case_ (self ) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/google-research/tree/master/rouge"] , reference_urls=[ "https://en.wikipedia.org/wiki/ROUGE_(metric)", "https://github.com/google-research/google-research/tree/master/rouge", ] , ) def snake_case_ (self , __a , __a , __a=None , __a=True , __a=False ) -> Union[str, Any]: if rouge_types is None: UpperCamelCase = ["rouge1", "rouge2", "rougeL", "rougeLsum"] UpperCamelCase = rouge_scorer.RougeScorer(rouge_types=__a , use_stemmer=__a ) if use_aggregator: UpperCamelCase = scoring.BootstrapAggregator() else: UpperCamelCase = [] for ref, pred in zip(__a , __a ): UpperCamelCase = scorer.score(__a , __a ) if use_aggregator: aggregator.add_scores(__a ) else: scores.append(__a ) if use_aggregator: UpperCamelCase = aggregator.aggregate() else: UpperCamelCase = {} for key in scores[0]: UpperCamelCase = [score[key] for score in scores] return result
153
0
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf UpperCAmelCase : str = logging.get_logger(__name__) @dataclass class __lowerCAmelCase ( UpperCamelCase__): _lowercase : Dict = [ """no_inference""", """no_cuda""", """no_tpu""", """no_speed""", """no_memory""", """no_env_print""", """no_multi_process""", ] def __init__( self , **lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: a__ : List[str] =deprecated_arg[3:] a__ : Tuple =not kwargs.pop(lowerCAmelCase__ ) logger.warning( F'''{deprecated_arg} is depreciated. Please use --no-{positive_arg} or''' F''' {positive_arg}={kwargs[positive_arg]}''' ) a__ : str =kwargs.pop("tpu_name" , self.tpu_name ) a__ : Any =kwargs.pop("device_idx" , self.device_idx ) a__ : Tuple =kwargs.pop("eager_mode" , self.eager_mode ) a__ : List[Any] =kwargs.pop("use_xla" , self.use_xla ) super().__init__(**lowerCAmelCase__ ) _lowercase : str = field( default=UpperCamelCase__ , metadata={"""help""": """Name of TPU"""} , ) _lowercase : int = field( default=0 , metadata={"""help""": """CPU / GPU device index. Defaults to 0."""} , ) _lowercase : bool = field(default=UpperCamelCase__ , metadata={"""help""": """Benchmark models in eager model."""}) _lowercase : bool = field( default=UpperCamelCase__ , metadata={ """help""": """Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.""" } , ) @cached_property def _lowercase ( self ) -> Tuple["tf.distribute.cluster_resolver.TPUClusterResolver"]: '''simple docstring''' requires_backends(self , ["tf"] ) a__ : Optional[int] =None if self.tpu: try: if self.tpu_name: a__ : List[Any] =tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: a__ : Dict =tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: a__ : Optional[int] =None return tpu @cached_property def _lowercase ( self ) -> Tuple["tf.distribute.Strategy", "tf.distribute.cluster_resolver.TPUClusterResolver"]: '''simple docstring''' requires_backends(self , ["tf"] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) a__ : Any =tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , "GPU" ) a__ : Dict =tf.distribute.OneDeviceStrategy(device=F'''/gpu:{self.device_idx}''' ) else: tf.config.set_visible_devices([] , "GPU" ) # disable GPU a__ : List[str] =tf.distribute.OneDeviceStrategy(device=F'''/cpu:{self.device_idx}''' ) return strategy @property def _lowercase ( self ) -> bool: '''simple docstring''' requires_backends(self , ["tf"] ) return self._setup_tpu is not None @property def _lowercase ( self ) -> "tf.distribute.Strategy": '''simple docstring''' requires_backends(self , ["tf"] ) return self._setup_strategy @property def _lowercase ( self ) -> str: '''simple docstring''' requires_backends(self , ["tf"] ) return tf.config.list_physical_devices("GPU" ) @property def _lowercase ( self ) -> int: '''simple docstring''' requires_backends(self , ["tf"] ) if self.cuda: return len(self.gpu_list ) return 0 @property def _lowercase ( self ) -> bool: '''simple docstring''' return self.n_gpu > 0
148
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Dict = logging.get_logger(__name__) UpperCAmelCase : Dict = { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json""", } class __lowerCAmelCase ( UpperCamelCase__): _lowercase : List[Any] = """mvp""" _lowercase : Tuple = ["""past_key_values"""] _lowercase : Dict = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowerCAmelCase__=5_0_2_6_7 , lowerCAmelCase__=1_0_2_4 , lowerCAmelCase__=1_2 , lowerCAmelCase__=4_0_9_6 , lowerCAmelCase__=1_6 , lowerCAmelCase__=1_2 , lowerCAmelCase__=4_0_9_6 , lowerCAmelCase__=1_6 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__="gelu" , lowerCAmelCase__=1_0_2_4 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.02 , lowerCAmelCase__=0.0 , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=1 , lowerCAmelCase__=0 , lowerCAmelCase__=2 , lowerCAmelCase__=True , lowerCAmelCase__=2 , lowerCAmelCase__=2 , lowerCAmelCase__=False , lowerCAmelCase__=1_0_0 , lowerCAmelCase__=8_0_0 , **lowerCAmelCase__ , ) -> str: '''simple docstring''' a__ : Dict =vocab_size a__ : List[str] =max_position_embeddings a__ : List[str] =d_model a__ : Optional[Any] =encoder_ffn_dim a__ : Dict =encoder_layers a__ : List[str] =encoder_attention_heads a__ : List[str] =decoder_ffn_dim a__ : Tuple =decoder_layers a__ : Tuple =decoder_attention_heads a__ : Any =dropout a__ : str =attention_dropout a__ : str =activation_dropout a__ : Optional[int] =activation_function a__ : Union[str, Any] =init_std a__ : Dict =encoder_layerdrop a__ : List[str] =decoder_layerdrop a__ : Union[str, Any] =classifier_dropout a__ : Union[str, Any] =use_cache a__ : Optional[Any] =encoder_layers a__ : List[str] =scale_embedding # scale factor will be sqrt(d_model) if True a__ : int =use_prompt a__ : Union[str, Any] =prompt_length a__ : Dict =prompt_mid_dim super().__init__( pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , is_encoder_decoder=lowerCAmelCase__ , decoder_start_token_id=lowerCAmelCase__ , forced_eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ , ) if self.forced_bos_token_id is None and kwargs.get("force_bos_token_to_be_generated" , lowerCAmelCase__ ): a__ : Tuple =self.bos_token_id warnings.warn( F'''Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ''' "The config can simply be saved and uploaded again to be fixed." )
148
1
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() _A = { 'bart': ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'bert': ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-base-cased-finetuned-mrpc': ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'dpr': ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'gpt2': ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlnet': ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm': ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm-roberta': ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'transfo-xl': ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'openai-gpt': ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'roberta': ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'layoutlm': ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'roberta-large-mnli': ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'camembert': ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'flaubert': ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert': ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert-base-distilled-squad': ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert': ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert-visual-feature-encoder': ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'ctrl': ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'albert': ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 't5': ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'electra': ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'wav2vec2': ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : str=False , SCREAMING_SNAKE_CASE__ : Optional[int]=True ): if model_type not in MODEL_CLASSES: raise ValueError(F'Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: __UpperCamelCase =cached_file(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , force_download=not use_cached_models ) __UpperCamelCase =config_class.from_json_file(SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =True __UpperCamelCase =True print(F'Building TensorFlow model from configuration: {config}' ) __UpperCamelCase =model_class(SCREAMING_SNAKE_CASE__ ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): __UpperCamelCase =cached_file( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: __UpperCamelCase =load_pytorch_checkpoint_in_tfa_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if compare_with_pt_model: __UpperCamelCase =tf_model(tf_model.dummy_inputs , training=SCREAMING_SNAKE_CASE__ ) # build the network __UpperCamelCase =torch.load(SCREAMING_SNAKE_CASE__ , map_location='cpu' ) __UpperCamelCase =pt_model_class.from_pretrained( pretrained_model_name_or_path=SCREAMING_SNAKE_CASE__ , config=SCREAMING_SNAKE_CASE__ , state_dict=SCREAMING_SNAKE_CASE__ ) with torch.no_grad(): __UpperCamelCase =pt_model(**pt_model.dummy_inputs ) __UpperCamelCase =pto[0].numpy() __UpperCamelCase =tfo[0].numpy() __UpperCamelCase =np.amax(np.abs(np_pt - np_tf ) ) print(F'Max absolute difference between models outputs {diff}' ) assert diff <= 2E-2, F'Error, model absolute difference is >2e-2: {diff}' # Save pytorch-model print(F'Save TensorFlow model to {tf_dump_path}' ) tf_model.save_weights(SCREAMING_SNAKE_CASE__ , save_format='h5' ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[str]=None , SCREAMING_SNAKE_CASE__ : Dict=None , SCREAMING_SNAKE_CASE__ : Dict=False , SCREAMING_SNAKE_CASE__ : Any=False , SCREAMING_SNAKE_CASE__ : Dict=False , SCREAMING_SNAKE_CASE__ : str=False , ): if args_model_type is None: __UpperCamelCase =list(MODEL_CLASSES.keys() ) else: __UpperCamelCase =[args_model_type] for j, model_type in enumerate(SCREAMING_SNAKE_CASE__ , start=1 ): print('=' * 1_00 ) print(F' Converting model type {j}/{len(SCREAMING_SNAKE_CASE__ )}: {model_type}' ) print('=' * 1_00 ) if model_type not in MODEL_CLASSES: raise ValueError(F'Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: __UpperCamelCase =list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: __UpperCamelCase =model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , start=1 ): print('-' * 1_00 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(F' Skipping finetuned checkpoint {model_shortcut_name}' ) continue __UpperCamelCase =model_shortcut_name elif only_convert_finetuned_models: print(F' Skipping not finetuned checkpoint {model_shortcut_name}' ) continue print( F' Converting checkpoint {i}/{len(SCREAMING_SNAKE_CASE__ )}: {model_shortcut_name} - model_type {model_type}' ) print('-' * 1_00 ) if config_shortcut_name in aws_config_map: __UpperCamelCase =cached_file(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , force_download=not use_cached_models ) else: __UpperCamelCase =config_shortcut_name if model_shortcut_name in aws_model_maps: __UpperCamelCase =cached_file(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , force_download=not use_cached_models ) else: __UpperCamelCase =model_shortcut_name if os.path.isfile(SCREAMING_SNAKE_CASE__ ): __UpperCamelCase ='converted_model' convert_pt_checkpoint_to_tf( model_type=SCREAMING_SNAKE_CASE__ , pytorch_checkpoint_path=SCREAMING_SNAKE_CASE__ , config_file=SCREAMING_SNAKE_CASE__ , tf_dump_path=os.path.join(SCREAMING_SNAKE_CASE__ , model_shortcut_name + '-tf_model.h5' ) , compare_with_pt_model=SCREAMING_SNAKE_CASE__ , ) if remove_cached_files: os.remove(SCREAMING_SNAKE_CASE__ ) os.remove(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": _A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_dump_path', default=None, type=str, required=True, help='Path to the output Tensorflow dump file.' ) parser.add_argument( '--model_type', default=None, type=str, help=( f"""Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and """ 'convert all the models from AWS.' ), ) parser.add_argument( '--pytorch_checkpoint_path', default=None, type=str, help=( 'Path to the PyTorch checkpoint path or shortcut name to download from AWS. ' 'If not given, will download and convert all the checkpoints from AWS.' ), ) parser.add_argument( '--config_file', default=None, type=str, help=( 'The config json file corresponding to the pre-trained model. \n' 'This specifies the model architecture. If not given and ' '--pytorch_checkpoint_path is not given or is a shortcut name ' 'use the configuration associated to the shortcut name on the AWS' ), ) parser.add_argument( '--compare_with_pt_model', action='store_true', help='Compare Tensorflow and PyTorch model predictions.' ) parser.add_argument( '--use_cached_models', action='store_true', help='Use cached models if possible instead of updating to latest checkpoint versions.', ) parser.add_argument( '--remove_cached_files', action='store_true', help='Remove pytorch models after conversion (save memory when converting in batches).', ) parser.add_argument('--only_convert_finetuned_models', action='store_true', help='Only convert finetuned models.') _A = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
62
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class A__ : """simple docstring""" def __init__( self : Any , lowerCAmelCase__ : Any ) -> Dict: """simple docstring""" _UpperCAmelCase : Union[str, Any] = str(id_ ) _UpperCAmelCase : Union[str, Any] = None _UpperCAmelCase : Dict = None _UpperCAmelCase : Tuple = [] _UpperCAmelCase : int = {} # {vertex:distance} def __lt__( self : List[str] , lowerCAmelCase__ : str ) -> List[str]: """simple docstring""" return self.key < other.key def __repr__( self : int ) -> Any: """simple docstring""" return self.id def _lowerCAmelCase ( self : Union[str, Any] , lowerCAmelCase__ : Optional[Any] ) -> Any: """simple docstring""" self.neighbors.append(lowerCAmelCase__ ) def _lowerCAmelCase ( self : Any , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Optional[int] ) -> Tuple: """simple docstring""" _UpperCAmelCase : Optional[Any] = weight def __UpperCAmelCase ( a_: Any, a_: Optional[Any], a_: Optional[Any], a_: List[str] ): # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1], a_ ) graph[b - 1].add_edge(graph[a - 1], a_ ) def __UpperCAmelCase ( a_: list, a_: Vertex ): _UpperCAmelCase : Optional[int] = [] for u in graph: _UpperCAmelCase : Dict = math.inf _UpperCAmelCase : Any = None _UpperCAmelCase : List[str] = 0 _UpperCAmelCase : Union[str, Any] = graph[:] while q: _UpperCAmelCase : List[Any] = min(a_ ) q.remove(a_ ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): _UpperCAmelCase : Optional[Any] = u _UpperCAmelCase : List[Any] = u.edges[v.id] for i in range(1, len(a_ ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def __UpperCAmelCase ( a_: list, a_: Vertex ): for u in graph: _UpperCAmelCase : Optional[Any] = math.inf _UpperCAmelCase : str = None _UpperCAmelCase : Optional[Any] = 0 _UpperCAmelCase : List[str] = list(a_ ) hq.heapify(a_ ) while h: _UpperCAmelCase : str = hq.heappop(a_ ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): _UpperCAmelCase : Any = u _UpperCAmelCase : Optional[int] = u.edges[v.id] hq.heapify(a_ ) for i in range(1, len(a_ ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def __UpperCAmelCase ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
145
0
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter a__ = '''Create a default config file for Accelerate with only a few flags set.''' def __UpperCAmelCase ( __a : Optional[int]="no" ,__a : str = default_json_config_file ,__a : bool = False ) -> Any: """simple docstring""" _a : List[Any] = Path(__a ) path.parent.mkdir(parents=__a ,exist_ok=__a ) if path.exists(): print( F"""Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.""" ) return False _a : List[Any] = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( F"""`mixed_precision` should be one of 'no', 'fp16', 'bf16', or 'fp8'. Received {mixed_precision}""" ) _a : Optional[int] = { '''compute_environment''': '''LOCAL_MACHINE''', '''mixed_precision''': mixed_precision, } if torch.cuda.is_available(): _a : Any = torch.cuda.device_count() _a : List[Any] = num_gpus _a : Dict = False if num_gpus > 1: _a : str = '''MULTI_GPU''' else: _a : Union[str, Any] = '''NO''' elif is_xpu_available() and use_xpu: _a : str = torch.xpu.device_count() _a : List[str] = num_xpus _a : Union[str, Any] = False if num_xpus > 1: _a : Tuple = '''MULTI_XPU''' else: _a : Tuple = '''NO''' elif is_npu_available(): _a : Dict = torch.npu.device_count() _a : Dict = num_npus _a : Optional[int] = False if num_npus > 1: _a : List[str] = '''MULTI_NPU''' else: _a : int = '''NO''' else: _a : str = 0 _a : Optional[Any] = True _a : int = 1 _a : int = '''NO''' _a : Tuple = ClusterConfig(**__a ) config.to_json_file(__a ) return path def __UpperCAmelCase ( __a : Tuple ,__a : Any ) -> List[Any]: """simple docstring""" _a : List[str] = parser.add_parser('''default''' ,parents=__a ,help=__a ,formatter_class=__a ) parser.add_argument( '''--config_file''' ,default=__a ,help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) ,dest='''save_location''' ,) parser.add_argument( '''--mixed_precision''' ,choices=['''no''', '''fp16''', '''bf16'''] ,type=__a ,help='''Whether or not to use mixed precision training. ''' '''Choose between FP16 and BF16 (bfloat16) training. ''' '''BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later.''' ,default='''no''' ,) parser.set_defaults(func=__a ) return parser def __UpperCAmelCase ( __a : Dict ) -> List[Any]: """simple docstring""" _a : List[Any] = write_basic_config(args.mixed_precision ,args.save_location ) if config_file: print(F"""accelerate configuration saved at {config_file}""" )
364
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ = { '''configuration_xmod''': [ '''XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XmodConfig''', '''XmodOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ = [ '''XMOD_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XmodForCausalLM''', '''XmodForMaskedLM''', '''XmodForMultipleChoice''', '''XmodForQuestionAnswering''', '''XmodForSequenceClassification''', '''XmodForTokenClassification''', '''XmodModel''', '''XmodPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys a__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
15
0
"""simple docstring""" import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow __UpperCamelCase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ '''text-classification''', '''language-modeling''', '''summarization''', '''token-classification''', '''question-answering''', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) __UpperCamelCase = logging.getLogger() def lowercase () -> int: SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument('-f' ) SCREAMING_SNAKE_CASE = parser.parse_args() return args.f def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Optional[Any]="eval" ) -> Tuple: SCREAMING_SNAKE_CASE = os.path.join(SCREAMING_SNAKE_CASE_ , F'{split}_results.json' ) if os.path.exists(SCREAMING_SNAKE_CASE_ ): with open(SCREAMING_SNAKE_CASE_ , 'r' ) as f: return json.load(SCREAMING_SNAKE_CASE_ ) raise ValueError(F'can\'t find {path}' ) __UpperCamelCase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' def __A ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE = F'\n run_glue.py\n --model_name_or_path distilbert-base-uncased\n --output_dir {tmp_dir}\n --train_file ./tests/fixtures/tests_samples/MRPC/train.csv\n --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --learning_rate=1e-4\n --eval_steps=2\n --warmup_steps=2\n --seed=42\n --max_seq_length=128\n '.split() with patch.object(lowerCAmelCase__ , 'argv' , lowerCAmelCase__ ): run_flax_glue.main() SCREAMING_SNAKE_CASE = get_results(lowerCAmelCase__ ) self.assertGreaterEqual(result['eval_accuracy'] , 0.75 ) @slow def __A ( self ) -> Dict: SCREAMING_SNAKE_CASE = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE = F'\n run_clm_flax.py\n --model_name_or_path distilgpt2\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --do_train\n --do_eval\n --block_size 128\n --per_device_train_batch_size 4\n --per_device_eval_batch_size 4\n --num_train_epochs 2\n --logging_steps 2 --eval_steps 2\n --output_dir {tmp_dir}\n --overwrite_output_dir\n '.split() with patch.object(lowerCAmelCase__ , 'argv' , lowerCAmelCase__ ): run_clm_flax.main() SCREAMING_SNAKE_CASE = get_results(lowerCAmelCase__ ) self.assertLess(result['eval_perplexity'] , 100 ) @slow def __A ( self ) -> Optional[Any]: SCREAMING_SNAKE_CASE = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE = F'\n run_summarization.py\n --model_name_or_path t5-small\n --train_file tests/fixtures/tests_samples/xsum/sample.json\n --validation_file tests/fixtures/tests_samples/xsum/sample.json\n --test_file tests/fixtures/tests_samples/xsum/sample.json\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --num_train_epochs=3\n --warmup_steps=8\n --do_train\n --do_eval\n --do_predict\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --predict_with_generate\n '.split() with patch.object(lowerCAmelCase__ , 'argv' , lowerCAmelCase__ ): run_summarization_flax.main() SCREAMING_SNAKE_CASE = get_results(lowerCAmelCase__ , split='test' ) self.assertGreaterEqual(result['test_rouge1'] , 10 ) self.assertGreaterEqual(result['test_rouge2'] , 2 ) self.assertGreaterEqual(result['test_rougeL'] , 7 ) self.assertGreaterEqual(result['test_rougeLsum'] , 7 ) @slow def __A ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE = F'\n run_mlm.py\n --model_name_or_path distilroberta-base\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --max_seq_length 128\n --per_device_train_batch_size 4\n --per_device_eval_batch_size 4\n --logging_steps 2 --eval_steps 2\n --do_train\n --do_eval\n --num_train_epochs=1\n '.split() with patch.object(lowerCAmelCase__ , 'argv' , lowerCAmelCase__ ): run_mlm_flax.main() SCREAMING_SNAKE_CASE = get_results(lowerCAmelCase__ ) self.assertLess(result['eval_perplexity'] , 42 ) @slow def __A ( self ) -> str: SCREAMING_SNAKE_CASE = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE = F'\n run_t5_mlm_flax.py\n --model_name_or_path t5-small\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --do_train\n --do_eval\n --max_seq_length 128\n --per_device_train_batch_size 4\n --per_device_eval_batch_size 4\n --num_train_epochs 2\n --logging_steps 2 --eval_steps 2\n --output_dir {tmp_dir}\n --overwrite_output_dir\n '.split() with patch.object(lowerCAmelCase__ , 'argv' , lowerCAmelCase__ ): run_ta_mlm_flax.main() SCREAMING_SNAKE_CASE = get_results(lowerCAmelCase__ ) self.assertGreaterEqual(result['eval_accuracy'] , 0.42 ) @slow def __A ( self ) -> Dict: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu SCREAMING_SNAKE_CASE = 7 if get_gpu_count() > 1 else 2 SCREAMING_SNAKE_CASE = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE = F'\n run_flax_ner.py\n --model_name_or_path bert-base-uncased\n --train_file tests/fixtures/tests_samples/conll/sample.json\n --validation_file tests/fixtures/tests_samples/conll/sample.json\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --do_train\n --do_eval\n --warmup_steps=2\n --learning_rate=2e-4\n --logging_steps 2 --eval_steps 2\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=2\n --num_train_epochs={epochs}\n --seed 7\n '.split() with patch.object(lowerCAmelCase__ , 'argv' , lowerCAmelCase__ ): run_flax_ner.main() SCREAMING_SNAKE_CASE = get_results(lowerCAmelCase__ ) self.assertGreaterEqual(result['eval_accuracy'] , 0.75 ) self.assertGreaterEqual(result['eval_f1'] , 0.3 ) @slow def __A ( self ) -> str: SCREAMING_SNAKE_CASE = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE = F'\n run_qa.py\n --model_name_or_path bert-base-uncased\n --version_2_with_negative\n --train_file tests/fixtures/tests_samples/SQUAD/sample.json\n --validation_file tests/fixtures/tests_samples/SQUAD/sample.json\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --num_train_epochs=3\n --warmup_steps=2\n --do_train\n --do_eval\n --logging_steps 2 --eval_steps 2\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n '.split() with patch.object(lowerCAmelCase__ , 'argv' , lowerCAmelCase__ ): run_qa.main() SCREAMING_SNAKE_CASE = get_results(lowerCAmelCase__ ) self.assertGreaterEqual(result['eval_f1'] , 30 ) self.assertGreaterEqual(result['eval_exact'] , 30 )
113
"""simple docstring""" import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) __UpperCamelCase = [ '''cross_validation.py''', '''gradient_accumulation.py''', '''local_sgd.py''', '''multi_process_metrics.py''', '''memory.py''', '''automatic_gradient_accumulation.py''', '''fsdp_with_peak_mem_tracking.py''', '''deepspeed_with_config_support.py''', '''megatron_lm_gpt_pretraining.py''', ] class lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None ) -> int: SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = os.path.abspath(os.path.join('examples' , 'by_feature' ) ) SCREAMING_SNAKE_CASE = os.path.abspath('examples' ) for item in os.listdir(lowerCAmelCase__ ): if item not in EXCLUDE_EXAMPLES: SCREAMING_SNAKE_CASE = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) if os.path.isfile(lowerCAmelCase__ ) and ".py" in item_path: with self.subTest( tested_script=lowerCAmelCase__ , feature_script=lowerCAmelCase__ , tested_section='main()' if parser_only else 'training_function()' , ): SCREAMING_SNAKE_CASE = compare_against_test( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = '\n'.join(lowerCAmelCase__ ) if special_strings is not None: for string in special_strings: SCREAMING_SNAKE_CASE = diff.replace(lowerCAmelCase__ , '' ) self.assertEqual(lowerCAmelCase__ , '' ) def __A ( self ) -> Optional[int]: self.one_complete_example('complete_nlp_example.py' , lowerCAmelCase__ ) self.one_complete_example('complete_nlp_example.py' , lowerCAmelCase__ ) def __A ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE = os.path.abspath(os.path.join('examples' , 'cv_example.py' ) ) SCREAMING_SNAKE_CASE = [ ' ' * 16 + '{\n\n', ' ' * 20 + '"accuracy": eval_metric["accuracy"],\n\n', ' ' * 20 + '"f1": eval_metric["f1"],\n\n', ' ' * 20 + '"train_loss": total_loss.item() / len(train_dataloader),\n\n', ' ' * 20 + '"epoch": epoch,\n\n', ' ' * 16 + '},\n\n', ' ' * 16 + 'step=epoch,\n', ' ' * 12, ' ' * 8 + 'for step, batch in enumerate(active_dataloader):\n', ] self.one_complete_example('complete_cv_example.py' , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) self.one_complete_example('complete_cv_example.py' , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) @mock.patch.dict(os.environ , {"""TESTING_MOCKED_DATALOADERS""": """1"""} ) class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = False @classmethod def __A ( cls ) -> List[str]: super().setUpClass() SCREAMING_SNAKE_CASE = tempfile.mkdtemp() SCREAMING_SNAKE_CASE = os.path.join(cls._tmpdir , 'default_config.yml' ) write_basic_config(save_location=cls.configPath ) SCREAMING_SNAKE_CASE = ['accelerate', 'launch', '--config_file', cls.configPath] @classmethod def __A ( cls ) -> Dict: super().tearDownClass() shutil.rmtree(cls._tmpdir ) def __A ( self ) -> int: SCREAMING_SNAKE_CASE = F'\n examples/by_feature/checkpointing.py\n --checkpointing_steps epoch\n --output_dir {self.tmpdir}\n '.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'epoch_0' ) ) ) def __A ( self ) -> List[Any]: SCREAMING_SNAKE_CASE = F'\n examples/by_feature/checkpointing.py\n --checkpointing_steps 1\n --output_dir {self.tmpdir}\n '.split() SCREAMING_SNAKE_CASE = run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'step_2' ) ) ) def __A ( self ) -> Union[str, Any]: SCREAMING_SNAKE_CASE = F'\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir , "epoch_0" )}\n '.split() SCREAMING_SNAKE_CASE = run_command(self._launch_args + testargs , return_stdout=lowerCAmelCase__ ) self.assertNotIn('epoch 0:' , lowerCAmelCase__ ) self.assertIn('epoch 1:' , lowerCAmelCase__ ) def __A ( self ) -> int: SCREAMING_SNAKE_CASE = F'\n examples/by_feature/checkpointing.py\n --resume_from_checkpoint {os.path.join(self.tmpdir , "step_2" )}\n '.split() SCREAMING_SNAKE_CASE = run_command(self._launch_args + testargs , return_stdout=lowerCAmelCase__ ) if torch.cuda.is_available(): SCREAMING_SNAKE_CASE = torch.cuda.device_count() else: SCREAMING_SNAKE_CASE = 1 if num_processes > 1: self.assertNotIn('epoch 0:' , lowerCAmelCase__ ) self.assertIn('epoch 1:' , lowerCAmelCase__ ) else: self.assertIn('epoch 0:' , lowerCAmelCase__ ) self.assertIn('epoch 1:' , lowerCAmelCase__ ) @slow def __A ( self ) -> Optional[Any]: SCREAMING_SNAKE_CASE = '\n examples/by_feature/cross_validation.py\n --num_folds 2\n '.split() with mock.patch.dict(os.environ , {'TESTING_MOCKED_DATALOADERS': '0'} ): SCREAMING_SNAKE_CASE = run_command(self._launch_args + testargs , return_stdout=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = re.findall('({.+})' , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE = [r for r in results if 'accuracy' in r][-1] SCREAMING_SNAKE_CASE = ast.literal_eval(lowerCAmelCase__ ) self.assertGreaterEqual(results['accuracy'] , 0.75 ) def __A ( self ) -> str: SCREAMING_SNAKE_CASE = ['examples/by_feature/multi_process_metrics.py'] run_command(self._launch_args + testargs ) @require_trackers @mock.patch.dict(os.environ , {'WANDB_MODE': 'offline'} ) def __A ( self ) -> Union[str, Any]: with tempfile.TemporaryDirectory() as tmpdir: SCREAMING_SNAKE_CASE = F'\n examples/by_feature/tracking.py\n --with_tracking\n --project_dir {tmpdir}\n '.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(lowerCAmelCase__ , 'tracking' ) ) ) def __A ( self ) -> Dict: SCREAMING_SNAKE_CASE = ['examples/by_feature/gradient_accumulation.py'] run_command(self._launch_args + testargs ) def __A ( self ) -> List[Any]: SCREAMING_SNAKE_CASE = ['examples/by_feature/local_sgd.py'] run_command(self._launch_args + testargs )
113
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase :Any = logging.get_logger(__name__) _lowerCAmelCase :Dict = { 'andreasmadsen/efficient_mlm_m0.40': ( 'https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json' ), } class _UpperCAmelCase ( a ): '''simple docstring''' a__ ='''roberta-prelayernorm''' def __init__( self , A=5_0_2_6_5 , A=7_6_8 , A=1_2 , A=1_2 , A=3_0_7_2 , A="gelu" , A=0.1 , A=0.1 , A=5_1_2 , A=2 , A=0.02 , A=1E-12 , A=1 , A=0 , A=2 , A="absolute" , A=True , A=None , **A , ) -> Optional[int]: super().__init__(pad_token_id=A , bos_token_id=A , eos_token_id=A , **A ) _UpperCAmelCase : Any = vocab_size _UpperCAmelCase : Tuple = hidden_size _UpperCAmelCase : List[Any] = num_hidden_layers _UpperCAmelCase : Optional[int] = num_attention_heads _UpperCAmelCase : str = hidden_act _UpperCAmelCase : Tuple = intermediate_size _UpperCAmelCase : Optional[Any] = hidden_dropout_prob _UpperCAmelCase : Optional[int] = attention_probs_dropout_prob _UpperCAmelCase : Union[str, Any] = max_position_embeddings _UpperCAmelCase : str = type_vocab_size _UpperCAmelCase : str = initializer_range _UpperCAmelCase : Union[str, Any] = layer_norm_eps _UpperCAmelCase : List[str] = position_embedding_type _UpperCAmelCase : int = use_cache _UpperCAmelCase : Optional[int] = classifier_dropout class _UpperCAmelCase ( a ): '''simple docstring''' @property def __lowerCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _UpperCAmelCase : List[Any] = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _UpperCAmelCase : Dict = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
68
"""simple docstring""" import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures _lowerCAmelCase :str = logging.get_logger(__name__) @dataclass class _UpperCAmelCase : '''simple docstring''' a__ =field(metadata={'''help''': '''The name of the task to train on: ''' + ''', '''.join(glue_processors.keys() )} ) a__ =field( metadata={'''help''': '''The input data dir. Should contain the .tsv files (or other data files) for the task.'''} ) a__ =field( default=1_2_8 ,metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } ,) a__ =field( default=a ,metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def __lowerCAmelCase ( self ) -> int: _UpperCAmelCase : str = self.task_name.lower() class _UpperCAmelCase ( a ): '''simple docstring''' a__ ='''train''' a__ ='''dev''' a__ ='''test''' class _UpperCAmelCase ( a ): '''simple docstring''' a__ =42 a__ =42 a__ =42 def __init__( self , A , A , A = None , A = Split.train , A = None , ) -> Dict: warnings.warn( '''This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets ''' '''library. You can have a look at this example script for pointers: ''' '''https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py''' , A , ) _UpperCAmelCase : Dict = args _UpperCAmelCase : int = glue_processors[args.task_name]() _UpperCAmelCase : Any = glue_output_modes[args.task_name] if isinstance(A , A ): try: _UpperCAmelCase : int = Split[mode] except KeyError: raise KeyError('''mode is not a valid split name''' ) # Load data features from cache or dataset file _UpperCAmelCase : Optional[int] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}' , ) _UpperCAmelCase : List[Any] = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = label_list[2], label_list[1] _UpperCAmelCase : Optional[Any] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _UpperCAmelCase : List[Any] = cached_features_file + '''.lock''' with FileLock(A ): if os.path.exists(A ) and not args.overwrite_cache: _UpperCAmelCase : str = time.time() _UpperCAmelCase : Dict = torch.load(A ) logger.info( f'Loading features from cached file {cached_features_file} [took %.3f s]' , time.time() - start ) else: logger.info(f'Creating features from dataset file at {args.data_dir}' ) if mode == Split.dev: _UpperCAmelCase : Union[str, Any] = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: _UpperCAmelCase : List[Any] = self.processor.get_test_examples(args.data_dir ) else: _UpperCAmelCase : Optional[Any] = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: _UpperCAmelCase : Tuple = examples[:limit_length] _UpperCAmelCase : str = glue_convert_examples_to_features( A , A , max_length=args.max_seq_length , label_list=A , output_mode=self.output_mode , ) _UpperCAmelCase : Optional[int] = time.time() torch.save(self.features , A ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]' ) def __len__( self ) -> Union[str, Any]: return len(self.features ) def __getitem__( self , A ) -> InputFeatures: return self.features[i] def __lowerCAmelCase ( self ) -> List[Any]: return self.label_list
68
1
from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class __lowercase (UpperCamelCase__ ): """simple docstring""" def __init__( self , A , A ) -> str: super().__init__() # make sure scheduler can always be converted to DDIM snake_case : str = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=A , scheduler=A ) @torch.no_grad() def __call__( self , A = 1 , A = None , A = 0.0 , A = 5_0 , A = None , A = "pil" , A = True , ) -> Union[ImagePipelineOutput, Tuple]: # Sample gaussian noise to begin loop if isinstance(self.unet.config.sample_size , A ): snake_case : str = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: snake_case : List[Any] = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(A , A ) and len(A ) != batch_size: raise ValueError( f"""You have passed a list of generators of length {len(A )}, but requested an effective batch""" f""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) snake_case : Dict = randn_tensor(A , generator=A , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(A ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output snake_case : Tuple = self.unet(A , A ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 snake_case : Dict = self.scheduler.step( A , A , A , eta=A , use_clipped_model_output=A , generator=A ).prev_sample snake_case : Optional[int] = (image / 2 + 0.5).clamp(0 , 1 ) snake_case : Any = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case : Optional[Any] = self.numpy_to_pil(A ) if not return_dict: return (image,) return ImagePipelineOutput(images=A )
124
from __future__ import annotations import math lowerCamelCase : Optional[int] = '2020.9.26' lowerCamelCase : int = 'xcodz-dot, cclaus, dhruvmanila' def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase ,lowercase ) -> tuple[float, float]: if not all(isinstance(lowercase ,(float, int) ) for val in locals().values() ): snake_case : Dict = f"""Input values must either be float or int: {list(locals().values() )}""" raise TypeError(lowercase ) snake_case : List[str] = ((x * distance) / (z + distance)) * scale snake_case : Dict = ((y * distance) / (z + distance)) * scale return projected_x, projected_y def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase ,lowercase ) -> tuple[float, float, float]: if not isinstance(lowercase ,lowercase ): raise TypeError("""Axis must be a str""" ) snake_case : Tuple = locals() del input_variables["axis"] if not all(isinstance(lowercase ,(float, int) ) for val in input_variables.values() ): snake_case : int = ( """Input values except axis must either be float or int: """ f"""{list(input_variables.values() )}""" ) raise TypeError(lowercase ) snake_case : int = (angle % 360) / 450 * 180 / math.pi if axis == "z": snake_case : str = x * math.cos(lowercase ) - y * math.sin(lowercase ) snake_case : List[Any] = y * math.cos(lowercase ) + x * math.sin(lowercase ) snake_case : Optional[int] = z elif axis == "x": snake_case : Optional[Any] = y * math.cos(lowercase ) - z * math.sin(lowercase ) snake_case : Optional[int] = z * math.cos(lowercase ) + y * math.sin(lowercase ) snake_case : Optional[int] = x elif axis == "y": snake_case : List[str] = x * math.cos(lowercase ) - z * math.sin(lowercase ) snake_case : Tuple = z * math.cos(lowercase ) + x * math.sin(lowercase ) snake_case : Optional[int] = y else: raise ValueError("""not a valid axis, choose one of 'x', 'y', 'z'""" ) return new_x, new_y, new_z if __name__ == "__main__": import doctest doctest.testmod() print(f"""{convert_to_ad(1.0, 2.0, 3.0, 10.0, 10.0) = }""") print(f"""{rotate(1.0, 2.0, 3.0, "y", 90.0) = }""")
124
1
"""simple docstring""" import math import random from typing import Any from .hill_climbing import SearchProblem def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase = True , lowerCAmelCase = math.inf , lowerCAmelCase = -math.inf , lowerCAmelCase = math.inf , lowerCAmelCase = -math.inf , lowerCAmelCase = False , lowerCAmelCase = 100 , lowerCAmelCase = 0.01 , lowerCAmelCase = 1 , ): '''simple docstring''' UpperCAmelCase = False UpperCAmelCase = search_prob UpperCAmelCase = start_temperate UpperCAmelCase = [] UpperCAmelCase = 0 UpperCAmelCase = None while not search_end: UpperCAmelCase = current_state.score() if best_state is None or current_score > best_state.score(): UpperCAmelCase = current_state scores.append(lowerCAmelCase ) iterations += 1 UpperCAmelCase = None UpperCAmelCase = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to UpperCAmelCase = random.randint(0 , len(lowerCAmelCase ) - 1 ) # picking a random neighbor UpperCAmelCase = neighbors.pop(lowerCAmelCase ) UpperCAmelCase = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: UpperCAmelCase = change * -1 # in case we are finding minimum if change > 0: # improves the solution UpperCAmelCase = picked_neighbor else: UpperCAmelCase = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability UpperCAmelCase = picked_neighbor UpperCAmelCase = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor UpperCAmelCase = True else: UpperCAmelCase = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(lowerCAmelCase ) , lowerCAmelCase ) plt.xlabel("""Iterations""" ) plt.ylabel("""Function values""" ) plt.show() return best_state if __name__ == "__main__": def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) lowerCAmelCase_ : List[str] = SearchProblem(x=1_2, y=4_7, step_size=1, function_to_optimize=test_fa) lowerCAmelCase_ : List[str] = simulated_annealing( prob, find_max=False, max_x=1_0_0, min_x=5, max_y=5_0, min_y=-5, visualization=True ) print( '''The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F'and 50 > y > - 5 found via hill climbing: {local_min.score()}' ) # starting the problem with initial coordinates (12, 47) lowerCAmelCase_ : int = SearchProblem(x=1_2, y=4_7, step_size=1, function_to_optimize=test_fa) lowerCAmelCase_ : Optional[Any] = simulated_annealing( prob, find_max=True, max_x=1_0_0, min_x=5, max_y=5_0, min_y=-5, visualization=True ) print( '''The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F'and 50 > y > - 5 found via hill climbing: {local_min.score()}' ) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' return (3 * x**2) - (6 * y) lowerCAmelCase_ : Dict = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase_ : List[Any] = simulated_annealing(prob, find_max=False, visualization=True) print( '''The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F'{local_min.score()}' ) lowerCAmelCase_ : List[str] = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase_ : List[Any] = simulated_annealing(prob, find_max=True, visualization=True) print( '''The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F'{local_min.score()}' )
248
"""simple docstring""" import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline lowerCAmelCase_ : Tuple = version.parse(version.parse(torch.__version__).base_version) < version.parse('''1.11''') def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False , ): '''simple docstring''' output_path.parent.mkdir(parents=lowerCAmelCase , exist_ok=lowerCAmelCase ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( lowerCAmelCase , lowerCAmelCase , f=output_path.as_posix() , input_names=lowerCAmelCase , output_names=lowerCAmelCase , dynamic_axes=lowerCAmelCase , do_constant_folding=lowerCAmelCase , use_external_data_format=lowerCAmelCase , enable_onnx_checker=lowerCAmelCase , opset_version=lowerCAmelCase , ) else: export( lowerCAmelCase , lowerCAmelCase , f=output_path.as_posix() , input_names=lowerCAmelCase , output_names=lowerCAmelCase , dynamic_axes=lowerCAmelCase , do_constant_folding=lowerCAmelCase , opset_version=lowerCAmelCase , ) @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = False ): '''simple docstring''' UpperCAmelCase = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): UpperCAmelCase = """cuda""" elif fpaa and not torch.cuda.is_available(): raise ValueError("""`float16` model export is only supported on GPUs with CUDA""" ) else: UpperCAmelCase = """cpu""" UpperCAmelCase = StableDiffusionPipeline.from_pretrained(lowerCAmelCase , torch_dtype=lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase = Path(lowerCAmelCase ) # TEXT ENCODER UpperCAmelCase = pipeline.text_encoder.config.max_position_embeddings UpperCAmelCase = pipeline.text_encoder.config.hidden_size UpperCAmelCase = pipeline.tokenizer( """A sample prompt""" , padding="""max_length""" , max_length=pipeline.tokenizer.model_max_length , truncation=lowerCAmelCase , return_tensors="""pt""" , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=lowerCAmelCase , dtype=torch.intaa )) , output_path=output_path / """text_encoder""" / """model.onnx""" , ordered_input_names=["""input_ids"""] , output_names=["""last_hidden_state""", """pooler_output"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """sequence"""}, } , opset=lowerCAmelCase , ) del pipeline.text_encoder # UNET UpperCAmelCase = pipeline.unet.config.in_channels UpperCAmelCase = pipeline.unet.config.sample_size UpperCAmelCase = output_path / """unet""" / """model.onnx""" onnx_export( pipeline.unet , model_args=( torch.randn(2 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), torch.randn(2 ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), torch.randn(2 , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), False, ) , output_path=lowerCAmelCase , ordered_input_names=["""sample""", """timestep""", """encoder_hidden_states""", """return_dict"""] , output_names=["""out_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """timestep""": {0: """batch"""}, """encoder_hidden_states""": {0: """batch""", 1: """sequence"""}, } , opset=lowerCAmelCase , use_external_data_format=lowerCAmelCase , ) UpperCAmelCase = str(unet_path.absolute().as_posix() ) UpperCAmelCase = os.path.dirname(lowerCAmelCase ) UpperCAmelCase = onnx.load(lowerCAmelCase ) # clean up existing tensor files shutil.rmtree(lowerCAmelCase ) os.mkdir(lowerCAmelCase ) # collate external tensor files into one onnx.save_model( lowerCAmelCase , lowerCAmelCase , save_as_external_data=lowerCAmelCase , all_tensors_to_one_file=lowerCAmelCase , location="""weights.pb""" , convert_attribute=lowerCAmelCase , ) del pipeline.unet # VAE ENCODER UpperCAmelCase = pipeline.vae UpperCAmelCase = vae_encoder.config.in_channels UpperCAmelCase = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder UpperCAmelCase = lambda lowerCAmelCase , lowerCAmelCase : vae_encoder.encode(lowerCAmelCase , lowerCAmelCase )[0].sample() onnx_export( lowerCAmelCase , model_args=( torch.randn(1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), False, ) , output_path=output_path / """vae_encoder""" / """model.onnx""" , ordered_input_names=["""sample""", """return_dict"""] , output_names=["""latent_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase , ) # VAE DECODER UpperCAmelCase = pipeline.vae UpperCAmelCase = vae_decoder.config.latent_channels UpperCAmelCase = vae_decoder.config.out_channels # forward only through the decoder part UpperCAmelCase = vae_encoder.decode onnx_export( lowerCAmelCase , model_args=( torch.randn(1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), False, ) , output_path=output_path / """vae_decoder""" / """model.onnx""" , ordered_input_names=["""latent_sample""", """return_dict"""] , output_names=["""sample"""] , dynamic_axes={ """latent_sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: UpperCAmelCase = pipeline.safety_checker UpperCAmelCase = safety_checker.config.vision_config.num_channels UpperCAmelCase = safety_checker.config.vision_config.image_size UpperCAmelCase = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), torch.randn(1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ).to(device=lowerCAmelCase , dtype=lowerCAmelCase ), ) , output_path=output_path / """safety_checker""" / """model.onnx""" , ordered_input_names=["""clip_input""", """images"""] , output_names=["""out_images""", """has_nsfw_concepts"""] , dynamic_axes={ """clip_input""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """images""": {0: """batch""", 1: """height""", 2: """width""", 3: """channels"""}, } , opset=lowerCAmelCase , ) del pipeline.safety_checker UpperCAmelCase = OnnxRuntimeModel.from_pretrained(output_path / """safety_checker""" ) UpperCAmelCase = pipeline.feature_extractor else: UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_encoder""" ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_decoder""" ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / """text_encoder""" ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / """unet""" ) , scheduler=pipeline.scheduler , safety_checker=lowerCAmelCase , feature_extractor=lowerCAmelCase , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(lowerCAmelCase ) print("""ONNX pipeline saved to""" , lowerCAmelCase ) del pipeline del onnx_pipeline UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(lowerCAmelCase , provider="""CPUExecutionProvider""" ) print("""ONNX pipeline is loadable""" ) if __name__ == "__main__": lowerCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model_path''', type=str, required=True, help='''Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).''', ) parser.add_argument('''--output_path''', type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--opset''', default=1_4, type=int, help='''The version of the ONNX operator set to use.''', ) parser.add_argument('''--fp16''', action='''store_true''', default=False, help='''Export the models in `float16` mode''') lowerCAmelCase_ : Union[str, Any] = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
248
1
import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process UpperCAmelCase_ : Dict = logging.getLogger(__name__) UpperCAmelCase_ : Optional[Any] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) UpperCAmelCase_ : str = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class _SCREAMING_SNAKE_CASE : snake_case__ : Optional[str] = field( default=_a , metadata={ """help""": ( """The model checkpoint for weights initialization. Leave None if you want to train a model from""" """ scratch.""" ) } , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """If training from scratch, pass a model type from the list: """ + """, """.join(_a )} , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class _SCREAMING_SNAKE_CASE : snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """The input training data file (a text file)."""} ) snake_case__ : Optional[str] = field( default=_a , metadata={ """help""": ( """The input training data files (multiple files in glob format). """ """Very often splitting large files to smaller files can prevent tokenizer going out of memory""" ) } , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """An optional input train ref data file for whole word mask in Chinese."""} , ) snake_case__ : Optional[str] = field( default=_a , metadata={"""help""": """An optional input eval ref data file for whole word mask in Chinese."""} , ) snake_case__ : bool = field( default=_a , metadata={"""help""": """Whether distinct lines of text in the dataset are to be handled as distinct sequences."""} , ) snake_case__ : bool = field( default=_a , metadata={"""help""": """Train with masked-language modeling loss instead of language modeling."""} ) snake_case__ : bool = field(default=_a , metadata={"""help""": """Whether ot not to use whole word mask."""} ) snake_case__ : float = field( default=0.15 , metadata={"""help""": """Ratio of tokens to mask for masked language modeling loss"""} ) snake_case__ : float = field( default=1 / 6 , metadata={ """help""": ( """Ratio of length of a span of masked tokens to surrounding context length for permutation language""" """ modeling.""" ) } , ) snake_case__ : int = field( default=5 , metadata={"""help""": """Maximum length of a span of masked tokens for permutation language modeling."""} ) snake_case__ : int = field( default=-1 , metadata={ """help""": ( """Optional input sequence length after tokenization.""" """The training dataset will be truncated in block of this size for training.""" """Default to the model max input length for single sentence inputs (take into account special tokens).""" ) } , ) snake_case__ : bool = field( default=_a , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : DataTrainingArguments , __magic_name__ : PreTrainedTokenizer , __magic_name__ : bool = False , __magic_name__ : Optional[str] = None , ) -> List[Any]: """simple docstring""" def _dataset(__magic_name__ : Union[str, Any] , __magic_name__ : Optional[Any]=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError("""You need to set world whole masking and mlm to True for Chinese Whole Word Mask""" ) return LineByLineWithRefDataset( tokenizer=__magic_name__ , file_path=__magic_name__ , block_size=args.block_size , ref_path=__magic_name__ , ) return LineByLineTextDataset(tokenizer=__magic_name__ , file_path=__magic_name__ , block_size=args.block_size ) else: return TextDataset( tokenizer=__magic_name__ , file_path=__magic_name__ , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=__magic_name__ , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(__magic_name__ ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def SCREAMING_SNAKE_CASE_ ( ) -> Optional[int]: """simple docstring""" UpperCamelCase :List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCamelCase , UpperCamelCase , UpperCamelCase :Dict = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( """Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file """ """or remove the --do_eval argument.""" ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" """ --overwrite_output_dir to overcome.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" , __magic_name__ ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: UpperCamelCase :Tuple = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCamelCase :Optional[Any] = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: UpperCamelCase :Dict = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.tokenizer_name: UpperCamelCase :Tuple = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCamelCase :int = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( """You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another""" """ script, save it,and load it from here, using --tokenizer_name""" ) if model_args.model_name_or_path: UpperCamelCase :Dict = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=__magic_name__ , cache_dir=model_args.cache_dir , ) else: logger.info("""Training new model from scratch""" ) UpperCamelCase :Any = AutoModelWithLMHead.from_config(__magic_name__ ) model.resize_token_embeddings(len(__magic_name__ ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( """BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the""" """--mlm flag (masked language modeling).""" ) if data_args.block_size <= 0: UpperCamelCase :Dict = tokenizer.max_len # Our input block size will be the max possible for the model else: UpperCamelCase :int = min(data_args.block_size , tokenizer.max_len ) # Get datasets UpperCamelCase :Optional[int] = ( get_dataset(__magic_name__ , tokenizer=__magic_name__ , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) UpperCamelCase :Any = ( get_dataset(__magic_name__ , tokenizer=__magic_name__ , evaluate=__magic_name__ , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": UpperCamelCase :int = DataCollatorForPermutationLanguageModeling( tokenizer=__magic_name__ , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: UpperCamelCase :str = DataCollatorForWholeWordMask( tokenizer=__magic_name__ , mlm_probability=data_args.mlm_probability ) else: UpperCamelCase :Any = DataCollatorForLanguageModeling( tokenizer=__magic_name__ , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer UpperCamelCase :List[Any] = Trainer( model=__magic_name__ , args=__magic_name__ , data_collator=__magic_name__ , train_dataset=__magic_name__ , eval_dataset=__magic_name__ , prediction_loss_only=__magic_name__ , ) # Training if training_args.do_train: UpperCamelCase :Optional[Any] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=__magic_name__ ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCamelCase :Optional[int] = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) UpperCamelCase :List[Any] = trainer.evaluate() UpperCamelCase :List[Any] = math.exp(eval_output["""eval_loss"""] ) UpperCamelCase :Optional[Any] = {"""perplexity""": perplexity} UpperCamelCase :Tuple = os.path.join(training_args.output_dir , """eval_results_lm.txt""" ) if trainer.is_world_master(): with open(__magic_name__ , """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key in sorted(result.keys() ): logger.info(""" %s = %s""" , __magic_name__ , str(result[key] ) ) writer.write("""%s = %s\n""" % (key, str(result[key] )) ) results.update(__magic_name__ ) return results def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Union[str, Any] ) -> int: """simple docstring""" main() if __name__ == "__main__": main()
38
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
38
1
"""simple docstring""" def __lowerCamelCase ( a_ : int , a_ : int ) -> int: while b: __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :Optional[int] = b, a % b return a def __lowerCamelCase ( a_ : int , a_ : int ) -> int: return a if b == 0 else euclidean_gcd_recursive(a_ , a % b ) def __lowerCamelCase ( ) -> List[str]: print(f'''euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}''' ) print(f'''euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}''' ) print(f'''euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}''' ) print(f'''euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}''' ) print(f'''euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}''' ) print(f'''euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}''' ) print(f'''euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}''' ) print(f'''euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}''' ) print(f'''euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}''' ) print(f'''euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}''' ) if __name__ == "__main__": main()
239
"""simple docstring""" import math import random def __lowerCamelCase ( a_ : float , a_ : bool = False ) -> float: if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value lowerCamelCase_ = 0.02 def __lowerCamelCase ( a_ : int , a_ : int ) -> float: __SCREAMING_SNAKE_CASE :Any = float(2 * (random.randint(1 , 1_00 )) - 1 ) for _ in range(a_ ): # Forward propagation __SCREAMING_SNAKE_CASE :Any = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? __SCREAMING_SNAKE_CASE :Tuple = (expected / 1_00) - layer_a # Error delta __SCREAMING_SNAKE_CASE :Union[str, Any] = layer_1_error * sigmoid_function(a_ , a_ ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 1_00 if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase_ = int(input("Expected value: ")) lowerCamelCase_ = int(input("Number of propagations: ")) print(forward_propagation(expected, number_propagations))
239
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase = logging.get_logger(__name__) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False ): """simple docstring""" lowercase__ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'deit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'deit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'deit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'deit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'deit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'deit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'deit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'deit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'deit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'deit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''deit.embeddings.cls_token'''), ('''dist_token''', '''deit.embeddings.distillation_token'''), ('''patch_embed.proj.weight''', '''deit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''deit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''deit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" lowercase__ = [(pair[0], pair[1][4:]) if pair[1].startswith('''deit''' ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ('''norm.weight''', '''deit.layernorm.weight'''), ('''norm.bias''', '''deit.layernorm.bias'''), ('''head.weight''', '''cls_classifier.weight'''), ('''head.bias''', '''cls_classifier.bias'''), ('''head_dist.weight''', '''distillation_classifier.weight'''), ('''head_dist.bias''', '''distillation_classifier.bias'''), ] ) return rename_keys def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: lowercase__ = '''''' else: lowercase__ = '''deit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowercase__ = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) lowercase__ = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[ : config.hidden_size, : ] lowercase__ = in_proj_bias[: config.hidden_size] lowercase__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowercase__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowercase__ = in_proj_weight[ -config.hidden_size :, : ] lowercase__ = in_proj_bias[-config.hidden_size :] def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = dct.pop(__UpperCAmelCase ) lowercase__ = val def _a ( ): """simple docstring""" lowercase__ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowercase__ = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ) return im @torch.no_grad() def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = DeiTConfig() # all deit models have fine-tuned heads lowercase__ = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size lowercase__ = 10_00 lowercase__ = '''huggingface/label-files''' lowercase__ = '''imagenet-1k-id2label.json''' lowercase__ = json.load(open(hf_hub_download(__UpperCAmelCase , __UpperCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) lowercase__ = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = int(deit_name[-6:-4] ) lowercase__ = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith('''tiny''' ): lowercase__ = 1_92 lowercase__ = 7_68 lowercase__ = 12 lowercase__ = 3 elif deit_name[9:].startswith('''small''' ): lowercase__ = 3_84 lowercase__ = 15_36 lowercase__ = 12 lowercase__ = 6 if deit_name[9:].startswith('''base''' ): pass elif deit_name[4:].startswith('''large''' ): lowercase__ = 10_24 lowercase__ = 40_96 lowercase__ = 24 lowercase__ = 16 # load original model from timm lowercase__ = timm.create_model(__UpperCAmelCase , pretrained=__UpperCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys lowercase__ = timm_model.state_dict() lowercase__ = create_rename_keys(__UpperCAmelCase , __UpperCAmelCase ) for src, dest in rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) read_in_q_k_v(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # load HuggingFace model lowercase__ = DeiTForImageClassificationWithTeacher(__UpperCAmelCase ).eval() model.load_state_dict(__UpperCAmelCase ) # Check outputs on an image, prepared by DeiTImageProcessor lowercase__ = int( (2_56 / 2_24) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 lowercase__ = DeiTImageProcessor(size=__UpperCAmelCase , crop_size=config.image_size ) lowercase__ = image_processor(images=prepare_img() , return_tensors='''pt''' ) lowercase__ = encoding['''pixel_values'''] lowercase__ = model(__UpperCAmelCase ) lowercase__ = timm_model(__UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__UpperCAmelCase , outputs.logits , atol=1E-3 ) Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(f'Saving model {deit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__UpperCAmelCase ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) lowerCAmelCase = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
110
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class lowerCamelCase_ : '''simple docstring''' def __init__( self : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Tuple=13 , _lowerCAmelCase : List[str]=7 , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Optional[int]=False , _lowerCAmelCase : Any=True , _lowerCAmelCase : str=99 , _lowerCAmelCase : List[str]=32 , _lowerCAmelCase : Tuple=5 , _lowerCAmelCase : List[str]=4 , _lowerCAmelCase : str=37 , _lowerCAmelCase : Any="gelu" , _lowerCAmelCase : str=0.1 , _lowerCAmelCase : Union[str, Any]=0.1 , _lowerCAmelCase : Optional[int]=512 , _lowerCAmelCase : Union[str, Any]=16 , _lowerCAmelCase : Tuple=2 , _lowerCAmelCase : Tuple=0.02 , _lowerCAmelCase : Tuple=3 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : Union[str, Any]=None , ): SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = seq_length SCREAMING_SNAKE_CASE_ = is_training SCREAMING_SNAKE_CASE_ = use_input_mask SCREAMING_SNAKE_CASE_ = use_token_type_ids SCREAMING_SNAKE_CASE_ = use_labels SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = hidden_dropout_prob SCREAMING_SNAKE_CASE_ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ = max_position_embeddings SCREAMING_SNAKE_CASE_ = type_vocab_size SCREAMING_SNAKE_CASE_ = type_sequence_label_size SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = num_labels SCREAMING_SNAKE_CASE_ = num_choices SCREAMING_SNAKE_CASE_ = scope def lowerCAmelCase_ ( self : List[str] ): SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE_ = None if self.use_input_mask: SCREAMING_SNAKE_CASE_ = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE_ = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None if self.use_labels: SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size] , self.num_choices ) SCREAMING_SNAKE_CASE_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase_ ( self : Optional[int] ): return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def lowerCAmelCase_ ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] ): SCREAMING_SNAKE_CASE_ = LlamaModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase_ ( self : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[str] , ): SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = LlamaModel(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase_ ( self : int , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any , _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , ): SCREAMING_SNAKE_CASE_ = LlamaForCausalLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase_ ( self : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] , ): SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = LlamaForCausalLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() # first forward pass SCREAMING_SNAKE_CASE_ = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , use_cache=_lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids SCREAMING_SNAKE_CASE_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE_ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and SCREAMING_SNAKE_CASE_ = torch.cat([input_ids, next_tokens] , dim=-1 ) SCREAMING_SNAKE_CASE_ = torch.cat([input_mask, next_mask] , dim=-1 ) SCREAMING_SNAKE_CASE_ = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , output_hidden_states=_lowerCAmelCase , )['hidden_states'][0] SCREAMING_SNAKE_CASE_ = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , past_key_values=_lowerCAmelCase , output_hidden_states=_lowerCAmelCase , )['hidden_states'][0] # select random slice SCREAMING_SNAKE_CASE_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() SCREAMING_SNAKE_CASE_ = output_from_no_past[:, -3:, random_slice_idx].detach() SCREAMING_SNAKE_CASE_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1E-3 ) ) def lowerCAmelCase_ ( self : List[str] ): SCREAMING_SNAKE_CASE_ = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ) = config_and_inputs SCREAMING_SNAKE_CASE_ = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () lowercase_ = (LlamaForCausalLM,) if is_torch_available() else () lowercase_ = ( { "feature-extraction": LlamaModel, "text-classification": LlamaForSequenceClassification, "text-generation": LlamaForCausalLM, "zero-shot": LlamaForSequenceClassification, } if is_torch_available() else {} ) lowercase_ = False lowercase_ = False def lowerCAmelCase_ ( self : Dict ): SCREAMING_SNAKE_CASE_ = LlamaModelTester(self ) SCREAMING_SNAKE_CASE_ = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=37 ) def lowerCAmelCase_ ( self : Any ): self.config_tester.run_common_tests() def lowerCAmelCase_ ( self : List[Any] ): SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def lowerCAmelCase_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: SCREAMING_SNAKE_CASE_ = type self.model_tester.create_and_check_model(*_lowerCAmelCase ) def lowerCAmelCase_ ( self : str ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_ = 3 SCREAMING_SNAKE_CASE_ = input_dict['input_ids'] SCREAMING_SNAKE_CASE_ = input_ids.ne(1 ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) SCREAMING_SNAKE_CASE_ = LlamaForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCAmelCase_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_ = 3 SCREAMING_SNAKE_CASE_ = 'single_label_classification' SCREAMING_SNAKE_CASE_ = input_dict['input_ids'] SCREAMING_SNAKE_CASE_ = input_ids.ne(1 ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) SCREAMING_SNAKE_CASE_ = LlamaForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCAmelCase_ ( self : List[Any] ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_ = 3 SCREAMING_SNAKE_CASE_ = 'multi_label_classification' SCREAMING_SNAKE_CASE_ = input_dict['input_ids'] SCREAMING_SNAKE_CASE_ = input_ids.ne(1 ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) SCREAMING_SNAKE_CASE_ = LlamaForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('LLaMA buffers include complex numbers, which breaks this test' ) def lowerCAmelCase_ ( self : int ): pass @parameterized.expand([('linear',), ('dynamic',)] ) def lowerCAmelCase_ ( self : str , _lowerCAmelCase : Tuple ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_ = ids_tensor([1, 10] , config.vocab_size ) SCREAMING_SNAKE_CASE_ = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights SCREAMING_SNAKE_CASE_ = LlamaModel(_lowerCAmelCase ) original_model.to(_lowerCAmelCase ) original_model.eval() SCREAMING_SNAKE_CASE_ = original_model(_lowerCAmelCase ).last_hidden_state SCREAMING_SNAKE_CASE_ = original_model(_lowerCAmelCase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights SCREAMING_SNAKE_CASE_ = {'type': scaling_type, 'factor': 10.0} SCREAMING_SNAKE_CASE_ = LlamaModel(_lowerCAmelCase ) scaled_model.to(_lowerCAmelCase ) scaled_model.eval() SCREAMING_SNAKE_CASE_ = scaled_model(_lowerCAmelCase ).last_hidden_state SCREAMING_SNAKE_CASE_ = scaled_model(_lowerCAmelCase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1E-5 ) ) @require_torch class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowerCAmelCase_ ( self : List[str] ): SCREAMING_SNAKE_CASE_ = [1, 306, 4_658, 278, 6_593, 310, 2_834, 338] SCREAMING_SNAKE_CASE_ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf' , device_map='auto' ) SCREAMING_SNAKE_CASE_ = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 SCREAMING_SNAKE_CASE_ = torch.tensor([[-6.6550, -4.1227, -4.9859, -3.2406, 0.8262, -3.0033, 1.2964, -3.3699]] ) torch.testing.assert_close(out.mean(-1 ) , _lowerCAmelCase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off SCREAMING_SNAKE_CASE_ = torch.tensor([-12.8281, -7.4453, -0.4639, -8.0625, -7.2500, -8.0000, -6.4883, -7.7695, -7.8438, -7.0312, -6.2188, -7.1328, -1.8496, 1.9961, -8.6250, -6.7227, -12.8281, -6.9492, -7.0742, -7.7852, -7.5820, -7.9062, -6.9375, -7.9805, -8.3438, -8.1562, -8.0469, -7.6250, -7.7422, -7.3398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , _lowerCAmelCase , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowerCAmelCase_ ( self : Optional[int] ): SCREAMING_SNAKE_CASE_ = [1, 306, 4_658, 278, 6_593, 310, 2_834, 338] SCREAMING_SNAKE_CASE_ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-hf' , device_map='auto' ) SCREAMING_SNAKE_CASE_ = model(torch.tensor(_lowerCAmelCase ) ) # Expected mean on dim = -1 SCREAMING_SNAKE_CASE_ = torch.tensor([[-2.0622, -1.2794, -1.1638, -0.9788, -1.4603, -1.0238, -1.7893, -1.4411]] ) torch.testing.assert_close(out.mean(-1 ) , _lowerCAmelCase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off SCREAMING_SNAKE_CASE_ = torch.tensor([-8.1406, -8.0547, 2.7461, -1.2344, -0.1448, -1.8262, -1.0020, -1.8154, -1.6895, -1.8516, -2.3574, -0.9277, 3.7598, 6.5742, -1.2998, -0.1177, -8.1406, -2.9688, -2.9199, -3.1699, -3.5254, -2.3555, -2.7988, -3.4141, -2.8262, -4.5195, -3.3379, -3.3164, -2.7832, -3.0273] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , _lowerCAmelCase , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowerCAmelCase_ ( self : int ): SCREAMING_SNAKE_CASE_ = [1, 306, 4_658, 278, 6_593, 310, 2_834, 338] SCREAMING_SNAKE_CASE_ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-chat-hf' , device_map='auto' ) SCREAMING_SNAKE_CASE_ = model(torch.tensor(_lowerCAmelCase ) ) # Expected mean on dim = -1 SCREAMING_SNAKE_CASE_ = torch.tensor([[-0.8562, -1.8520, -0.7551, -0.4162, -1.5161, -1.2038, -2.4823, -2.3254]] ) torch.testing.assert_close(out.mean(-1 ) , _lowerCAmelCase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off SCREAMING_SNAKE_CASE_ = torch.tensor([-2.2227, 4.8828, 0.9023, -0.4578, -0.7871, -0.1033, -0.6221, -0.5786, -0.7803, -1.0674, -1.2920, -0.1570, 0.8008, 2.0723, -0.9497, 0.2771, -2.2227, -0.7612, -1.4346, -1.2061, -1.6426, -0.3000, -0.7139, -1.1934, -1.8691, -1.6973, -1.5947, -1.2705, -0.3523, -0.5513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , _lowerCAmelCase , atol=1E-2 , rtol=1E-2 ) @unittest.skip( 'Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test' ) @slow def lowerCAmelCase_ ( self : int ): SCREAMING_SNAKE_CASE_ = [1, 306, 4_658, 278, 6_593, 310, 2_834, 338] SCREAMING_SNAKE_CASE_ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-70b-hf' , device_map='auto' ) SCREAMING_SNAKE_CASE_ = model(torch.tensor(_lowerCAmelCase ) ) SCREAMING_SNAKE_CASE_ = torch.tensor( [[-4.2327, -3.3360, -4.6665, -4.7631, -1.8180, -3.4170, -1.4211, -3.1810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , _lowerCAmelCase , atol=1E-2 , rtol=1E-2 ) # fmt: off SCREAMING_SNAKE_CASE_ = torch.tensor([-9.4922, -3.9551, 1.7998, -5.6758, -5.1055, -5.8984, -4.8320, -6.8086, -6.5391, -5.6172, -5.5820, -5.5352, 1.7881, 3.6289, -6.5117, -3.4785, -9.5000, -6.0352, -6.8125, -6.0195, -6.6836, -5.4727, -6.2812, -6.0391, -7.3398, -7.4297, -7.4844, -6.5820, -5.8789, -5.5312] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , _lowerCAmelCase , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Model is curently gated' ) @slow def lowerCAmelCase_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE_ = 'Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi' SCREAMING_SNAKE_CASE_ = 'Simply put, the theory of relativity states that ' SCREAMING_SNAKE_CASE_ = LlamaTokenizer.from_pretrained('meta-llama/Llama-2-13b-chat-hf' ) SCREAMING_SNAKE_CASE_ = tokenizer.encode(_lowerCAmelCase , return_tensors='pt' ) SCREAMING_SNAKE_CASE_ = LlamaForCausalLM.from_pretrained( 'meta-llama/Llama-2-13b-chat-hf' , device_map='sequential' , use_safetensors=_lowerCAmelCase ) # greedy generation outputs SCREAMING_SNAKE_CASE_ = model.generate(_lowerCAmelCase , max_new_tokens=64 , top_p=_lowerCAmelCase , temperature=1 , do_sample=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = tokenizer.decode(generated_ids[0] , skip_special_tokens=_lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase )
225
0
import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class _a ( unittest.TestCase ): """simple docstring""" def __A ( self : str ): A_ = 'laion/clap-htsat-unfused' A_ = tempfile.mkdtemp() def __A ( self : Optional[int] , **UpperCAmelCase : Optional[Any] ): return RobertaTokenizer.from_pretrained(self.checkpoint , **_A ) def __A ( self : Tuple , **UpperCAmelCase : List[str] ): return ClapFeatureExtractor.from_pretrained(self.checkpoint , **_A ) def __A ( self : str ): shutil.rmtree(self.tmpdirname ) def __A ( self : Optional[int] ): A_ = self.get_tokenizer() A_ = self.get_feature_extractor() A_ = ClapProcessor(tokenizer=_A , feature_extractor=_A ) processor.save_pretrained(self.tmpdirname ) A_ = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _A ) def __A ( self : Dict ): A_ = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) A_ = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) A_ = self.get_feature_extractor(do_normalize=_A , padding_value=1.0 ) A_ = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=_A , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _A ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _A ) def __A ( self : str ): A_ = self.get_feature_extractor() A_ = self.get_tokenizer() A_ = ClapProcessor(tokenizer=_A , feature_extractor=_A ) A_ = floats_list((3, 1000) ) A_ = feature_extractor(_A , return_tensors="np" ) A_ = processor(audios=_A , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __A ( self : Optional[int] ): A_ = self.get_feature_extractor() A_ = self.get_tokenizer() A_ = ClapProcessor(tokenizer=_A , feature_extractor=_A ) A_ = 'This is a test string' A_ = processor(text=_A ) A_ = tokenizer(_A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __A ( self : Any ): A_ = self.get_feature_extractor() A_ = self.get_tokenizer() A_ = ClapProcessor(tokenizer=_A , feature_extractor=_A ) A_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] A_ = processor.batch_decode(_A ) A_ = tokenizer.batch_decode(_A ) self.assertListEqual(_A , _A ) def __A ( self : Optional[Any] ): A_ = self.get_feature_extractor() A_ = self.get_tokenizer() A_ = ClapProcessor(tokenizer=_A , feature_extractor=_A ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="`processor` and `feature_extractor` model input names do not match" , )
350
from maths.prime_factors import prime_factors def __snake_case ( __UpperCamelCase : int ): """simple docstring""" if not isinstance(__UpperCamelCase ,__UpperCamelCase ): A_ = f'''Input value of [number={number}] must be an integer''' raise TypeError(__UpperCamelCase ) if number < 1: raise ValueError("Input must be a positive integer" ) return -1 if len(prime_factors(__UpperCamelCase ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
329
0
"""simple docstring""" from __future__ import annotations from typing import Any class _UpperCAmelCase : '''simple docstring''' def __init__( self , A = 6 ) -> None: _UpperCAmelCase : Node | None = None _UpperCAmelCase : Node | None = None self.create_linked_list(A ) def __lowerCAmelCase ( self , A ) -> None: _UpperCAmelCase : int = Node() _UpperCAmelCase : Any = current_node _UpperCAmelCase : Optional[Any] = current_node _UpperCAmelCase : Tuple = current_node for _ in range(1 , A ): _UpperCAmelCase : List[str] = Node() _UpperCAmelCase : List[str] = current_node _UpperCAmelCase : List[str] = previous_node _UpperCAmelCase : Dict = current_node _UpperCAmelCase : int = self.front _UpperCAmelCase : Optional[int] = previous_node def __lowerCAmelCase ( self ) -> bool: return ( self.front == self.rear and self.front is not None and self.front.data is None ) def __lowerCAmelCase ( self ) -> Any | None: self.check_can_perform_operation() return self.front.data if self.front else None def __lowerCAmelCase ( self , A ) -> None: if self.rear is None: return self.check_is_full() if not self.is_empty(): _UpperCAmelCase : int = self.rear.next if self.rear: _UpperCAmelCase : Dict = data def __lowerCAmelCase ( self ) -> Any: self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: _UpperCAmelCase : int = self.front.data _UpperCAmelCase : List[Any] = None return data _UpperCAmelCase : Optional[Any] = self.front _UpperCAmelCase : Dict = old_front.next _UpperCAmelCase : Dict = old_front.data _UpperCAmelCase : List[str] = None return data def __lowerCAmelCase ( self ) -> None: if self.is_empty(): raise Exception('''Empty Queue''' ) def __lowerCAmelCase ( self ) -> None: if self.rear and self.rear.next == self.front: raise Exception('''Full Queue''' ) class _UpperCAmelCase : '''simple docstring''' def __init__( self ) -> None: _UpperCAmelCase : Any | None = None _UpperCAmelCase : Node | None = None _UpperCAmelCase : Node | None = None if __name__ == "__main__": import doctest doctest.testmod()
263
"""simple docstring""" import argparse import os import torch from transformers.utils import WEIGHTS_NAME _lowerCAmelCase :int = ['small', 'medium', 'large'] _lowerCAmelCase :int = 'lm_head.decoder.weight' _lowerCAmelCase :Dict = 'lm_head.weight' def lowerCamelCase_ (UpperCamelCase__ : str , UpperCamelCase__ : str ): _UpperCAmelCase : List[Any] = torch.load(UpperCamelCase__ ) _UpperCAmelCase : List[str] = d.pop(UpperCamelCase__ ) os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) torch.save(UpperCamelCase__ , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) if __name__ == "__main__": _lowerCAmelCase :Dict = argparse.ArgumentParser() parser.add_argument('--dialogpt_path', default='.', type=str) _lowerCAmelCase :str = parser.parse_args() for MODEL in DIALOGPT_MODELS: _lowerCAmelCase :Tuple = os.path.join(args.dialogpt_path, f"{MODEL}_ft.pkl") _lowerCAmelCase :int = f"./DialoGPT-{MODEL}" convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
263
1
import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __snake_case ( lowerCamelCase_ , unittest.TestCase ): lowerCAmelCase_ = AudioLDMPipeline lowerCAmelCase_ = TEXT_TO_AUDIO_PARAMS lowerCAmelCase_ = TEXT_TO_AUDIO_BATCH_PARAMS lowerCAmelCase_ = frozenset( [ "num_inference_steps", "num_waveforms_per_prompt", "generator", "latents", "output_type", "return_dict", "callback", "callback_steps", ] ) def __a ( self : Dict ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=(32, 64) , class_embed_type="""simple_projection""" , projection_class_embeddings_input_dim=32 , class_embeddings_concat=_lowercase , ) SCREAMING_SNAKE_CASE__ = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=_lowercase , set_alpha_to_one=_lowercase , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , projection_dim=32 , ) SCREAMING_SNAKE_CASE__ = ClapTextModelWithProjection(_lowercase ) SCREAMING_SNAKE_CASE__ = RobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-random-roberta""" , model_max_length=77 ) SCREAMING_SNAKE_CASE__ = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=1_60_00 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=_lowercase , ) SCREAMING_SNAKE_CASE__ = SpeechTaHifiGan(_lowercase ) SCREAMING_SNAKE_CASE__ = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """vocoder""": vocoder, } return components def __a ( self : Dict , _lowercase : List[Any] , _lowercase : Union[str, Any]=0 ): """simple docstring""" if str(_lowercase ).startswith("""mps""" ): SCREAMING_SNAKE_CASE__ = torch.manual_seed(_lowercase ) else: SCREAMING_SNAKE_CASE__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) SCREAMING_SNAKE_CASE__ = { """prompt""": """A hammer hitting a wooden surface""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, } return inputs def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """cpu""" # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = AudioLDMPipeline(**_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe(**_lowercase ) SCREAMING_SNAKE_CASE__ = output.audios[0] assert audio.ndim == 1 assert len(_lowercase ) == 2_56 SCREAMING_SNAKE_CASE__ = audio[:10] SCREAMING_SNAKE_CASE__ = np.array( [-0.00_50, 0.00_50, -0.00_60, 0.00_33, -0.00_26, 0.00_33, -0.00_27, 0.00_33, -0.00_28, 0.00_33] ) assert np.abs(audio_slice - expected_slice ).max() < 1E-2 def __a ( self : List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = AudioLDMPipeline(**_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = 3 * [inputs["""prompt"""]] # forward SCREAMING_SNAKE_CASE__ = audioldm_pipe(**_lowercase ) SCREAMING_SNAKE_CASE__ = output.audios[0] SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = 3 * [inputs.pop("""prompt""" )] SCREAMING_SNAKE_CASE__ = audioldm_pipe.tokenizer( _lowercase , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=_lowercase , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ = text_inputs["""input_ids"""].to(_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.text_encoder( _lowercase , ) SCREAMING_SNAKE_CASE__ = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state SCREAMING_SNAKE_CASE__ = F.normalize(_lowercase , dim=-1 ) SCREAMING_SNAKE_CASE__ = prompt_embeds # forward SCREAMING_SNAKE_CASE__ = audioldm_pipe(**_lowercase ) SCREAMING_SNAKE_CASE__ = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1E-2 def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = AudioLDMPipeline(**_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = 3 * ["""this is a negative prompt"""] SCREAMING_SNAKE_CASE__ = negative_prompt SCREAMING_SNAKE_CASE__ = 3 * [inputs["""prompt"""]] # forward SCREAMING_SNAKE_CASE__ = audioldm_pipe(**_lowercase ) SCREAMING_SNAKE_CASE__ = output.audios[0] SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = 3 * [inputs.pop("""prompt""" )] SCREAMING_SNAKE_CASE__ = [] for p in [prompt, negative_prompt]: SCREAMING_SNAKE_CASE__ = audioldm_pipe.tokenizer( _lowercase , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=_lowercase , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ = text_inputs["""input_ids"""].to(_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.text_encoder( _lowercase , ) SCREAMING_SNAKE_CASE__ = text_embeds.text_embeds # additional L_2 normalization over each hidden-state SCREAMING_SNAKE_CASE__ = F.normalize(_lowercase , dim=-1 ) embeds.append(_lowercase ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = embeds # forward SCREAMING_SNAKE_CASE__ = audioldm_pipe(**_lowercase ) SCREAMING_SNAKE_CASE__ = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1E-2 def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """cpu""" # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = PNDMScheduler(skip_prk_steps=_lowercase ) SCREAMING_SNAKE_CASE__ = AudioLDMPipeline(**_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = """egg cracking""" SCREAMING_SNAKE_CASE__ = audioldm_pipe(**_lowercase , negative_prompt=_lowercase ) SCREAMING_SNAKE_CASE__ = output.audios[0] assert audio.ndim == 1 assert len(_lowercase ) == 2_56 SCREAMING_SNAKE_CASE__ = audio[:10] SCREAMING_SNAKE_CASE__ = np.array( [-0.00_51, 0.00_50, -0.00_60, 0.00_34, -0.00_26, 0.00_33, -0.00_27, 0.00_33, -0.00_28, 0.00_32] ) assert np.abs(audio_slice - expected_slice ).max() < 1E-2 def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """cpu""" # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = PNDMScheduler(skip_prk_steps=_lowercase ) SCREAMING_SNAKE_CASE__ = AudioLDMPipeline(**_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = """A hammer hitting a wooden surface""" # test num_waveforms_per_prompt=1 (default) SCREAMING_SNAKE_CASE__ = audioldm_pipe(_lowercase , num_inference_steps=2 ).audios assert audios.shape == (1, 2_56) # test num_waveforms_per_prompt=1 (default) for batch of prompts SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 2_56) # test num_waveforms_per_prompt for single prompt SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = audioldm_pipe(_lowercase , num_inference_steps=2 , num_waveforms_per_prompt=_lowercase ).audios assert audios.shape == (num_waveforms_per_prompt, 2_56) # test num_waveforms_per_prompt for batch of prompts SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=_lowercase ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 2_56) def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """cpu""" # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = AudioLDMPipeline(**_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.vocoder.config.sampling_rate SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe(audio_length_in_s=0.0_16 , **_lowercase ) SCREAMING_SNAKE_CASE__ = output.audios[0] assert audio.ndim == 1 assert len(_lowercase ) / vocoder_sampling_rate == 0.0_16 SCREAMING_SNAKE_CASE__ = audioldm_pipe(audio_length_in_s=0.0_32 , **_lowercase ) SCREAMING_SNAKE_CASE__ = output.audios[0] assert audio.ndim == 1 assert len(_lowercase ) / vocoder_sampling_rate == 0.0_32 def __a ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = AudioLDMPipeline(**_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = ["""hey"""] SCREAMING_SNAKE_CASE__ = audioldm_pipe(_lowercase , num_inference_steps=1 ) SCREAMING_SNAKE_CASE__ = output.audios.shape assert audio_shape == (1, 2_56) SCREAMING_SNAKE_CASE__ = audioldm_pipe.vocoder.config config.model_in_dim *= 2 SCREAMING_SNAKE_CASE__ = SpeechTaHifiGan(_lowercase ).to(_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe(_lowercase , num_inference_steps=1 ) SCREAMING_SNAKE_CASE__ = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 2_56) def __a ( self : Optional[Any] ): """simple docstring""" self._test_attention_slicing_forward_pass(test_mean_pixel_difference=_lowercase ) def __a ( self : Dict ): """simple docstring""" self._test_inference_batch_single_identical(test_mean_pixel_difference=_lowercase ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def __a ( self : Dict ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=_lowercase ) @slow class __snake_case ( unittest.TestCase ): def __a ( self : Union[str, Any] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self : Any , _lowercase : Tuple , _lowercase : Any="cpu" , _lowercase : List[str]=torch.floataa , _lowercase : List[Any]=0 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) SCREAMING_SNAKE_CASE__ = np.random.RandomState(_lowercase ).standard_normal((1, 8, 1_28, 16) ) SCREAMING_SNAKE_CASE__ = torch.from_numpy(_lowercase ).to(device=_lowercase , dtype=_lowercase ) SCREAMING_SNAKE_CASE__ = { """prompt""": """A hammer hitting a wooden surface""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 2.5, } return inputs def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = self.get_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = 25 SCREAMING_SNAKE_CASE__ = audioldm_pipe(**_lowercase ).audios[0] assert audio.ndim == 1 assert len(_lowercase ) == 8_19_20 SCREAMING_SNAKE_CASE__ = audio[7_72_30:7_72_40] SCREAMING_SNAKE_CASE__ = np.array( [-0.48_84, -0.46_07, 0.00_23, 0.50_07, 0.58_96, 0.51_51, 0.38_13, -0.02_08, -0.36_87, -0.43_15] ) SCREAMING_SNAKE_CASE__ = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1E-2 def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) SCREAMING_SNAKE_CASE__ = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) SCREAMING_SNAKE_CASE__ = audioldm_pipe.to(_lowercase ) audioldm_pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = self.get_inputs(_lowercase ) SCREAMING_SNAKE_CASE__ = audioldm_pipe(**_lowercase ).audios[0] assert audio.ndim == 1 assert len(_lowercase ) == 8_19_20 SCREAMING_SNAKE_CASE__ = audio[2_77_80:2_77_90] SCREAMING_SNAKE_CASE__ = np.array([-0.21_31, -0.08_73, -0.01_24, -0.01_89, 0.05_69, 0.13_73, 0.18_83, 0.28_86, 0.32_97, 0.22_12] ) SCREAMING_SNAKE_CASE__ = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3E-2
204
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch __lowerCamelCase : int = logging.get_logger(__name__) class __snake_case ( lowerCamelCase_ ): lowerCAmelCase_ = ["pixel_values"] def __init__( self : int , _lowercase : bool = True , _lowercase : Dict[str, int] = None , _lowercase : PILImageResampling = PILImageResampling.BILINEAR , _lowercase : bool = True , _lowercase : Union[int, float] = 1 / 2_55 , _lowercase : bool = True , _lowercase : Dict[str, int] = None , _lowercase : bool = True , **_lowercase : List[Any] , ): """simple docstring""" super().__init__(**_lowercase ) SCREAMING_SNAKE_CASE__ = size if size is not None else {"""shortest_edge""": 2_24} SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) SCREAMING_SNAKE_CASE__ = crop_size if crop_size is not None else {"""height""": 2_56, """width""": 2_56} SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , param_name="""crop_size""" ) SCREAMING_SNAKE_CASE__ = do_resize SCREAMING_SNAKE_CASE__ = size SCREAMING_SNAKE_CASE__ = resample SCREAMING_SNAKE_CASE__ = do_rescale SCREAMING_SNAKE_CASE__ = rescale_factor SCREAMING_SNAKE_CASE__ = do_center_crop SCREAMING_SNAKE_CASE__ = crop_size SCREAMING_SNAKE_CASE__ = do_flip_channel_order def __a ( self : List[Any] , _lowercase : np.ndarray , _lowercase : Dict[str, int] , _lowercase : PILImageResampling = PIL.Image.BILINEAR , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : str , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) if "shortest_edge" not in size: raise ValueError(f"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) SCREAMING_SNAKE_CASE__ = get_resize_output_image_size(_lowercase , size=size["""shortest_edge"""] , default_to_square=_lowercase ) return resize(_lowercase , size=_lowercase , resample=_lowercase , data_format=_lowercase , **_lowercase ) def __a ( self : str , _lowercase : np.ndarray , _lowercase : Dict[str, int] , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : Any , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(_lowercase , size=(size["""height"""], size["""width"""]) , data_format=_lowercase , **_lowercase ) def __a ( self : Optional[Any] , _lowercase : np.ndarray , _lowercase : Union[int, float] , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : Any , ): """simple docstring""" return rescale(_lowercase , scale=_lowercase , data_format=_lowercase , **_lowercase ) def __a ( self : Tuple , _lowercase : np.ndarray , _lowercase : Optional[Union[str, ChannelDimension]] = None ): """simple docstring""" return flip_channel_order(_lowercase , data_format=_lowercase ) def __a ( self : List[str] , _lowercase : ImageInput , _lowercase : bool = None , _lowercase : Dict[str, int] = None , _lowercase : PILImageResampling = None , _lowercase : bool = None , _lowercase : float = None , _lowercase : bool = None , _lowercase : Dict[str, int] = None , _lowercase : bool = None , _lowercase : Optional[Union[str, TensorType]] = None , _lowercase : ChannelDimension = ChannelDimension.FIRST , **_lowercase : int , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE__ = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE__ = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE__ = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE__ = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE__ = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) SCREAMING_SNAKE_CASE__ = size if size is not None else self.size SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , default_to_square=_lowercase ) SCREAMING_SNAKE_CASE__ = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE__ = get_size_dict(_lowercase , param_name="""crop_size""" ) SCREAMING_SNAKE_CASE__ = make_list_of_images(_lowercase ) if not valid_images(_lowercase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE__ = [to_numpy_array(_lowercase ) for image in images] if do_resize: SCREAMING_SNAKE_CASE__ = [self.resize(image=_lowercase , size=_lowercase , resample=_lowercase ) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE__ = [self.center_crop(image=_lowercase , size=_lowercase ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE__ = [self.rescale(image=_lowercase , scale=_lowercase ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: SCREAMING_SNAKE_CASE__ = [self.flip_channel_order(image=_lowercase ) for image in images] SCREAMING_SNAKE_CASE__ = [to_channel_dimension_format(_lowercase , _lowercase ) for image in images] SCREAMING_SNAKE_CASE__ = {"""pixel_values""": images} return BatchFeature(data=_lowercase , tensor_type=_lowercase ) def __a ( self : List[Any] , _lowercase : Dict , _lowercase : List[Tuple] = None ): """simple docstring""" SCREAMING_SNAKE_CASE__ = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_lowercase ) != len(_lowercase ): raise ValueError( """Make sure that you pass in as many target sizes as the batch dimension of the logits""" ) if is_torch_tensor(_lowercase ): SCREAMING_SNAKE_CASE__ = target_sizes.numpy() SCREAMING_SNAKE_CASE__ = [] for idx in range(len(_lowercase ) ): SCREAMING_SNAKE_CASE__ = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="""bilinear""" , align_corners=_lowercase ) SCREAMING_SNAKE_CASE__ = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_lowercase ) else: SCREAMING_SNAKE_CASE__ = logits.argmax(dim=1 ) SCREAMING_SNAKE_CASE__ = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
204
1
"""simple docstring""" from __future__ import annotations from typing import Any class UpperCAmelCase_ : def __init__( self : Tuple , __UpperCamelCase : Tuple ) -> Any: _UpperCamelCase = num_of_nodes _UpperCamelCase = [] _UpperCamelCase = {} def _UpperCamelCase ( self : Optional[int] , __UpperCamelCase : Dict , __UpperCamelCase : int , __UpperCamelCase : List[str] ) -> Tuple: self.m_edges.append([u_node, v_node, weight] ) def _UpperCamelCase ( self : Optional[int] , __UpperCamelCase : List[Any] ) -> int: if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def _UpperCamelCase ( self : List[str] , __UpperCamelCase : Any ) -> List[str]: if self.m_component[u_node] != u_node: for k in self.m_component: _UpperCamelCase = self.find_component(__lowerCamelCase ) def _UpperCamelCase ( self : Union[str, Any] , __UpperCamelCase : Optional[int] , __UpperCamelCase : int , __UpperCamelCase : int ) -> Union[str, Any]: if component_size[u_node] <= component_size[v_node]: _UpperCamelCase = v_node component_size[v_node] += component_size[u_node] self.set_component(__lowerCamelCase ) elif component_size[u_node] >= component_size[v_node]: _UpperCamelCase = self.find_component(__lowerCamelCase ) component_size[u_node] += component_size[v_node] self.set_component(__lowerCamelCase ) def _UpperCamelCase ( self : int ) -> Dict: _UpperCamelCase = [] _UpperCamelCase = 0 _UpperCamelCase = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) _UpperCamelCase = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = edge _UpperCamelCase = self.m_component[u] _UpperCamelCase = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): _UpperCamelCase = [u, v, w] for edge in minimum_weight_edge: if isinstance(__lowerCamelCase , __lowerCamelCase ): _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = edge _UpperCamelCase = self.m_component[u] _UpperCamelCase = self.m_component[v] if u_component != v_component: mst_weight += w self.union(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) print(F'''Added edge [{u} - {v}]\nAdded weight: {w}\n''' ) num_of_components -= 1 _UpperCamelCase = [-1] * self.m_num_of_nodes print(F'''The total weight of the minimal spanning tree is: {mst_weight}''' ) def lowercase ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
256
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE__ ( metaclass=UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = ['''keras_nlp'''] def __init__( self,*__lowerCamelCase,**__lowerCamelCase ): requires_backends(self,['''keras_nlp'''] )
193
0
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml a = NewType('''DataClass''', Any) a = NewType('''DataClassType''', Any) def _snake_case ( _snake_case : int ) -> List[Any]: '''simple docstring''' if isinstance(__snake_case , __snake_case ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'''Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).''' ) def _snake_case ( _snake_case : Union[str, Any] ) -> Callable[[str], Any]: '''simple docstring''' _A = {str(__snake_case ): choice for choice in choices} return lambda _snake_case : str_to_choice.get(__snake_case , __snake_case ) def _snake_case ( *, _snake_case : Union[str, Any] = None , _snake_case : Any = None , _snake_case : int = dataclasses.MISSING , _snake_case : str = dataclasses.MISSING , _snake_case : Dict = None , **_snake_case : int , ) -> dataclasses.Field: '''simple docstring''' if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls _A = {} if aliases is not None: _A = aliases if help is not None: _A = help return dataclasses.field(metadata=__snake_case , default=__snake_case , default_factory=__snake_case , **__snake_case ) class lowercase_ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase : str = 42 def __init__( self : Union[str, Any] , _UpperCAmelCase : Union[DataClassType, Iterable[DataClassType]] , **_UpperCAmelCase : Tuple ): if "formatter_class" not in kwargs: _A = ArgumentDefaultsHelpFormatter super().__init__(**_SCREAMING_SNAKE_CASE ) if dataclasses.is_dataclass(_SCREAMING_SNAKE_CASE ): _A = [dataclass_types] _A = list(_SCREAMING_SNAKE_CASE ) for dtype in self.dataclass_types: self._add_dataclass_arguments(_SCREAMING_SNAKE_CASE ) @staticmethod def lowerCAmelCase_ ( _UpperCAmelCase : ArgumentParser , _UpperCAmelCase : dataclasses.Field ): _A = F'''--{field.name}''' _A = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , _SCREAMING_SNAKE_CASE ): raise RuntimeError( 'Unresolved type detected, which should have been done with the help of ' '`typing.get_type_hints` method by default' ) _A = kwargs.pop('aliases' , [] ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _A = [aliases] _A = getattr(field.type , '__origin__' , field.type ) if origin_type is Union or (hasattr(_SCREAMING_SNAKE_CASE , 'UnionType' ) and isinstance(_SCREAMING_SNAKE_CASE , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(_SCREAMING_SNAKE_CASE ) not in field.type.__args__ ): raise ValueError( 'Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because' ' the argument parser only supports one type per argument.' F''' Problem encountered in field \'{field.name}\'.''' ) if type(_SCREAMING_SNAKE_CASE ) not in field.type.__args__: # filter `str` in Union _A = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] _A = getattr(field.type , '__origin__' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) _A = ( field.type.__args__[0] if isinstance(_SCREAMING_SNAKE_CASE , field.type.__args__[1] ) else field.type.__args__[1] ) _A = getattr(field.type , '__origin__' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) _A = {} if origin_type is Literal or (isinstance(field.type , _SCREAMING_SNAKE_CASE ) and issubclass(field.type , _SCREAMING_SNAKE_CASE )): if origin_type is Literal: _A = field.type.__args__ else: _A = [x.value for x in field.type] _A = make_choice_type_function(kwargs['choices'] ) if field.default is not dataclasses.MISSING: _A = field.default else: _A = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument _A = copy(_SCREAMING_SNAKE_CASE ) # Hack because type=bool in argparse does not behave as we want. _A = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. _A = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way _A = default # This tells argparse we accept 0 or 1 value after --field_name _A = "?" # This is the value that will get picked if we do --field_name (without value) _A = True elif isclass(_SCREAMING_SNAKE_CASE ) and issubclass(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _A = field.type.__args__[0] _A = "+" if field.default_factory is not dataclasses.MISSING: _A = field.default_factory() elif field.default is dataclasses.MISSING: _A = True else: _A = field.type if field.default is not dataclasses.MISSING: _A = field.default elif field.default_factory is not dataclasses.MISSING: _A = field.default_factory() else: _A = True parser.add_argument(_SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): _A = False parser.add_argument(F'''--no_{field.name}''' , action='store_false' , dest=field.name , **_SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( self : int , _UpperCAmelCase : DataClassType ): if hasattr(_SCREAMING_SNAKE_CASE , '_argument_group_name' ): _A = self.add_argument_group(dtype._argument_group_name ) else: _A = self try: _A = get_type_hints(_SCREAMING_SNAKE_CASE ) except NameError: raise RuntimeError( F'''Type resolution failed for {dtype}. Try declaring the class in global scope or ''' 'removing line of `from __future__ import annotations` which opts in Postponed ' 'Evaluation of Annotations (PEP 563)' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(_SCREAMING_SNAKE_CASE ): _A = ".".join(map(_SCREAMING_SNAKE_CASE , sys.version_info[:3] ) ) raise RuntimeError( F'''Type resolution failed for {dtype} on Python {python_version}. Try removing ''' 'line of `from __future__ import annotations` which opts in union types as ' '`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ' 'support Python versions that lower than 3.10, you need to use ' '`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ' '`X | None`.' ) from ex raise for field in dataclasses.fields(_SCREAMING_SNAKE_CASE ): if not field.init: continue _A = type_hints[field.name] self._parse_dataclass_field(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : int=None , _UpperCAmelCase : List[Any]=False , _UpperCAmelCase : Any=True , _UpperCAmelCase : Any=None , _UpperCAmelCase : str=None , ): if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): _A = [] if args_filename: args_files.append(Path(_SCREAMING_SNAKE_CASE ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('.args' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values _A = ArgumentParser() args_file_parser.add_argument(_SCREAMING_SNAKE_CASE , type=_SCREAMING_SNAKE_CASE , action='append' ) # Use only remaining args for further parsing (remove the args_file_flag) _A = args_file_parser.parse_known_args(args=_SCREAMING_SNAKE_CASE ) _A = vars(_SCREAMING_SNAKE_CASE ).get(args_file_flag.lstrip('-' ) , _SCREAMING_SNAKE_CASE ) if cmd_args_file_paths: args_files.extend([Path(_SCREAMING_SNAKE_CASE ) for p in cmd_args_file_paths] ) _A = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last _A = file_args + args if args is not None else file_args + sys.argv[1:] _A = self.parse_known_args(args=_SCREAMING_SNAKE_CASE ) _A = [] for dtype in self.dataclass_types: _A = {f.name for f in dataclasses.fields(_SCREAMING_SNAKE_CASE ) if f.init} _A = {k: v for k, v in vars(_SCREAMING_SNAKE_CASE ).items() if k in keys} for k in keys: delattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _A = dtype(**_SCREAMING_SNAKE_CASE ) outputs.append(_SCREAMING_SNAKE_CASE ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(_SCREAMING_SNAKE_CASE ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F'''Some specified arguments are not used by the HfArgumentParser: {remaining_args}''' ) return (*outputs,) def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : Dict[str, Any] , _UpperCAmelCase : bool = False ): _A = set(args.keys() ) _A = [] for dtype in self.dataclass_types: _A = {f.name for f in dataclasses.fields(_SCREAMING_SNAKE_CASE ) if f.init} _A = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) _A = dtype(**_SCREAMING_SNAKE_CASE ) outputs.append(_SCREAMING_SNAKE_CASE ) if not allow_extra_keys and unused_keys: raise ValueError(F'''Some keys are not used by the HfArgumentParser: {sorted(_SCREAMING_SNAKE_CASE )}''' ) return tuple(_SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( self : Any , _UpperCAmelCase : str , _UpperCAmelCase : bool = False ): with open(Path(_SCREAMING_SNAKE_CASE ) , encoding='utf-8' ) as open_json_file: _A = json.loads(open_json_file.read() ) _A = self.parse_dict(_SCREAMING_SNAKE_CASE , allow_extra_keys=_SCREAMING_SNAKE_CASE ) return tuple(_SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : str , _UpperCAmelCase : bool = False ): _A = self.parse_dict(yaml.safe_load(Path(_SCREAMING_SNAKE_CASE ).read_text() ) , allow_extra_keys=_SCREAMING_SNAKE_CASE ) return tuple(_SCREAMING_SNAKE_CASE )
364
"""simple docstring""" import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def _snake_case ( _snake_case : int ) -> Any: '''simple docstring''' random.seed(_snake_case ) np.random.seed(_snake_case ) torch.manual_seed(_snake_case ) torch.cuda.manual_seed_all(_snake_case ) # ^^ safe to call this function even if cuda is not available class lowercase_ : '''simple docstring''' def __init__( self : Tuple , _UpperCAmelCase : Iterable[torch.nn.Parameter] , _UpperCAmelCase : float = 0.9999 , _UpperCAmelCase : float = 0.0 , _UpperCAmelCase : int = 0 , _UpperCAmelCase : bool = False , _UpperCAmelCase : Union[float, int] = 1.0 , _UpperCAmelCase : Union[float, int] = 2 / 3 , _UpperCAmelCase : Optional[Any] = None , _UpperCAmelCase : Dict[str, Any] = None , **_UpperCAmelCase : Optional[int] , ): if isinstance(_UpperCAmelCase , torch.nn.Module ): _A = ( 'Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ' 'Please pass the parameters of the module instead.' ) deprecate( 'passing a `torch.nn.Module` to `ExponentialMovingAverage`' , '1.0.0' , _UpperCAmelCase , standard_warn=_UpperCAmelCase , ) _A = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility _A = True if kwargs.get('max_value' , _UpperCAmelCase ) is not None: _A = 'The `max_value` argument is deprecated. Please use `decay` instead.' deprecate('max_value' , '1.0.0' , _UpperCAmelCase , standard_warn=_UpperCAmelCase ) _A = kwargs['max_value'] if kwargs.get('min_value' , _UpperCAmelCase ) is not None: _A = 'The `min_value` argument is deprecated. Please use `min_decay` instead.' deprecate('min_value' , '1.0.0' , _UpperCAmelCase , standard_warn=_UpperCAmelCase ) _A = kwargs['min_value'] _A = list(_UpperCAmelCase ) _A = [p.clone().detach() for p in parameters] if kwargs.get('device' , _UpperCAmelCase ) is not None: _A = 'The `device` argument is deprecated. Please use `to` instead.' deprecate('device' , '1.0.0' , _UpperCAmelCase , standard_warn=_UpperCAmelCase ) self.to(device=kwargs['device'] ) _A = None _A = decay _A = min_decay _A = update_after_step _A = use_ema_warmup _A = inv_gamma _A = power _A = 0 _A = None # set in `step()` _A = model_cls _A = model_config @classmethod def lowerCAmelCase_ ( cls : Any , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] ): _A , _A = model_cls.load_config(_UpperCAmelCase , return_unused_kwargs=_UpperCAmelCase ) _A = model_cls.from_pretrained(_UpperCAmelCase ) _A = cls(model.parameters() , model_cls=_UpperCAmelCase , model_config=model.config ) ema_model.load_state_dict(_UpperCAmelCase ) return ema_model def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : Tuple ): if self.model_cls is None: raise ValueError('`save_pretrained` can only be used if `model_cls` was defined at __init__.' ) if self.model_config is None: raise ValueError('`save_pretrained` can only be used if `model_config` was defined at __init__.' ) _A = self.model_cls.from_config(self.model_config ) _A = self.state_dict() state_dict.pop('shadow_params' , _UpperCAmelCase ) model.register_to_config(**_UpperCAmelCase ) self.copy_to(model.parameters() ) model.save_pretrained(_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : int ): _A = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: _A = 1 - (1 + step / self.inv_gamma) ** -self.power else: _A = (1 + step) / (10 + step) _A = min(_UpperCAmelCase , self.decay ) # make sure decay is not smaller than min_decay _A = max(_UpperCAmelCase , self.min_decay ) return cur_decay_value @torch.no_grad() def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : Iterable[torch.nn.Parameter] ): if isinstance(_UpperCAmelCase , torch.nn.Module ): _A = ( 'Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ' 'Please pass the parameters of the module instead.' ) deprecate( 'passing a `torch.nn.Module` to `ExponentialMovingAverage.step`' , '1.0.0' , _UpperCAmelCase , standard_warn=_UpperCAmelCase , ) _A = parameters.parameters() _A = list(_UpperCAmelCase ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. _A = self.get_decay(self.optimization_step ) _A = decay _A = 1 - decay _A = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , _UpperCAmelCase ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): _A = deepspeed.zero.GatheredParameters(_UpperCAmelCase , modifier_rank=_UpperCAmelCase ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(_UpperCAmelCase ) def lowerCAmelCase_ ( self : str , _UpperCAmelCase : Iterable[torch.nn.Parameter] ): _A = list(_UpperCAmelCase ) for s_param, param in zip(self.shadow_params , _UpperCAmelCase ): param.data.copy_(s_param.to(param.device ).data ) def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : Dict=None ): _A = [ p.to(device=_UpperCAmelCase , dtype=_UpperCAmelCase ) if p.is_floating_point() else p.to(device=_UpperCAmelCase ) for p in self.shadow_params ] def lowerCAmelCase_ ( self : Dict ): return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def lowerCAmelCase_ ( self : Any , _UpperCAmelCase : Iterable[torch.nn.Parameter] ): _A = [param.detach().cpu().clone() for param in parameters] def lowerCAmelCase_ ( self : List[Any] , _UpperCAmelCase : Iterable[torch.nn.Parameter] ): if self.temp_stored_params is None: raise RuntimeError('This ExponentialMovingAverage has no `store()`ed weights ' 'to `restore()`' ) for c_param, param in zip(self.temp_stored_params , _UpperCAmelCase ): param.data.copy_(c_param.data ) # Better memory-wise. _A = None def lowerCAmelCase_ ( self : int , _UpperCAmelCase : dict ): _A = copy.deepcopy(_UpperCAmelCase ) _A = state_dict.get('decay' , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError('Decay must be between 0 and 1' ) _A = state_dict.get('min_decay' , self.min_decay ) if not isinstance(self.min_decay , _UpperCAmelCase ): raise ValueError('Invalid min_decay' ) _A = state_dict.get('optimization_step' , self.optimization_step ) if not isinstance(self.optimization_step , _UpperCAmelCase ): raise ValueError('Invalid optimization_step' ) _A = state_dict.get('update_after_step' , self.update_after_step ) if not isinstance(self.update_after_step , _UpperCAmelCase ): raise ValueError('Invalid update_after_step' ) _A = state_dict.get('use_ema_warmup' , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , _UpperCAmelCase ): raise ValueError('Invalid use_ema_warmup' ) _A = state_dict.get('inv_gamma' , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError('Invalid inv_gamma' ) _A = state_dict.get('power' , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError('Invalid power' ) _A = state_dict.get('shadow_params' , _UpperCAmelCase ) if shadow_params is not None: _A = shadow_params if not isinstance(self.shadow_params , _UpperCAmelCase ): raise ValueError('shadow_params must be a list' ) if not all(isinstance(_UpperCAmelCase , torch.Tensor ) for p in self.shadow_params ): raise ValueError('shadow_params must all be Tensors' )
271
0
'''simple docstring''' import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _a ( __a , unittest.TestCase ): __a : Optional[int] = DDIMPipeline __a : Any = UNCONDITIONAL_IMAGE_GENERATION_PARAMS __a : int = PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } __a : Optional[Any] = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS __a : Tuple = False def A ( self : str ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) UpperCAmelCase = DDIMScheduler() UpperCAmelCase = {'''unet''': unet, '''scheduler''': scheduler} return components def A ( self : Optional[Any] , lowercase : int , lowercase : Dict=0 ): '''simple docstring''' if str(lowercase ).startswith('''mps''' ): UpperCAmelCase = torch.manual_seed(lowercase ) else: UpperCAmelCase = torch.Generator(device=lowercase ).manual_seed(lowercase ) UpperCAmelCase = { '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def A ( self : int ): '''simple docstring''' UpperCAmelCase = '''cpu''' UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) UpperCAmelCase = self.get_dummy_inputs(lowercase ) UpperCAmelCase = pipe(**lowercase ).images UpperCAmelCase = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) UpperCAmelCase = np.array( [1.000E00, 5.717E-01, 4.717E-01, 1.000E00, 0.000E00, 1.000E00, 3.000E-04, 0.000E00, 9.000E-04] ) UpperCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1E-3 ) def A ( self : Union[str, Any] ): '''simple docstring''' super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def A ( self : Optional[int] ): '''simple docstring''' super().test_save_load_local(expected_max_difference=3E-3 ) def A ( self : Optional[Any] ): '''simple docstring''' super().test_save_load_optional_components(expected_max_difference=3E-3 ) def A ( self : Optional[Any] ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _a ( unittest.TestCase ): def A ( self : Tuple ): '''simple docstring''' UpperCAmelCase = '''google/ddpm-cifar10-32''' UpperCAmelCase = UNetaDModel.from_pretrained(lowercase ) UpperCAmelCase = DDIMScheduler() UpperCAmelCase = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddim.to(lowercase ) ddim.set_progress_bar_config(disable=lowercase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = ddim(generator=lowercase , eta=0.0 , output_type='''numpy''' ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) UpperCAmelCase = np.array([0.1723, 0.1617, 0.1600, 0.1626, 0.1497, 0.1513, 0.1505, 0.1442, 0.1453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : int ): '''simple docstring''' UpperCAmelCase = '''google/ddpm-ema-bedroom-256''' UpperCAmelCase = UNetaDModel.from_pretrained(lowercase ) UpperCAmelCase = DDIMScheduler.from_pretrained(lowercase ) UpperCAmelCase = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddpm.to(lowercase ) ddpm.set_progress_bar_config(disable=lowercase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = ddpm(generator=lowercase , output_type='''numpy''' ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) UpperCAmelCase = np.array([0.0060, 0.0201, 0.0344, 0.0024, 0.0018, 0.0002, 0.0022, 0.0000, 0.0069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
34
import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from utils_ner import Split, TokenClassificationDataset, TokenClassificationTask import transformers from transformers import ( AutoConfig, AutoModelForTokenClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process A__ = logging.getLogger(__name__) @dataclass class a : __lowerCAmelCase : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __lowerCAmelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __lowerCAmelCase : Optional[str] = field( default="""NER""" , metadata={"""help""": """Task type to fine tune in training (e.g. NER, POS, etc)"""} ) __lowerCAmelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __lowerCAmelCase : bool = field(default=__lowerCamelCase , metadata={"""help""": """Set this flag to use fast tokenization."""} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. __lowerCAmelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class a : __lowerCAmelCase : str = field( metadata={"""help""": """The input data dir. Should contain the .txt files for a CoNLL-2003-formatted task."""} ) __lowerCAmelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Path to a file containing all labels. If not specified, CoNLL-2003 labels are used."""} , ) __lowerCAmelCase : int = field( default=1_28 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __lowerCAmelCase : bool = field( default=__lowerCamelCase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def _lowerCAmelCase ( ) -> Dict: """simple docstring""" snake_case__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case__ , snake_case__ , snake_case__ : str = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case__ , snake_case__ , snake_case__ : List[Any] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ''' --overwrite_output_dir to overcome.''' ) snake_case__ : int = import_module('''tasks''' ) try: snake_case__ : Optional[int] = getattr(__lowerCAmelCase , model_args.task_type ) snake_case__ : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( f"""Task {model_args.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. """ f"""Available tasks classes are: {TokenClassificationTask.__subclasses__()}""" ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , __lowerCAmelCase ) # Set seed set_seed(training_args.seed ) # Prepare CONLL-2003 task snake_case__ : Optional[int] = token_classification_task.get_labels(data_args.labels ) snake_case__ : Dict[int, str] = dict(enumerate(__lowerCAmelCase ) ) snake_case__ : Optional[Any] = len(__lowerCAmelCase ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case__ : List[str] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__lowerCAmelCase , idalabel=__lowerCAmelCase , labelaid={label: i for i, label in enumerate(__lowerCAmelCase )} , cache_dir=model_args.cache_dir , ) snake_case__ : List[str] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast , ) snake_case__ : Tuple = AutoModelForTokenClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , ) # Get datasets snake_case__ : Dict = ( TokenClassificationDataset( token_classification_task=__lowerCAmelCase , data_dir=data_args.data_dir , tokenizer=__lowerCAmelCase , labels=__lowerCAmelCase , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case__ : Optional[Any] = ( TokenClassificationDataset( token_classification_task=__lowerCAmelCase , data_dir=data_args.data_dir , tokenizer=__lowerCAmelCase , labels=__lowerCAmelCase , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def align_predictions(__lowerCAmelCase , __lowerCAmelCase ) -> Tuple[List[int], List[int]]: snake_case__ : Any = np.argmax(__lowerCAmelCase , axis=2 ) snake_case__ , snake_case__ : List[Any] = preds.shape snake_case__ : List[Any] = [[] for _ in range(__lowerCAmelCase )] snake_case__ : int = [[] for _ in range(__lowerCAmelCase )] for i in range(__lowerCAmelCase ): for j in range(__lowerCAmelCase ): if label_ids[i, j] != nn.CrossEntropyLoss().ignore_index: out_label_list[i].append(label_map[label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) return preds_list, out_label_list def compute_metrics(__lowerCAmelCase ) -> Dict: snake_case__ , snake_case__ : List[str] = align_predictions(p.predictions , p.label_ids ) return { "accuracy_score": accuracy_score(__lowerCAmelCase , __lowerCAmelCase ), "precision": precision_score(__lowerCAmelCase , __lowerCAmelCase ), "recall": recall_score(__lowerCAmelCase , __lowerCAmelCase ), "f1": fa_score(__lowerCAmelCase , __lowerCAmelCase ), } # Data collator snake_case__ : Any = DataCollatorWithPadding(__lowerCAmelCase , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case__ : Tuple = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=__lowerCAmelCase , eval_dataset=__lowerCAmelCase , compute_metrics=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_process_zero(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case__ : Any = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case__ : Tuple = trainer.evaluate() snake_case__ : Optional[Any] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_process_zero(): with open(__lowerCAmelCase , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , __lowerCAmelCase , __lowerCAmelCase ) writer.write('''%s = %s\n''' % (key, value) ) results.update(__lowerCAmelCase ) # Predict if training_args.do_predict: snake_case__ : Optional[Any] = TokenClassificationDataset( token_classification_task=__lowerCAmelCase , data_dir=data_args.data_dir , tokenizer=__lowerCAmelCase , labels=__lowerCAmelCase , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.test , ) snake_case__ , snake_case__ , snake_case__ : List[str] = trainer.predict(__lowerCAmelCase ) snake_case__ , snake_case__ : int = align_predictions(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : Optional[int] = os.path.join(training_args.output_dir , '''test_results.txt''' ) if trainer.is_world_process_zero(): with open(__lowerCAmelCase , '''w''' ) as writer: for key, value in metrics.items(): logger.info(''' %s = %s''' , __lowerCAmelCase , __lowerCAmelCase ) writer.write('''%s = %s\n''' % (key, value) ) # Save predictions snake_case__ : Any = os.path.join(training_args.output_dir , '''test_predictions.txt''' ) if trainer.is_world_process_zero(): with open(__lowerCAmelCase , '''w''' ) as writer: with open(os.path.join(data_args.data_dir , '''test.txt''' ) , '''r''' ) as f: token_classification_task.write_predictions_to_file(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) return results def _lowerCAmelCase ( __lowerCAmelCase ) -> Union[str, Any]: """simple docstring""" main() if __name__ == "__main__": main()
230
0
'''simple docstring''' import torch from diffusers import DPMSolverSDEScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import require_torchsde from .test_schedulers import SchedulerCommonTest @require_torchsde class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = (DPMSolverSDEScheduler,) lowerCamelCase = 10 def UpperCAmelCase_ ( self , **_lowerCamelCase ) -> List[Any]: A_ : Optional[int] = { """num_train_timesteps""": 1100, """beta_start""": 0.0001, """beta_end""": 0.02, """beta_schedule""": """linear""", """noise_sampler_seed""": 0, } config.update(**_lowerCamelCase ) return config def UpperCAmelCase_ ( self ) -> Any: for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: for beta_start, beta_end in zip([0.0_0001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=_lowerCamelCase , beta_end=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Any: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> int: A_ : Union[str, Any] = self.scheduler_classes[0] A_ : List[str] = self.get_scheduler_config() A_ : int = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps ) A_ : List[Any] = self.dummy_model() A_ : Tuple = self.dummy_sample_deter * scheduler.init_noise_sigma A_ : Dict = sample.to(_lowerCamelCase ) for i, t in enumerate(scheduler.timesteps ): A_ : Optional[int] = scheduler.scale_model_input(_lowerCamelCase , _lowerCamelCase ) A_ : int = model(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : List[Any] = output.prev_sample A_ : Optional[Any] = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : int = torch.mean(torch.abs(_lowerCamelCase ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.47_8210_4492_1875 ) < 1e-2 assert abs(result_mean.item() - 0.2178_7059_6456_5277 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_3521_1181_6406 ) < 1e-2 assert abs(result_mean.item() - 0.2_2342_9068_9229_9652 ) < 1e-3 else: assert abs(result_sum.item() - 162.52_3834_2285_1562 ) < 1e-2 assert abs(result_mean.item() - 0.211_6195_7085_1326 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : List[str] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config(prediction_type="""v_prediction""" ) A_ : List[Any] = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps ) A_ : List[Any] = self.dummy_model() A_ : Optional[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma A_ : Dict = sample.to(_lowerCamelCase ) for i, t in enumerate(scheduler.timesteps ): A_ : Dict = scheduler.scale_model_input(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = model(_lowerCamelCase , _lowerCamelCase ) A_ : List[Any] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Optional[Any] = output.prev_sample A_ : Any = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : str = torch.mean(torch.abs(_lowerCamelCase ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 124.77_1492_0043_9453 ) < 1e-2 assert abs(result_mean.item() - 0.1_6226_2890_1481_6284 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 128.1_6633_6059_5703 ) < 1e-2 assert abs(result_mean.item() - 0.1_6688_3260_0116_7297 ) < 1e-3 else: assert abs(result_sum.item() - 119.8_4875_4882_8125 ) < 1e-2 assert abs(result_mean.item() - 0.1560_5306_6253_6621 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Dict: A_ : Tuple = self.scheduler_classes[0] A_ : int = self.get_scheduler_config() A_ : Optional[Any] = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps , device=_lowerCamelCase ) A_ : Optional[int] = self.dummy_model() A_ : int = self.dummy_sample_deter.to(_lowerCamelCase ) * scheduler.init_noise_sigma for t in scheduler.timesteps: A_ : Optional[Any] = scheduler.scale_model_input(_lowerCamelCase , _lowerCamelCase ) A_ : Union[str, Any] = model(_lowerCamelCase , _lowerCamelCase ) A_ : List[str] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : List[Any] = output.prev_sample A_ : Optional[int] = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Tuple = torch.mean(torch.abs(_lowerCamelCase ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.46_9573_9746_0938 ) < 1e-2 assert abs(result_mean.item() - 0.2_1805_9346_0798_2635 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_3536_3769_5312 ) < 1e-2 assert abs(result_mean.item() - 0.2_2342_9083_8241_5771 ) < 1e-3 else: assert abs(result_sum.item() - 162.52_3834_2285_1562 ) < 1e-2 assert abs(result_mean.item() - 0.211_6195_7085_1326 ) < 1e-3 def UpperCAmelCase_ ( self ) -> List[str]: A_ : Union[str, Any] = self.scheduler_classes[0] A_ : str = self.get_scheduler_config() A_ : Optional[int] = scheduler_class(**_lowerCamelCase , use_karras_sigmas=_lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps , device=_lowerCamelCase ) A_ : Any = self.dummy_model() A_ : int = self.dummy_sample_deter.to(_lowerCamelCase ) * scheduler.init_noise_sigma A_ : List[str] = sample.to(_lowerCamelCase ) for t in scheduler.timesteps: A_ : int = scheduler.scale_model_input(_lowerCamelCase , _lowerCamelCase ) A_ : List[Any] = model(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = output.prev_sample A_ : List[str] = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Optional[int] = torch.mean(torch.abs(_lowerCamelCase ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 176.66_9741_3574_2188 ) < 1e-2 assert abs(result_mean.item() - 0.2_3003_8727_3098_1811 ) < 1e-2 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 177.63_6535_6445_3125 ) < 1e-2 assert abs(result_mean.item() - 0.2_3003_8727_3098_1811 ) < 1e-2 else: assert abs(result_sum.item() - 170.3_1352_2338_8672 ) < 1e-2 assert abs(result_mean.item() - 0.2_3003_8727_3098_1811 ) < 1e-2
370
'''simple docstring''' def UpperCAmelCase ( a_ = 5_0_0_0_0_0_0_0 ) -> int: """simple docstring""" A_ : Union[str, Any] = set() A_ : List[str] = int((limit - 2_4) ** (1 / 2) ) A_ : Dict = set(range(3 , prime_square_limit + 1 , 2 ) ) primes.add(2 ) for p in range(3 , prime_square_limit + 1 , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , prime_square_limit + 1 , a_ ) ) ) for primea in primes: A_ : Union[str, Any] = primea * primea for primea in primes: A_ : Optional[int] = primea * primea * primea if square + cube >= limit - 1_6: break for primea in primes: A_ : Tuple = primea * primea * primea * primea A_ : List[str] = square + cube + tetr if total >= limit: break ret.add(a_ ) return len(a_ ) if __name__ == "__main__": print(f'{solution() = }')
164
0
"""simple docstring""" import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features a_ = logging.get_logger(__name__) a_ = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) a_ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __snake_case : """simple docstring""" _lowerCamelCase = field( default=UpperCAmelCase__ , metadata={"""help""": """Model type selected in the list: """ + """, """.join(UpperCAmelCase__ )} ) _lowerCamelCase = field( default=UpperCAmelCase__ , metadata={"""help""": """The input data dir. Should contain the .json files for the SQuAD task."""} ) _lowerCamelCase = field( default=1_28 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) _lowerCamelCase = field( default=1_28 , metadata={"""help""": """When splitting up a long document into chunks, how much stride to take between chunks."""} , ) _lowerCamelCase = field( default=64 , metadata={ """help""": ( """The maximum number of tokens for the question. Questions longer than this will """ """be truncated to this length.""" ) } , ) _lowerCamelCase = field( default=30 , metadata={ """help""": ( """The maximum length of an answer that can be generated. This is needed because the start """ """and end predictions are not conditioned on one another.""" ) } , ) _lowerCamelCase = field( default=UpperCAmelCase__ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) _lowerCamelCase = field( default=UpperCAmelCase__ , metadata={"""help""": """If true, the SQuAD examples contain some that do not have an answer."""} ) _lowerCamelCase = field( default=0.0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) _lowerCamelCase = field( default=20 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) _lowerCamelCase = field( default=0 , metadata={ """help""": ( """language id of input for language-specific xlm models (see""" """ tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)""" ) } , ) _lowerCamelCase = field(default=1 , metadata={"""help""": """multiple threads for converting example to features"""} ) class __snake_case ( UpperCAmelCase__ ): """simple docstring""" _lowerCamelCase = """train""" _lowerCamelCase = """dev""" class __snake_case ( UpperCAmelCase__ ): """simple docstring""" _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = None , __lowerCamelCase = Split.train , __lowerCamelCase = False , __lowerCamelCase = None , __lowerCamelCase = "pt" , ): '''simple docstring''' __A : Optional[Any] = args __A : List[str] = is_language_sensitive __A : int = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(__lowercase , __lowercase ): try: __A : int = Split[mode] except KeyError: raise KeyError('''mode is not a valid split name''' ) __A : Any = mode # Load data features from cache or dataset file __A : List[Any] = '''v2''' if args.version_2_with_negative else '''v1''' __A : Tuple = os.path.join( cache_dir if cache_dir is not None else args.data_dir , F"""cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}""" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __A : Optional[Any] = cached_features_file + '''.lock''' with FileLock(__lowercase ): if os.path.exists(__lowercase ) and not args.overwrite_cache: __A : Any = time.time() __A : str = torch.load(__lowercase ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. __A : str = self.old_features['''features'''] __A : Optional[Any] = self.old_features.get('''dataset''' , __lowercase ) __A : List[Any] = self.old_features.get('''examples''' , __lowercase ) logger.info( F"""Loading features from cached file {cached_features_file} [took %.3f s]""" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( F"""Deleting cached file {cached_features_file} will allow dataset and examples to be cached in""" ''' future run''' ) else: if mode == Split.dev: __A : List[str] = self.processor.get_dev_examples(args.data_dir ) else: __A : List[str] = self.processor.get_train_examples(args.data_dir ) __A , __A : Optional[int] = squad_convert_examples_to_features( examples=self.examples , tokenizer=__lowercase , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=__lowercase , ) __A : Union[str, Any] = time.time() torch.save( {'''features''': self.features, '''dataset''': self.dataset, '''examples''': self.examples} , __lowercase , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F"""Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]""" ) def __len__( self ): '''simple docstring''' return len(self.features ) def __getitem__( self , __lowerCamelCase ): '''simple docstring''' __A : Dict = self.features[i] __A : Optional[Any] = torch.tensor(feature.input_ids , dtype=torch.long ) __A : Union[str, Any] = torch.tensor(feature.attention_mask , dtype=torch.long ) __A : Union[str, Any] = torch.tensor(feature.token_type_ids , dtype=torch.long ) __A : List[Any] = torch.tensor(feature.cls_index , dtype=torch.long ) __A : Dict = torch.tensor(feature.p_mask , dtype=torch.float ) __A : str = torch.tensor(feature.is_impossible , dtype=torch.float ) __A : int = { '''input_ids''': input_ids, '''attention_mask''': attention_mask, '''token_type_ids''': token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({'''cls_index''': cls_index, '''p_mask''': p_mask} ) if self.args.version_2_with_negative: inputs.update({'''is_impossible''': is_impossible} ) if self.is_language_sensitive: inputs.update({'''langs''': (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: __A : Optional[Any] = torch.tensor(feature.start_position , dtype=torch.long ) __A : int = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({'''start_positions''': start_positions, '''end_positions''': end_positions} ) return inputs
179
from __future__ import annotations def lowerCamelCase__ ( _A ): '''simple docstring''' snake_case_ = str(_A ) return len(_A ) == 9 and set(_A ) == set("123456789" ) def lowerCamelCase__ ( ): '''simple docstring''' for base_num in range(9999 , 4999 , -1 ): snake_case_ = 100002 * base_num if is_9_pandigital(_A ): return candidate for base_num in range(333 , 99 , -1 ): snake_case_ = 1002003 * base_num if is_9_pandigital(_A ): return candidate return None if __name__ == "__main__": print(f'''{solution() = }''')
187
0
"""simple docstring""" def A_ ( _lowerCAmelCase : list, _lowerCAmelCase : list, _lowerCAmelCase : int ): """simple docstring""" if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): raise ValueError('''The length of profit and weight must be same.''' ) if max_weight <= 0: raise ValueError('''max_weight must greater than zero.''' ) if any(p < 0 for p in profit ): raise ValueError('''Profit can not be negative.''' ) if any(w < 0 for w in weight ): raise ValueError('''Weight can not be negative.''' ) # List created to store profit gained for the 1kg in case of each weight # respectively. Calculate and append profit/weight for each element. _a = [p / w for p, w in zip(_lowerCAmelCase, _lowerCAmelCase )] # Creating a copy of the list and sorting profit/weight in ascending order _a = sorted(_lowerCAmelCase ) # declaring useful variables _a = len(_lowerCAmelCase ) _a = 0 _a = 0 _a = 0 # loop till the total weight do not reach max limit e.g. 15 kg and till i<length while limit <= max_weight and i < length: # flag value for encountered greatest element in sorted_profit_by_weight _a = sorted_profit_by_weight[length - i - 1] _a = profit_by_weight.index(_lowerCAmelCase ) _a = -1 # check if the weight encountered is less than the total weight # encountered before. if max_weight - limit >= weight[index]: limit += weight[index] # Adding profit gained for the given weight 1 === # weight[index]/weight[index] gain += 1 * profit[index] else: # Since the weight encountered is greater than limit, therefore take the # required number of remaining kgs and calculate profit for it. # weight remaining / weight[index] gain += (max_weight - limit) / weight[index] * profit[index] break i += 1 return gain if __name__ == "__main__": print( '''Input profits, weights, and then max_weight (all positive ints) separated by ''' '''spaces.''' ) __snake_case = [int(x) for x in input('''Input profits separated by spaces: ''').split()] __snake_case = [int(x) for x in input('''Input weights separated by spaces: ''').split()] __snake_case = int(input('''Max weight allowed: ''')) # Function Call calc_profit(profit, weight, max_weight)
153
"""simple docstring""" import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class __lowerCamelCase ( a__ ): '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> List[Any]: super().__init__( __UpperCAmelCase , split=__UpperCAmelCase , features=__UpperCAmelCase , cache_dir=__UpperCAmelCase , keep_in_memory=__UpperCAmelCase , streaming=__UpperCAmelCase , num_proc=__UpperCAmelCase , **__UpperCAmelCase , ) _a = field _a = path_or_paths if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else {self.split: path_or_paths} _a = Json( cache_dir=__UpperCAmelCase , data_files=__UpperCAmelCase , features=__UpperCAmelCase , field=__UpperCAmelCase , **__UpperCAmelCase , ) def _UpperCAmelCase ( self ) -> str: # Build iterable dataset if self.streaming: _a = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: _a = None _a = None _a = None _a = None self.builder.download_and_prepare( download_config=__UpperCAmelCase , download_mode=__UpperCAmelCase , verification_mode=__UpperCAmelCase , base_path=__UpperCAmelCase , num_proc=self.num_proc , ) _a = self.builder.as_dataset( split=self.split , verification_mode=__UpperCAmelCase , in_memory=self.keep_in_memory ) return dataset class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> Tuple: if num_proc is not None and num_proc <= 0: raise ValueError(F'num_proc {num_proc} must be an integer > 0.' ) _a = dataset _a = path_or_buf _a = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE _a = num_proc _a = '''utf-8''' _a = to_json_kwargs def _UpperCAmelCase ( self ) -> int: _a = self.to_json_kwargs.pop('''path_or_buf''' , __UpperCAmelCase ) _a = self.to_json_kwargs.pop('''orient''' , '''records''' ) _a = self.to_json_kwargs.pop('''lines''' , True if orient == '''records''' else False ) _a = self.to_json_kwargs.pop('''index''' , False if orient in ['''split''', '''table'''] else True ) _a = self.to_json_kwargs.pop('''compression''' , __UpperCAmelCase ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(F'`datasets` currently does not support {compression} compression' ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , '''wb''' , compression=__UpperCAmelCase ) as buffer: _a = self._write(file_obj=__UpperCAmelCase , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( F'The compression parameter is not supported when writing to a buffer, but compression={compression}' ''' was passed. Please provide a local path instead.''' ) _a = self._write( file_obj=self.path_or_buf , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **self.to_json_kwargs ) return written def _UpperCAmelCase ( self , __UpperCAmelCase ) -> Optional[int]: _a , _a , _a , _a , _a = args _a = query_table( table=self.dataset.data , key=slice(__UpperCAmelCase , offset + self.batch_size ) , indices=self.dataset._indices , ) _a = batch.to_pandas().to_json( path_or_buf=__UpperCAmelCase , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **__UpperCAmelCase ) if not json_str.endswith('''\n''' ): json_str += "\n" return json_str.encode(self.encoding ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) -> int: _a = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating json from Arrow format''' , ): _a = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(__UpperCAmelCase ) else: _a , _a = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , __UpperCAmelCase , __UpperCAmelCase )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating json from Arrow format''' , ): written += file_obj.write(__UpperCAmelCase ) return written
153
1
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging lowercase : List[Any] = logging.get_logger(__name__) lowercase : int = {"""vocab_file""": """spiece.model"""} lowercase : Optional[int] = { """vocab_file""": { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model""", } } lowercase : Optional[int] = { """xlnet-base-cased""": None, """xlnet-large-cased""": None, } # Segments (not really needed) lowercase : int = 0 lowercase : int = 1 lowercase : Optional[Any] = 2 lowercase : Tuple = 3 lowercase : str = 4 class __snake_case ( lowerCAmelCase ): _a : Dict= VOCAB_FILES_NAMES _a : Optional[Any]= PRETRAINED_VOCAB_FILES_MAP _a : int= PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a : Tuple= "left" def __init__( self ,snake_case ,snake_case=False ,snake_case=True ,snake_case=False ,snake_case="<s>" ,snake_case="</s>" ,snake_case="<unk>" ,snake_case="<sep>" ,snake_case="<pad>" ,snake_case="<cls>" ,snake_case="<mask>" ,snake_case=["<eop>", "<eod>"] ,snake_case = None ,**snake_case ,): '''simple docstring''' lowercase : List[str] = AddedToken(snake_case ,lstrip=snake_case ,rstrip=snake_case ) if isinstance(snake_case ,snake_case ) else mask_token lowercase : int = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case ,remove_space=snake_case ,keep_accents=snake_case ,bos_token=snake_case ,eos_token=snake_case ,unk_token=snake_case ,sep_token=snake_case ,pad_token=snake_case ,cls_token=snake_case ,mask_token=snake_case ,additional_special_tokens=snake_case ,sp_model_kwargs=self.sp_model_kwargs ,**snake_case ,) lowercase : List[Any] = 3 lowercase : str = do_lower_case lowercase : Any = remove_space lowercase : str = keep_accents lowercase : Any = vocab_file lowercase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case ) @property def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' return len(self.sp_model ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : Tuple = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' lowercase : Optional[int] = self.__dict__.copy() lowercase : Union[str, Any] = None return state def __setstate__( self ,snake_case ): '''simple docstring''' lowercase : Union[str, Any] = d # for backward compatibility if not hasattr(self ,"""sp_model_kwargs""" ): lowercase : Optional[int] = {} lowercase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' if self.remove_space: lowercase : Optional[Any] = """ """.join(inputs.strip().split() ) else: lowercase : Optional[int] = inputs lowercase : Optional[int] = outputs.replace("""``""" ,"""\"""" ).replace("""''""" ,"""\"""" ) if not self.keep_accents: lowercase : str = unicodedata.normalize("""NFKD""" ,snake_case ) lowercase : Optional[Any] = """""".join([c for c in outputs if not unicodedata.combining(snake_case )] ) if self.do_lower_case: lowercase : str = outputs.lower() return outputs def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase : Tuple = self.preprocess_text(snake_case ) lowercase : Any = self.sp_model.encode(snake_case ,out_type=snake_case ) lowercase : Any = [] for piece in pieces: if len(snake_case ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): lowercase : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case ,"""""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowercase : Tuple = cur_pieces[1:] else: lowercase : Optional[Any] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case ) else: new_pieces.append(snake_case ) return new_pieces def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' return self.sp_model.PieceToId(snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase : Dict = """""".join(snake_case ).replace(snake_case ,""" """ ).strip() return out_string def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case = False ,snake_case = None ,snake_case = True ,**snake_case ,): '''simple docstring''' lowercase : Optional[int] = kwargs.pop("""use_source_tokenizer""" ,snake_case ) lowercase : Optional[int] = self.convert_ids_to_tokens(snake_case ,skip_special_tokens=snake_case ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowercase : Optional[Any] = [] lowercase : Union[str, Any] = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) lowercase : Dict = [] sub_texts.append(snake_case ) else: current_sub_text.append(snake_case ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens lowercase : List[str] = """""".join(snake_case ) lowercase : Tuple = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowercase : Optional[Any] = self.clean_up_tokenization(snake_case ) return clean_text else: return text def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case = None ): '''simple docstring''' lowercase : int = [self.sep_token_id] lowercase : Optional[int] = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case = None ,snake_case = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case ,token_ids_a=snake_case ,already_has_special_tokens=snake_case ) if token_ids_a is not None: return ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1, 1] return ([0] * len(snake_case )) + [1, 1] def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case = None ): '''simple docstring''' lowercase : int = [self.sep_token_id] lowercase : List[Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return lowercase : List[Any] = os.path.join( snake_case ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case ,"""wb""" ) as fi: lowercase : Dict = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,)
20
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class _SCREAMING_SNAKE_CASE ( _a ): def __init__( self : List[Any] , __lowerCamelCase : Callable , __lowerCamelCase : Optional[Features] = None , __lowerCamelCase : str = None , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[dict] = None , __lowerCamelCase : Optional[int] = None , **__lowerCamelCase : List[Any] , ): super().__init__( features=__lowerCamelCase , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase , streaming=__lowerCamelCase , num_proc=__lowerCamelCase , **__lowerCamelCase , ) UpperCamelCase :Union[str, Any] = Generator( cache_dir=__lowerCamelCase , features=__lowerCamelCase , generator=__lowerCamelCase , gen_kwargs=__lowerCamelCase , **__lowerCamelCase , ) def _A ( self : List[str] ): # Build iterable dataset if self.streaming: UpperCamelCase :Any = self.builder.as_streaming_dataset(split="""train""" ) # Build regular (map-style) dataset else: UpperCamelCase :Tuple = None UpperCamelCase :Dict = None UpperCamelCase :Dict = None UpperCamelCase :List[str] = None self.builder.download_and_prepare( download_config=__lowerCamelCase , download_mode=__lowerCamelCase , verification_mode=__lowerCamelCase , base_path=__lowerCamelCase , num_proc=self.num_proc , ) UpperCamelCase :Tuple = self.builder.as_dataset( split="""train""" , verification_mode=__lowerCamelCase , in_memory=self.keep_in_memory ) return dataset
38
0
"""simple docstring""" def lowerCamelCase_ ( _lowerCamelCase ): lowerCamelCase__ : Union[str, Any] = [] lowerCamelCase__ : List[str] = [] lowerCamelCase__ : Tuple = { '^': 3, '*': 2, '/': 2, '%': 2, '+': 1, '-': 1, } # Priority of each operator lowerCamelCase__ : List[str] = len(_lowerCamelCase ) if (len(_lowerCamelCase ) > 7) else 7 # Print table header for output print( 'Symbol'.center(8 ) , 'Stack'.center(_lowerCamelCase ) , 'Postfix'.center(_lowerCamelCase ) , sep=' | ' , ) print('-' * (print_width * 3 + 7) ) for x in infix: if x.isalpha() or x.isdigit(): post_fix.append(_lowerCamelCase ) # if x is Alphabet / Digit, add it to Postfix elif x == "(": stack.append(_lowerCamelCase ) # if x is "(" push to Stack elif x == ")": # if x is ")" pop stack until "(" is encountered while stack[-1] != "(": post_fix.append(stack.pop() ) # Pop stack & add the content to Postfix stack.pop() else: if len(_lowerCamelCase ) == 0: stack.append(_lowerCamelCase ) # If stack is empty, push x to stack else: # while priority of x is not > priority of element in the stack while len(_lowerCamelCase ) > 0 and priority[x] <= priority[stack[-1]]: post_fix.append(stack.pop() ) # pop stack & add to Postfix stack.append(_lowerCamelCase ) # push x to stack print( x.center(8 ) , (''.join(_lowerCamelCase )).ljust(_lowerCamelCase ) , (''.join(_lowerCamelCase )).ljust(_lowerCamelCase ) , sep=' | ' , ) # Output in tabular format while len(_lowerCamelCase ) > 0: # while stack is not empty post_fix.append(stack.pop() ) # pop stack & add to Postfix print( ' '.center(8 ) , (''.join(_lowerCamelCase )).ljust(_lowerCamelCase ) , (''.join(_lowerCamelCase )).ljust(_lowerCamelCase ) , sep=' | ' , ) # Output in tabular format return "".join(_lowerCamelCase ) # return Postfix as str def lowerCamelCase_ ( _lowerCamelCase ): lowerCamelCase__ : Union[str, Any] = list(infix[::-1] ) # reverse the infix equation for i in range(len(_lowerCamelCase ) ): if infix[i] == "(": lowerCamelCase__ : List[Any] = ')' # change "(" to ")" elif infix[i] == ")": lowerCamelCase__ : Tuple = '(' # change ")" to "(" return (infix_2_postfix(''.join(_lowerCamelCase ) ))[ ::-1 ] # call infix_2_postfix on Infix, return reverse of Postfix if __name__ == "__main__": A_ : Tuple = input("\nEnter an Infix Equation = ") # Input an Infix equation A_ : List[str] = "".join(Infix.split()) # Remove spaces from the input print("\n\t", Infix, "(Infix) -> ", infix_2_prefix(Infix), "(Prefix)")
316
"""simple docstring""" import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 A_ : Any = { "return_dict": False, "output_hidden_states": True, "output_attentions": True, "torchscript": True, "torch_dtype": "float16", "use_bfloat16": True, "tf_legacy_loss": True, "pruned_heads": {"a": 1}, "tie_word_embeddings": False, "is_decoder": True, "cross_attention_hidden_size": 1_28, "add_cross_attention": True, "tie_encoder_decoder": True, "max_length": 50, "min_length": 3, "do_sample": True, "early_stopping": True, "num_beams": 3, "num_beam_groups": 3, "diversity_penalty": 0.5, "temperature": 2.0, "top_k": 10, "top_p": 0.7, "typical_p": 0.2, "repetition_penalty": 0.8, "length_penalty": 0.8, "no_repeat_ngram_size": 5, "encoder_no_repeat_ngram_size": 5, "bad_words_ids": [1, 2, 3], "num_return_sequences": 3, "chunk_size_feed_forward": 5, "output_scores": True, "return_dict_in_generate": True, "forced_bos_token_id": 2, "forced_eos_token_id": 3, "remove_invalid_values": True, "architectures": ["BertModel"], "finetuning_task": "translation", "id2label": {0: "label"}, "label2id": {"label": "0"}, "tokenizer_class": "BertTokenizerFast", "prefix": "prefix", "bos_token_id": 6, "pad_token_id": 7, "eos_token_id": 8, "sep_token_id": 9, "decoder_start_token_id": 10, "exponential_decay_length_penalty": (5, 1.01), "suppress_tokens": [0, 1], "begin_suppress_tokens": 2, "task_specific_params": {"translation": "some_params"}, "problem_type": "regression", } @is_staging_test class a_ ( unittest.TestCase ): '''simple docstring''' @classmethod def a__ (cls ): '''simple docstring''' lowerCamelCase__ : Tuple = TOKEN HfFolder.save_token(lowerCamelCase_ ) @classmethod def a__ (cls ): '''simple docstring''' try: delete_repo(token=cls._token, repo_id='test-config' ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id='valid_org/test-config-org' ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id='test-dynamic-config' ) except HTTPError: pass def a__ (self ): '''simple docstring''' lowerCamelCase__ : Optional[int] = BertConfig( vocab_size=9_9, hidden_size=3_2, num_hidden_layers=5, num_attention_heads=4, intermediate_size=3_7 ) config.push_to_hub('test-config', use_auth_token=self._token ) lowerCamelCase__ : List[str] = BertConfig.from_pretrained(f'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase_, getattr(lowerCamelCase_, lowerCamelCase_ ) ) # Reset repo delete_repo(token=self._token, repo_id='test-config' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase_, repo_id='test-config', push_to_hub=lowerCamelCase_, use_auth_token=self._token ) lowerCamelCase__ : List[Any] = BertConfig.from_pretrained(f'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase_, getattr(lowerCamelCase_, lowerCamelCase_ ) ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : List[Any] = BertConfig( vocab_size=9_9, hidden_size=3_2, num_hidden_layers=5, num_attention_heads=4, intermediate_size=3_7 ) config.push_to_hub('valid_org/test-config-org', use_auth_token=self._token ) lowerCamelCase__ : int = BertConfig.from_pretrained('valid_org/test-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase_, getattr(lowerCamelCase_, lowerCamelCase_ ) ) # Reset repo delete_repo(token=self._token, repo_id='valid_org/test-config-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase_, repo_id='valid_org/test-config-org', push_to_hub=lowerCamelCase_, use_auth_token=self._token ) lowerCamelCase__ : Tuple = BertConfig.from_pretrained('valid_org/test-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase_, getattr(lowerCamelCase_, lowerCamelCase_ ) ) def a__ (self ): '''simple docstring''' CustomConfig.register_for_auto_class() lowerCamelCase__ : str = CustomConfig(attribute=4_2 ) config.push_to_hub('test-dynamic-config', use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map, {'AutoConfig': 'custom_configuration.CustomConfig'} ) lowerCamelCase__ : List[str] = AutoConfig.from_pretrained(f'''{USER}/test-dynamic-config''', trust_remote_code=lowerCamelCase_ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__, 'CustomConfig' ) self.assertEqual(new_config.attribute, 4_2 ) class a_ ( unittest.TestCase ): '''simple docstring''' def a__ (self ): '''simple docstring''' lowerCamelCase__ : Tuple = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated lowerCamelCase__ : Union[str, Any] = c.n_embd + 1 # int lowerCamelCase__ : Optional[Any] = c.resid_pdrop + 1.0 # float lowerCamelCase__ : str = not c.scale_attn_weights # bool lowerCamelCase__ : Any = c.summary_type + 'foo' # str c.update_from_string( f'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(lowerCamelCase_, c.n_embd, 'mismatch for key: n_embd' ) self.assertEqual(lowerCamelCase_, c.resid_pdrop, 'mismatch for key: resid_pdrop' ) self.assertEqual(lowerCamelCase_, c.scale_attn_weights, 'mismatch for key: scale_attn_weights' ) self.assertEqual(lowerCamelCase_, c.summary_type, 'mismatch for key: summary_type' ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : Any = PretrainedConfig() lowerCamelCase__ : Union[str, Any] = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCamelCase_, ['is_encoder_decoder', '_name_or_path', '_commit_hash', 'transformers_version'] ) lowerCamelCase__ : str = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCamelCase_, lowerCamelCase_ )] if len(lowerCamelCase_ ) > 0: raise ValueError( 'The following keys are set with the default values in' ' `test_configuration_common.config_common_kwargs` pick another value for them:' f''' {', '.join(lowerCamelCase_ )}.''' ) def a__ (self ): '''simple docstring''' with self.assertRaises(lowerCamelCase_ ): # config is in subfolder, the following should not work without specifying the subfolder lowerCamelCase__ : Union[str, Any] = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert-subfolder' ) lowerCamelCase__ : Any = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert-subfolder', subfolder='bert' ) self.assertIsNotNone(lowerCamelCase_ ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : int = mock.Mock() lowerCamelCase__ : str = 5_0_0 lowerCamelCase__ : Union[str, Any] = {} lowerCamelCase__ : Any = HTTPError lowerCamelCase__ : str = {} # Download this model to make sure it's in the cache. lowerCamelCase__ : Dict = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('requests.Session.request', return_value=lowerCamelCase_ ) as mock_head: lowerCamelCase__ : Union[str, Any] = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() def a__ (self ): '''simple docstring''' lowerCamelCase__ : int = BertConfig.from_pretrained( 'https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json' ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : str = AutoConfig.from_pretrained('bert-base-cased' ) lowerCamelCase__ : Optional[Any] = ['config.4.0.0.json'] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCamelCase_ ) lowerCamelCase__ : Tuple = 2 json.dump(configuration.to_dict(), open(os.path.join(lowerCamelCase_, 'config.4.0.0.json' ), 'w' ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 lowerCamelCase__ : List[str] = AutoConfig.from_pretrained(lowerCamelCase_ ) self.assertEqual(new_configuration.hidden_size, 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 lowerCamelCase__ : Optional[Any] = ['config.42.0.0.json'] lowerCamelCase__ : List[Any] = 7_6_8 configuration.save_pretrained(lowerCamelCase_ ) shutil.move(os.path.join(lowerCamelCase_, 'config.4.0.0.json' ), os.path.join(lowerCamelCase_, 'config.42.0.0.json' ) ) lowerCamelCase__ : str = AutoConfig.from_pretrained(lowerCamelCase_ ) self.assertEqual(new_configuration.hidden_size, 7_6_8 ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : int = 'hf-internal-testing/test-two-configs' import transformers as new_transformers lowerCamelCase__ : Dict = 'v4.0.0' lowerCamelCase__ , lowerCamelCase__ : str = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCamelCase_, return_unused_kwargs=lowerCamelCase_ ) self.assertEqual(new_configuration.hidden_size, 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCamelCase_, {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers lowerCamelCase__ : Optional[Any] = 'v3.0.0' lowerCamelCase__ : Optional[int] = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCamelCase_ ) self.assertEqual(old_configuration.hidden_size, 7_6_8 )
316
1
import os import textwrap import pyarrow as pa import pytest from datasets import ClassLabel, Features, Image from datasets.packaged_modules.csv.csv import Csv from ..utils import require_pil @pytest.fixture def lowerCAmelCase_ ( __a ) -> int: """simple docstring""" lowerCamelCase__: Optional[int] =tmp_path / "file.csv" lowerCamelCase__: Optional[int] =textwrap.dedent( "\\n header1,header2\n 1,2\n 10,20\n " ) with open(__a , "w" ) as f: f.write(__a ) return str(__a ) @pytest.fixture def lowerCAmelCase_ ( __a ) -> Optional[Any]: """simple docstring""" lowerCamelCase__: str =tmp_path / "malformed_file.csv" lowerCamelCase__: List[Any] =textwrap.dedent( "\\n header1,header2\n 1,2\n 10,20,\n " ) with open(__a , "w" ) as f: f.write(__a ) return str(__a ) @pytest.fixture def lowerCAmelCase_ ( __a , __a ) -> Optional[int]: """simple docstring""" lowerCamelCase__: Tuple =tmp_path / "csv_with_image.csv" lowerCamelCase__: str =textwrap.dedent( F"""\ image {image_file} """ ) with open(__a , "w" ) as f: f.write(__a ) return str(__a ) @pytest.fixture def lowerCAmelCase_ ( __a ) -> Optional[int]: """simple docstring""" lowerCamelCase__: Tuple =tmp_path / "csv_with_label.csv" lowerCamelCase__: Optional[Any] =textwrap.dedent( "\\n label\n good\n bad\n good\n " ) with open(__a , "w" ) as f: f.write(__a ) return str(__a ) @pytest.fixture def lowerCAmelCase_ ( __a ) -> List[Any]: """simple docstring""" lowerCamelCase__: Dict =tmp_path / "csv_with_int_list.csv" lowerCamelCase__: str =textwrap.dedent( "\\n int_list\n 1 2 3\n 4 5 6\n 7 8 9\n " ) with open(__a , "w" ) as f: f.write(__a ) return str(__a ) def lowerCAmelCase_ ( __a , __a , __a ) -> str: """simple docstring""" lowerCamelCase__: List[Any] =Csv() lowerCamelCase__: Union[str, Any] =csv._generate_tables([[csv_file, malformed_csv_file]] ) with pytest.raises(__a , match="Error tokenizing data" ): for _ in generator: pass assert any( record.levelname == "ERROR" and "Failed to read file" in record.message and os.path.basename(__a ) in record.message for record in caplog.records ) @require_pil def lowerCAmelCase_ ( __a ) -> Dict: """simple docstring""" with open(__a , encoding="utf-8" ) as f: lowerCamelCase__: Optional[Any] =f.read().splitlines()[1] lowerCamelCase__: Any =Csv(encoding="utf-8" , features=Features({"image": Image()} ) ) lowerCamelCase__: int =csv._generate_tables([[csv_file_with_image]] ) lowerCamelCase__: Tuple =pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field("image" ).type == Image()() lowerCamelCase__: List[str] =pa_table.to_pydict()["image"] assert generated_content == [{"path": image_file, "bytes": None}] def lowerCAmelCase_ ( __a ) -> str: """simple docstring""" with open(__a , encoding="utf-8" ) as f: lowerCamelCase__: str =f.read().splitlines()[1:] lowerCamelCase__: Any =Csv(encoding="utf-8" , features=Features({"label": ClassLabel(names=["good", "bad"] )} ) ) lowerCamelCase__: Optional[int] =csv._generate_tables([[csv_file_with_label]] ) lowerCamelCase__: Dict =pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field("label" ).type == ClassLabel(names=["good", "bad"] )() lowerCamelCase__: Optional[int] =pa_table.to_pydict()["label"] assert generated_content == [ClassLabel(names=["good", "bad"] ).straint(__a ) for label in labels] def lowerCAmelCase_ ( __a ) -> List[str]: """simple docstring""" lowerCamelCase__: List[Any] =Csv(encoding="utf-8" , sep="," , converters={"int_list": lambda __a : [int(__a ) for i in x.split()]} ) lowerCamelCase__: Tuple =csv._generate_tables([[csv_file_with_int_list]] ) lowerCamelCase__: Dict =pa.concat_tables([table for _, table in generator] ) assert pa.types.is_list(pa_table.schema.field("int_list" ).type ) lowerCamelCase__: Any =pa_table.to_pydict()["int_list"] assert generated_content == [[1, 2, 3], [4, 5, 6], [7, 8, 9]]
10
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer __A = logging.get_logger(__name__) __A = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __A = { "vocab_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json" ), }, } __A = { "yjernite/retribert-base-uncased": 512, } __A = { "yjernite/retribert-base-uncased": {"do_lower_case": True}, } class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = PRETRAINED_INIT_CONFIGURATION lowercase_ = RetriBertTokenizer lowercase_ = ["input_ids", "attention_mask"] def __init__(self : int , UpperCAmelCase_ : List[str]=None , UpperCAmelCase_ : str=None , UpperCAmelCase_ : Optional[Any]=True , UpperCAmelCase_ : Union[str, Any]="[UNK]" , UpperCAmelCase_ : Any="[SEP]" , UpperCAmelCase_ : List[str]="[PAD]" , UpperCAmelCase_ : Optional[Any]="[CLS]" , UpperCAmelCase_ : Optional[Any]="[MASK]" , UpperCAmelCase_ : Dict=True , UpperCAmelCase_ : str=None , **UpperCAmelCase_ : str , ) ->List[Any]: '''simple docstring''' super().__init__( UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , tokenize_chinese_chars=UpperCAmelCase_ , strip_accents=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCamelCase__: List[Any] =json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get("lowercase" , UpperCAmelCase_) != do_lower_case or normalizer_state.get("strip_accents" , UpperCAmelCase_) != strip_accents or normalizer_state.get("handle_chinese_chars" , UpperCAmelCase_) != tokenize_chinese_chars ): lowerCamelCase__: Dict =getattr(UpperCAmelCase_ , normalizer_state.pop("type")) lowerCamelCase__: int =do_lower_case lowerCamelCase__: int =strip_accents lowerCamelCase__: List[str] =tokenize_chinese_chars lowerCamelCase__: Tuple =normalizer_class(**UpperCAmelCase_) lowerCamelCase__: Any =do_lower_case def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any]=None) ->List[str]: '''simple docstring''' lowerCamelCase__: Optional[Any] =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def SCREAMING_SNAKE_CASE_ (self : Dict , UpperCAmelCase_ : List[int] , UpperCAmelCase_ : Optional[List[int]] = None) ->List[int]: '''simple docstring''' lowerCamelCase__: Tuple =[self.sep_token_id] lowerCamelCase__: Optional[int] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : str , UpperCAmelCase_ : Optional[str] = None) ->Tuple[str]: '''simple docstring''' lowerCamelCase__: Tuple =self._tokenizer.model.save(UpperCAmelCase_ , name=UpperCAmelCase_) return tuple(UpperCAmelCase_)
10
1
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): lowerCAmelCase_ = 'pt' elif is_tf_available(): lowerCAmelCase_ = 'tf' else: lowerCAmelCase_ = 'jax' class _A ( _lowerCamelCase , unittest.TestCase ): _UpperCamelCase : Optional[int] = PerceiverTokenizer _UpperCamelCase : Any = False def __a ( self : int ) -> Optional[Any]: """simple docstring""" super().setUp() lowercase : int = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def __a ( self : Tuple ) -> Union[str, Any]: """simple docstring""" return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def __a ( self : List[str] , **_A : List[str] ) -> PerceiverTokenizer: """simple docstring""" return self.tokenizer_class.from_pretrained(self.tmpdirname , **_A ) def __a ( self : Optional[Any] , _A : Optional[Any] , _A : List[str]=False , _A : Tuple=20 , _A : List[Any]=5 ) -> Tuple[str, list]: """simple docstring""" lowercase : List[Any] = [] for i in range(len(_A ) ): try: lowercase : Dict = tokenizer.decode([i] , clean_up_tokenization_spaces=_A ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowercase : Optional[int] = list(filter(lambda _A : re.match(r'''^[ a-zA-Z]+$''' , t[1] ) , _A ) ) lowercase : Tuple = list(filter(lambda _A : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=_A ) , _A ) ) if max_length is not None and len(_A ) > max_length: lowercase : str = toks[:max_length] if min_length is not None and len(_A ) < min_length and len(_A ) > 0: while len(_A ) < min_length: lowercase : str = toks + toks # toks_str = [t[1] for t in toks] lowercase : Optional[Any] = [t[0] for t in toks] # Ensure consistency lowercase : int = tokenizer.decode(_A , clean_up_tokenization_spaces=_A ) if " " not in output_txt and len(_A ) > 1: lowercase : Dict = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=_A ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=_A ) ) if with_prefix_space: lowercase : Optional[Any] = ''' ''' + output_txt lowercase : List[str] = tokenizer.encode(_A , add_special_tokens=_A ) return output_txt, output_ids def __a ( self : str ) -> Dict: """simple docstring""" lowercase : Optional[int] = self.perceiver_tokenizer lowercase : List[Any] = '''Unicode €.''' lowercase : Dict = tokenizer(_A ) lowercase : int = [4, 91, 116, 111, 105, 117, 106, 107, 38, 232, 136, 178, 52, 5] self.assertEqual(encoded['''input_ids'''] , _A ) # decoding lowercase : List[Any] = tokenizer.decode(_A ) self.assertEqual(_A , '''[CLS]Unicode €.[SEP]''' ) lowercase : Union[str, Any] = tokenizer('''e è é ê ë''' ) lowercase : Optional[int] = [4, 107, 38, 201, 174, 38, 201, 175, 38, 201, 176, 38, 201, 177, 5] self.assertEqual(encoded['''input_ids'''] , _A ) # decoding lowercase : Dict = tokenizer.decode(_A ) self.assertEqual(_A , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def __a ( self : Dict ) -> int: """simple docstring""" lowercase : Union[str, Any] = self.perceiver_tokenizer lowercase : Optional[Any] = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off lowercase : Dict = [4, 71, 38, 114, 117, 116, 109, 38, 118, 103, 120, 103, 109, 120, 103, 118, 110, 38, 108, 117, 120, 38, 121, 123, 115, 115, 103, 120, 111, 128, 103, 122, 111, 117, 116, 52, 5, 0] # fmt: on lowercase : int = tokenizer(_A , padding=_A , return_tensors=_A ) self.assertIsInstance(_A , _A ) if FRAMEWORK != "jax": lowercase : int = list(batch.input_ids.numpy()[0] ) else: lowercase : Optional[int] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(_A , _A ) self.assertEqual((2, 38) , batch.input_ids.shape ) self.assertEqual((2, 38) , batch.attention_mask.shape ) def __a ( self : List[Any] ) -> Optional[Any]: """simple docstring""" lowercase : str = self.perceiver_tokenizer lowercase : str = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase : Dict = tokenizer(_A , padding=_A , return_tensors=_A ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , _A ) self.assertIn('''attention_mask''' , _A ) self.assertNotIn('''decoder_input_ids''' , _A ) self.assertNotIn('''decoder_attention_mask''' , _A ) def __a ( self : str ) -> str: """simple docstring""" lowercase : Tuple = self.perceiver_tokenizer lowercase : str = [ '''Summary of the text.''', '''Another summary.''', ] lowercase : Optional[Any] = tokenizer( text_target=_A , max_length=32 , padding='''max_length''' , truncation=_A , return_tensors=_A ) self.assertEqual(32 , targets['''input_ids'''].shape[1] ) def __a ( self : Any ) -> Union[str, Any]: """simple docstring""" lowercase : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test lowercase : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowercase : Any = tempfile.mkdtemp() lowercase : Optional[Any] = ''' He is very happy, UNwant\u00E9d,running''' lowercase : Dict = tokenizer.encode(_A , add_special_tokens=_A ) tokenizer.save_pretrained(_A ) lowercase : Optional[int] = tokenizer.__class__.from_pretrained(_A ) lowercase : Tuple = after_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) shutil.rmtree(_A ) lowercase : List[str] = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc lowercase : Optional[Any] = tempfile.mkdtemp() lowercase : Optional[int] = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) lowercase : Dict = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) lowercase : Union[str, Any] = tokenizer.encode(_A , add_special_tokens=_A ) tokenizer.save_pretrained(_A ) lowercase : int = tokenizer.__class__.from_pretrained(_A ) lowercase : Any = after_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) lowercase : int = tokenizer.__class__.from_pretrained(_A , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(_A ) def __a ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" lowercase : Optional[int] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_A ) with open(os.path.join(_A , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: lowercase : Any = json.load(_A ) with open(os.path.join(_A , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: lowercase : int = json.load(_A ) lowercase : Any = [f"""<extra_id_{i}>""" for i in range(125 )] lowercase : int = added_tokens_extra_ids + [ '''an_additional_special_token''' ] lowercase : Optional[Any] = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(_A , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(_A , _A ) with open(os.path.join(_A , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(_A , _A ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowercase : Optional[int] = tokenizer_class.from_pretrained( _A , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowercase : Any = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=_A )] lowercase : Optional[int] = tokenizer_class.from_pretrained( _A , additional_special_tokens=_A , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def __a ( self : Any ) -> Optional[int]: """simple docstring""" lowercase : Optional[int] = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([178] ) , '''�''' ) def __a ( self : Any ) -> Union[str, Any]: """simple docstring""" pass def __a ( self : int ) -> Dict: """simple docstring""" pass def __a ( self : List[str] ) -> Tuple: """simple docstring""" pass def __a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" pass def __a ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowercase : Optional[Any] = self.get_tokenizers(fast=_A , do_lower_case=_A ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): lowercase : Tuple = ['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] lowercase : str = tokenizer.convert_tokens_to_string(_A ) self.assertIsInstance(_A , _A )
116
from __future__ import annotations import time lowerCAmelCase_ = list[tuple[int, int]] lowerCAmelCase_ = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] lowerCAmelCase_ = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class _A : def __init__( self : List[Any] , _A : int , _A : int , _A : int , _A : int , _A : Node | None ) -> str: """simple docstring""" lowercase : Any = pos_x lowercase : Union[str, Any] = pos_y lowercase : int = (pos_y, pos_x) lowercase : List[str] = goal_x lowercase : Optional[int] = goal_y lowercase : Union[str, Any] = parent class _A : def __init__( self : Optional[Any] , _A : tuple[int, int] , _A : tuple[int, int] ) -> List[Any]: """simple docstring""" lowercase : int = Node(start[1] , start[0] , goal[1] , goal[0] , _A ) lowercase : int = Node(goal[1] , goal[0] , goal[1] , goal[0] , _A ) lowercase : List[str] = [self.start] lowercase : Optional[int] = False def __a ( self : Dict ) -> Path | None: """simple docstring""" while self.node_queue: lowercase : Optional[Any] = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: lowercase : int = True return self.retrace_path(_A ) lowercase : Union[str, Any] = self.get_successors(_A ) for node in successors: self.node_queue.append(_A ) if not self.reached: return [self.start.pos] return None def __a ( self : str , _A : Node ) -> list[Node]: """simple docstring""" lowercase : Union[str, Any] = [] for action in delta: lowercase : List[str] = parent.pos_x + action[1] lowercase : str = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(_A ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(_A , _A , self.target.pos_y , self.target.pos_x , _A ) ) return successors def __a ( self : Any , _A : Node | None ) -> Path: """simple docstring""" lowercase : Optional[int] = node lowercase : Dict = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) lowercase : Tuple = current_node.parent path.reverse() return path class _A : def __init__( self : Optional[int] , _A : List[str] , _A : List[str] ) -> int: """simple docstring""" lowercase : str = BreadthFirstSearch(_A , _A ) lowercase : Optional[Any] = BreadthFirstSearch(_A , _A ) lowercase : List[str] = False def __a ( self : Tuple ) -> Path | None: """simple docstring""" while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: lowercase : str = self.fwd_bfs.node_queue.pop(0 ) lowercase : Tuple = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: lowercase : Optional[Any] = True return self.retrace_bidirectional_path( _A , _A ) lowercase : Optional[int] = current_bwd_node lowercase : Optional[int] = current_fwd_node lowercase : Union[str, Any] = { self.fwd_bfs: self.fwd_bfs.get_successors(_A ), self.bwd_bfs: self.bwd_bfs.get_successors(_A ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(_A ) if not self.reached: return [self.fwd_bfs.start.pos] return None def __a ( self : List[Any] , _A : Node , _A : Node ) -> Path: """simple docstring""" lowercase : Optional[Any] = self.fwd_bfs.retrace_path(_A ) lowercase : Union[str, Any] = self.bwd_bfs.retrace_path(_A ) bwd_path.pop() bwd_path.reverse() lowercase : Optional[int] = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() lowerCAmelCase_ = (0, 0) lowerCAmelCase_ = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) lowerCAmelCase_ = time.time() lowerCAmelCase_ = BreadthFirstSearch(init, goal) lowerCAmelCase_ = bfs.search() lowerCAmelCase_ = time.time() - start_bfs_time print('Unidirectional BFS computation time : ', bfs_time) lowerCAmelCase_ = time.time() lowerCAmelCase_ = BidirectionalBreadthFirstSearch(init, goal) lowerCAmelCase_ = bd_bfs.search() lowerCAmelCase_ = time.time() - start_bd_bfs_time print('Bidirectional BFS computation time : ', bd_bfs_time)
116
1
from graphs.minimum_spanning_tree_kruskal import kruskal def lowerCAmelCase_ ( ) -> str: """simple docstring""" lowerCamelCase__: Optional[int] =9 lowerCamelCase__: Union[str, Any] =[ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] lowerCamelCase__: Optional[Any] =kruskal(__a , __a ) lowerCamelCase__: str =[ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] assert sorted(__a ) == sorted(__a )
10
import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType UpperCamelCase = None UpperCamelCase = """<""" if sys.byteorder == """little""" else """>""" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image UpperCamelCase = [ np.dtype("""|b1"""), np.dtype("""|u1"""), np.dtype("""<u2"""), np.dtype(""">u2"""), np.dtype("""<i2"""), np.dtype(""">i2"""), np.dtype("""<u4"""), np.dtype(""">u4"""), np.dtype("""<i4"""), np.dtype(""">i4"""), np.dtype("""<f4"""), np.dtype(""">f4"""), np.dtype("""<f8"""), np.dtype(""">f8"""), ] @dataclass class _lowerCamelCase : """simple docstring""" snake_case = True snake_case = None # Automatically constructed snake_case = "PIL.Image.Image" snake_case = pa.struct({"bytes": pa.binary(), "path": pa.string()} ) snake_case = field(default="Image" , init=UpperCamelCase , repr=UpperCamelCase ) def __call__( self )->int: '''simple docstring''' return self.pa_type def _snake_case ( self , _SCREAMING_SNAKE_CASE )->dict: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): A_ : Optional[int] = np.array(_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return {"path": value, "bytes": None} elif isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return {"path": None, "bytes": value} elif isinstance(_SCREAMING_SNAKE_CASE , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(_SCREAMING_SNAKE_CASE ) elif isinstance(_SCREAMING_SNAKE_CASE , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(_SCREAMING_SNAKE_CASE ) elif value.get('''path''' ) is not None and os.path.isfile(value['''path'''] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get('''path''' )} elif value.get('''bytes''' ) is not None or value.get('''path''' ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get('''bytes''' ), "path": value.get('''path''' )} else: raise ValueError( F'''An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None )->"PIL.Image.Image": '''simple docstring''' if not self.decode: raise RuntimeError('''Decoding is disabled for this feature. Please use Image(decode=True) instead.''' ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support decoding images, please install \'Pillow\'.''' ) if token_per_repo_id is None: A_ : List[str] = {} A_ , A_ : str = value['''path'''], value['''bytes'''] if bytes_ is None: if path is None: raise ValueError(F'''An image should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) else: if is_local_path(_SCREAMING_SNAKE_CASE ): A_ : List[str] = PIL.Image.open(_SCREAMING_SNAKE_CASE ) else: A_ : List[str] = path.split('''::''' )[-1] try: A_ : int = string_to_dict(_SCREAMING_SNAKE_CASE , config.HUB_DATASETS_URL )['''repo_id'''] A_ : Optional[int] = token_per_repo_id.get(_SCREAMING_SNAKE_CASE ) except ValueError: A_ : Any = None with xopen(_SCREAMING_SNAKE_CASE , '''rb''' , use_auth_token=_SCREAMING_SNAKE_CASE ) as f: A_ : Optional[Any] = BytesIO(f.read() ) A_ : Dict = PIL.Image.open(bytes_ ) else: A_ : Optional[int] = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def _snake_case ( self )->Union["FeatureType", Dict[str, "FeatureType"]]: '''simple docstring''' from .features import Value return ( self if self.decode else { "bytes": Value('''binary''' ), "path": Value('''string''' ), } ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )->pa.StructArray: '''simple docstring''' if pa.types.is_string(storage.type ): A_ : Dict = pa.array([None] * len(_SCREAMING_SNAKE_CASE ) , type=pa.binary() ) A_ : List[Any] = pa.StructArray.from_arrays([bytes_array, storage] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): A_ : Dict = pa.array([None] * len(_SCREAMING_SNAKE_CASE ) , type=pa.string() ) A_ : List[str] = pa.StructArray.from_arrays([storage, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index('''bytes''' ) >= 0: A_ : Tuple = storage.field('''bytes''' ) else: A_ : Optional[int] = pa.array([None] * len(_SCREAMING_SNAKE_CASE ) , type=pa.binary() ) if storage.type.get_field_index('''path''' ) >= 0: A_ : Optional[Any] = storage.field('''path''' ) else: A_ : Optional[int] = pa.array([None] * len(_SCREAMING_SNAKE_CASE ) , type=pa.string() ) A_ : Union[str, Any] = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): A_ : Optional[Any] = pa.array( [encode_np_array(np.array(_SCREAMING_SNAKE_CASE ) )['''bytes'''] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) A_ : str = pa.array([None] * len(_SCREAMING_SNAKE_CASE ) , type=pa.string() ) A_ : List[Any] = pa.StructArray.from_arrays( [bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null() ) return array_cast(_SCREAMING_SNAKE_CASE , self.pa_type ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )->pa.StructArray: '''simple docstring''' @no_op_if_value_is_null def path_to_bytes(_SCREAMING_SNAKE_CASE ): with xopen(_SCREAMING_SNAKE_CASE , '''rb''' ) as f: A_ : Any = f.read() return bytes_ A_ : Dict = pa.array( [ (path_to_bytes(x['''path'''] ) if x['''bytes'''] is None else x['''bytes''']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) A_ : List[Any] = pa.array( [os.path.basename(_SCREAMING_SNAKE_CASE ) if path is not None else None for path in storage.field('''path''' ).to_pylist()] , type=pa.string() , ) A_ : str = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null() ) return array_cast(_SCREAMING_SNAKE_CASE , self.pa_type ) def _SCREAMING_SNAKE_CASE ( ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() A_ : Dict = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): A_ : Dict = BytesIO() if image.format in list_image_compression_formats(): A_ : Tuple = image.format else: A_ : List[str] = '''PNG''' if image.mode in ['''1''', '''L''', '''LA''', '''RGB''', '''RGBA'''] else '''TIFF''' image.save(SCREAMING_SNAKE_CASE , format=SCREAMING_SNAKE_CASE ) return buffer.getvalue() def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): if hasattr(SCREAMING_SNAKE_CASE , '''filename''' ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(SCREAMING_SNAKE_CASE )} def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) A_ : Union[str, Any] = array.dtype A_ : Dict = dtype.byteorder if dtype.byteorder != '''=''' else _NATIVE_BYTEORDER A_ : Any = dtype.kind A_ : Any = dtype.itemsize A_ : Dict = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: A_ : List[Any] = np.dtype('''|u1''' ) if dtype_kind not in ["u", "i"]: raise TypeError( f'''Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.''' ) if dtype is not dest_dtype: warnings.warn(f'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: A_ : int = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: A_ : Any = dtype_byteorder + dtype_kind + str(SCREAMING_SNAKE_CASE ) A_ : Optional[int] = np.dtype(SCREAMING_SNAKE_CASE ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(f'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( f'''Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}''' ) A_ : Tuple = PIL.Image.fromarray(array.astype(SCREAMING_SNAKE_CASE ) ) return {"path": None, "bytes": image_to_bytes(SCREAMING_SNAKE_CASE )} def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) if objs: A_ , A_ : Union[str, Any] = first_non_null_value(SCREAMING_SNAKE_CASE ) if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(SCREAMING_SNAKE_CASE , np.ndarray ): A_ : Tuple = no_op_if_value_is_null(SCREAMING_SNAKE_CASE ) return [obj_to_image_dict_func(SCREAMING_SNAKE_CASE ) for obj in objs] elif isinstance(SCREAMING_SNAKE_CASE , PIL.Image.Image ): A_ : List[str] = no_op_if_value_is_null(SCREAMING_SNAKE_CASE ) return [obj_to_image_dict_func(SCREAMING_SNAKE_CASE ) for obj in objs] else: return objs else: return objs
186
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() _lowerCAmelCase : Union[str, Any] = logging.get_logger(__name__) _lowerCAmelCase : Optional[Any] = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def UpperCamelCase_( _snake_case : Optional[int] , _snake_case : Any , _snake_case : Optional[int] , _snake_case : str , _snake_case : List[Any] ): """simple docstring""" for attribute in key.split('.' ): __a =getattr(_snake_case , _snake_case ) if weight_type is not None: __a =getattr(_snake_case , _snake_case ).shape else: __a =hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": __a =value elif weight_type == "weight_g": __a =value elif weight_type == "weight_v": __a =value elif weight_type == "bias": __a =value else: __a =value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def UpperCamelCase_( _snake_case : Tuple , _snake_case : str , _snake_case : int ): """simple docstring""" __a =[] __a =fairseq_model.state_dict() __a =hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): __a =False if "conv_layers" in name: load_conv_layer( _snake_case , _snake_case , _snake_case , _snake_case , hf_model.config.feat_extract_norm == 'group' , ) __a =True else: for key, mapped_key in MAPPING.items(): __a ='sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: __a =True if "*" in mapped_key: __a =name.split(_snake_case )[0].split('.' )[-2] __a =mapped_key.replace('*' , _snake_case ) if "weight_g" in name: __a ='weight_g' elif "weight_v" in name: __a ='weight_v' elif "weight" in name: __a ='weight' elif "bias" in name: __a ='bias' else: __a =None set_recursively(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) continue if not is_used: unused_weights.append(_snake_case ) logger.warning(F'Unused weights: {unused_weights}' ) def UpperCamelCase_( _snake_case : List[str] , _snake_case : Any , _snake_case : Optional[Any] , _snake_case : int , _snake_case : Dict ): """simple docstring""" __a =full_name.split('conv_layers.' )[-1] __a =name.split('.' ) __a =int(items[0] ) __a =int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) __a =value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) __a =value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) __a =value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) __a =value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(_snake_case ) def UpperCamelCase_( _snake_case : Union[str, Any] , _snake_case : Any ): """simple docstring""" __a =SEWConfig() if is_finetuned: __a =model.wav_encoder.wav_model.cfg else: __a =model.cfg __a =fs_config.conv_bias __a =eval(fs_config.conv_feature_layers ) __a =[x[0] for x in conv_layers] __a =[x[1] for x in conv_layers] __a =[x[2] for x in conv_layers] __a ='gelu' __a ='layer' if fs_config.extractor_mode == 'layer_norm' else 'group' __a =0.0 __a =fs_config.activation_fn.name __a =fs_config.encoder_embed_dim __a =0.02 __a =fs_config.encoder_ffn_embed_dim __a =1e-5 __a =fs_config.encoder_layerdrop __a =fs_config.encoder_attention_heads __a =fs_config.conv_pos_groups __a =fs_config.conv_pos __a =len(_snake_case ) __a =fs_config.encoder_layers __a =fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: __a =model.cfg __a =fs_config.final_dropout __a =fs_config.layerdrop __a =fs_config.activation_dropout __a =fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 __a =fs_config.attention_dropout __a =fs_config.dropout_input __a =fs_config.dropout __a =fs_config.mask_channel_length __a =fs_config.mask_channel_prob __a =fs_config.mask_length __a =fs_config.mask_prob __a ='Wav2Vec2FeatureExtractor' __a ='Wav2Vec2CTCTokenizer' return config @torch.no_grad() def UpperCamelCase_( _snake_case : str , _snake_case : Optional[int] , _snake_case : List[Any]=None , _snake_case : Optional[int]=None , _snake_case : Dict=True ): """simple docstring""" if is_finetuned: __a , __a , __a =fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: __a , __a , __a =fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: __a =SEWConfig.from_pretrained(_snake_case ) else: __a =convert_config(model[0] , _snake_case ) __a =model[0].eval() __a =True if config.feat_extract_norm == 'layer' else False __a =WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_snake_case , return_attention_mask=_snake_case , ) if is_finetuned: if dict_path: __a =Dictionary.load(_snake_case ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __a =target_dict.pad_index __a =target_dict.bos_index __a =target_dict.pad_index __a =target_dict.bos_index __a =target_dict.eos_index __a =len(target_dict.symbols ) __a =os.path.join(_snake_case , 'vocab.json' ) if not os.path.isdir(_snake_case ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(_snake_case ) ) return os.makedirs(_snake_case , exist_ok=_snake_case ) with open(_snake_case , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , _snake_case ) __a =WavaVecaCTCTokenizer( _snake_case , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=_snake_case , ) __a =WavaVecaProcessor(feature_extractor=_snake_case , tokenizer=_snake_case ) processor.save_pretrained(_snake_case ) __a =SEWForCTC(_snake_case ) else: __a =SEWModel(_snake_case ) feature_extractor.save_pretrained(_snake_case ) recursively_load_weights(_snake_case , _snake_case , _snake_case ) hf_model.save_pretrained(_snake_case ) if __name__ == "__main__": _lowerCAmelCase : str = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) _lowerCAmelCase : str = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
308
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : Any = logging.get_logger(__name__) _lowerCAmelCase : int = { "caidas/swin2sr-classicalsr-x2-64": ( "https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json" ), } class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = 'swin2sr' SCREAMING_SNAKE_CASE = { 'hidden_size': 'embed_dim', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , __snake_case=64 , __snake_case=1 , __snake_case=3 , __snake_case=180 , __snake_case=[6, 6, 6, 6, 6, 6] , __snake_case=[6, 6, 6, 6, 6, 6] , __snake_case=8 , __snake_case=2.0 , __snake_case=True , __snake_case=0.0 , __snake_case=0.0 , __snake_case=0.1 , __snake_case="gelu" , __snake_case=False , __snake_case=0.02 , __snake_case=1e-5 , __snake_case=2 , __snake_case=1.0 , __snake_case="1conv" , __snake_case="pixelshuffle" , **__snake_case , ) -> Dict: '''simple docstring''' super().__init__(**__snake_case ) __a =image_size __a =patch_size __a =num_channels __a =embed_dim __a =depths __a =len(__snake_case ) __a =num_heads __a =window_size __a =mlp_ratio __a =qkv_bias __a =hidden_dropout_prob __a =attention_probs_dropout_prob __a =drop_path_rate __a =hidden_act __a =use_absolute_embeddings __a =layer_norm_eps __a =initializer_range __a =upscale __a =img_range __a =resi_connection __a =upsampler
308
1
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def A__ ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = s.rsplit(snake_case__, snake_case__ ) return new.join(snake_case__ ) def A__ ( __lowerCamelCase ): return sum(param.float().sum() if '''encoder.embeddings''' not in key else 0 for key, param in state_dict.items() ) def A__ ( __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = {} SCREAMING_SNAKE_CASE_ = ['''group_1''', '''group_2''', '''group_3''', '''group_4'''] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: SCREAMING_SNAKE_CASE_ = key.replace(F'''{group_key}.''', F'''{group_key}.group.''' ) if "res_path" in key: SCREAMING_SNAKE_CASE_ = key.replace('''res_path.''', '''res_path.path.''' ) if key.endswith('''.w''' ): SCREAMING_SNAKE_CASE_ = rreplace(snake_case__, '''.w''', '''.weight''', 1 ) if key.endswith('''.b''' ): SCREAMING_SNAKE_CASE_ = rreplace(snake_case__, '''.b''', '''.bias''', 1 ) SCREAMING_SNAKE_CASE_ = value.float() return upgrade @torch.no_grad() def A__ ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=None, __lowerCamelCase=True ): from dall_e import Encoder SCREAMING_SNAKE_CASE_ = Encoder() if os.path.exists(snake_case__ ): SCREAMING_SNAKE_CASE_ = torch.load(snake_case__ ) else: SCREAMING_SNAKE_CASE_ = torch.hub.load_state_dict_from_url(snake_case__ ) if isinstance(snake_case__, snake_case__ ): SCREAMING_SNAKE_CASE_ = ckpt.state_dict() encoder.load_state_dict(snake_case__ ) if config_path is not None: SCREAMING_SNAKE_CASE_ = FlavaImageCodebookConfig.from_pretrained(snake_case__ ) else: SCREAMING_SNAKE_CASE_ = FlavaImageCodebookConfig() SCREAMING_SNAKE_CASE_ = FlavaImageCodebook(snake_case__ ).eval() SCREAMING_SNAKE_CASE_ = encoder.state_dict() SCREAMING_SNAKE_CASE_ = upgrade_state_dict(snake_case__ ) hf_model.load_state_dict(snake_case__ ) SCREAMING_SNAKE_CASE_ = hf_model.state_dict() SCREAMING_SNAKE_CASE_ = count_parameters(snake_case__ ) SCREAMING_SNAKE_CASE_ = count_parameters(snake_case__ ) assert torch.allclose(snake_case__, snake_case__, atol=1E-3 ) if save_checkpoint: hf_model.save_pretrained(snake_case__ ) else: return hf_state_dict if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to flava checkpoint") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") __UpperCAmelCase = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
299
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> Optional[int]: """simple docstring""" assert torch_layer.weight.shape == weight.shape, F'{torch_layer} layer.weight does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'{torch_layer} layer.bias does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = weights[0][0][0] _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[0] ) _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # lsh weights + output _SCREAMING_SNAKE_CASE = weights[0][1] if len(snake_case__ ) < 4: set_layer_weights_in_torch_lsh(snake_case__ ,torch_block.attention ,snake_case__ ) else: set_layer_weights_in_torch_local(snake_case__ ,torch_block.attention ,snake_case__ ) # intermediate weighs _SCREAMING_SNAKE_CASE = weights[2][0][1][2] # Chunked Feed Forward if len(snake_case__ ) == 4: _SCREAMING_SNAKE_CASE = intermediate_weights[2] # layernorm 2 _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # intermediate dense _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) # intermediate out _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = torch_model.reformer # word embeds _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings ,torch.tensor(snake_case__ ) ,) if isinstance(weights[3] ,snake_case__ ): _SCREAMING_SNAKE_CASE = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _SCREAMING_SNAKE_CASE = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'{position_embeddings[emb_idx]} emb does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(torch.tensor(snake_case__ ) ) _SCREAMING_SNAKE_CASE = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( snake_case__ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _SCREAMING_SNAKE_CASE = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(snake_case__ ,snake_case__ ,snake_case__ ) # output layer norm _SCREAMING_SNAKE_CASE = np.asarray(weights[7][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # output embeddings _SCREAMING_SNAKE_CASE = np.asarray(weights[9][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = ReformerConfig.from_json_file(snake_case__ ) print(F'Building PyTorch model from configuration: {config}' ) _SCREAMING_SNAKE_CASE = ReformerModelWithLMHead(snake_case__ ) with open(snake_case__ ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = pickle.load(snake_case__ )["""weights"""] set_model_weights_in_torch(snake_case__ ,snake_case__ ,config.hidden_size ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() ,snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
306
0
def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase ): if density <= 0: raise ValueError("Impossible fluid density" ) if bulk_modulus <= 0: raise ValueError("Impossible bulk modulus" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
158
from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig _UpperCAmelCase : Tuple = logging.get_logger(__name__) # General docstring _UpperCAmelCase : int = "RegNetConfig" # Base docstring _UpperCAmelCase : Optional[int] = "facebook/regnet-y-040" _UpperCAmelCase : Any = [1, 1088, 7, 7] # Image classification docstring _UpperCAmelCase : Any = "facebook/regnet-y-040" _UpperCAmelCase : List[str] = "tabby, tabby cat" _UpperCAmelCase : Optional[int] = [ "facebook/regnet-y-040", # See all regnet models at https://huggingface.co/models?filter=regnet ] class __lowerCAmelCase ( tf.keras.layers.Layer): def __init__( self: List[str] , _lowerCAmelCase: int , _lowerCAmelCase: int = 3 , _lowerCAmelCase: int = 1 , _lowerCAmelCase: int = 1 , _lowerCAmelCase: Optional[str] = "relu" , **_lowerCAmelCase: Any , ): super().__init__(**_lowerCAmelCase ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb lowercase :List[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) lowercase :Dict = tf.keras.layers.ConvaD( filters=_lowerCAmelCase , kernel_size=_lowerCAmelCase , strides=_lowerCAmelCase , padding="VALID" , groups=_lowerCAmelCase , use_bias=_lowerCAmelCase , name="convolution" , ) lowercase :Tuple = tf.keras.layers.BatchNormalization(epsilon=1e-5 , momentum=0.9 , name="normalization" ) lowercase :Optional[int] = ACTaFN[activation] if activation is not None else tf.identity def SCREAMING_SNAKE_CASE ( self: List[str] , _lowerCAmelCase: Tuple ): lowercase :Dict = self.convolution(self.padding(_lowerCAmelCase ) ) lowercase :Union[str, Any] = self.normalization(_lowerCAmelCase ) lowercase :Union[str, Any] = self.activation(_lowerCAmelCase ) return hidden_state class __lowerCAmelCase ( tf.keras.layers.Layer): def __init__( self: Dict , _lowerCAmelCase: RegNetConfig , **_lowerCAmelCase: Dict ): super().__init__(**_lowerCAmelCase ) lowercase :Optional[Any] = config.num_channels lowercase :str = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: Optional[int] ): lowercase :List[Any] = shape_list(_lowerCAmelCase )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) lowercase :Tuple = tf.transpose(_lowerCAmelCase , perm=(0, 2, 3, 1) ) lowercase :List[Any] = self.embedder(_lowerCAmelCase ) return hidden_state class __lowerCAmelCase ( tf.keras.layers.Layer): def __init__( self: List[Any] , _lowerCAmelCase: int , _lowerCAmelCase: int = 2 , **_lowerCAmelCase: Union[str, Any] ): super().__init__(**_lowerCAmelCase ) lowercase :Union[str, Any] = tf.keras.layers.ConvaD( filters=_lowerCAmelCase , kernel_size=1 , strides=_lowerCAmelCase , use_bias=_lowerCAmelCase , name="convolution" ) lowercase :List[Any] = tf.keras.layers.BatchNormalization(epsilon=1e-5 , momentum=0.9 , name="normalization" ) def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: tf.Tensor , _lowerCAmelCase: bool = False ): return self.normalization(self.convolution(_lowerCAmelCase ) , training=_lowerCAmelCase ) class __lowerCAmelCase ( tf.keras.layers.Layer): def __init__( self: Union[str, Any] , _lowerCAmelCase: int , _lowerCAmelCase: int , **_lowerCAmelCase: Optional[Any] ): super().__init__(**_lowerCAmelCase ) lowercase :Optional[Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=_lowerCAmelCase , name="pooler" ) lowercase :Any = [ tf.keras.layers.ConvaD(filters=_lowerCAmelCase , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=_lowerCAmelCase , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: str ): # [batch_size, h, w, num_channels] -> [batch_size, 1, 1, num_channels] lowercase :List[Any] = self.pooler(_lowerCAmelCase ) for layer_module in self.attention: lowercase :Optional[Any] = layer_module(_lowerCAmelCase ) lowercase :List[Any] = hidden_state * pooled return hidden_state class __lowerCAmelCase ( tf.keras.layers.Layer): def __init__( self: Dict , _lowerCAmelCase: RegNetConfig , _lowerCAmelCase: int , _lowerCAmelCase: int , _lowerCAmelCase: int = 1 , **_lowerCAmelCase: Tuple ): super().__init__(**_lowerCAmelCase ) lowercase :Union[str, Any] = in_channels != out_channels or stride != 1 lowercase :int = max(1 , out_channels // config.groups_width ) lowercase :Any = ( TFRegNetShortCut(_lowerCAmelCase , stride=_lowerCAmelCase , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. lowercase :Optional[int] = [ TFRegNetConvLayer(_lowerCAmelCase , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( _lowerCAmelCase , stride=_lowerCAmelCase , groups=_lowerCAmelCase , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(_lowerCAmelCase , kernel_size=1 , activation=_lowerCAmelCase , name="layer.2" ), ] lowercase :Dict = ACTaFN[config.hidden_act] def SCREAMING_SNAKE_CASE ( self: Optional[int] , _lowerCAmelCase: Optional[int] ): lowercase :List[str] = hidden_state for layer_module in self.layers: lowercase :int = layer_module(_lowerCAmelCase ) lowercase :Optional[int] = self.shortcut(_lowerCAmelCase ) hidden_state += residual lowercase :Optional[int] = self.activation(_lowerCAmelCase ) return hidden_state class __lowerCAmelCase ( tf.keras.layers.Layer): def __init__( self: int , _lowerCAmelCase: RegNetConfig , _lowerCAmelCase: int , _lowerCAmelCase: int , _lowerCAmelCase: int = 1 , **_lowerCAmelCase: Any ): super().__init__(**_lowerCAmelCase ) lowercase :Union[str, Any] = in_channels != out_channels or stride != 1 lowercase :Optional[int] = max(1 , out_channels // config.groups_width ) lowercase :Union[str, Any] = ( TFRegNetShortCut(_lowerCAmelCase , stride=_lowerCAmelCase , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) lowercase :Union[str, Any] = [ TFRegNetConvLayer(_lowerCAmelCase , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( _lowerCAmelCase , stride=_lowerCAmelCase , groups=_lowerCAmelCase , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(_lowerCAmelCase , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(_lowerCAmelCase , kernel_size=1 , activation=_lowerCAmelCase , name="layer.3" ), ] lowercase :Optional[int] = ACTaFN[config.hidden_act] def SCREAMING_SNAKE_CASE ( self: Optional[int] , _lowerCAmelCase: Dict ): lowercase :List[Any] = hidden_state for layer_module in self.layers: lowercase :str = layer_module(_lowerCAmelCase ) lowercase :Tuple = self.shortcut(_lowerCAmelCase ) hidden_state += residual lowercase :str = self.activation(_lowerCAmelCase ) return hidden_state class __lowerCAmelCase ( tf.keras.layers.Layer): def __init__( self: Dict , _lowerCAmelCase: RegNetConfig , _lowerCAmelCase: int , _lowerCAmelCase: int , _lowerCAmelCase: int = 2 , _lowerCAmelCase: int = 2 , **_lowerCAmelCase: Union[str, Any] ): super().__init__(**_lowerCAmelCase ) lowercase :List[str] = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer lowercase :Union[str, Any] = [ # downsampling is done in the first layer with stride of 2 layer(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , stride=_lowerCAmelCase , name="layers.0" ), *[layer(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , name=F"layers.{i+1}" ) for i in range(depth - 1 )], ] def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: Optional[Any] ): for layer_module in self.layers: lowercase :Tuple = layer_module(_lowerCAmelCase ) return hidden_state class __lowerCAmelCase ( tf.keras.layers.Layer): def __init__( self: Union[str, Any] , _lowerCAmelCase: RegNetConfig , **_lowerCAmelCase: Optional[Any] ): super().__init__(**_lowerCAmelCase ) lowercase :Any = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( _lowerCAmelCase , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) lowercase :Dict = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(_lowerCAmelCase , config.depths[1:] ) ): self.stages.append(TFRegNetStage(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , depth=_lowerCAmelCase , name=F"stages.{i+1}" ) ) def SCREAMING_SNAKE_CASE ( self: str , _lowerCAmelCase: tf.Tensor , _lowerCAmelCase: bool = False , _lowerCAmelCase: bool = True ): lowercase :str = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: lowercase :int = hidden_states + (hidden_state,) lowercase :List[Any] = stage_module(_lowerCAmelCase ) if output_hidden_states: lowercase :Union[str, Any] = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=_lowerCAmelCase , hidden_states=_lowerCAmelCase ) @keras_serializable class __lowerCAmelCase ( tf.keras.layers.Layer): _a = RegNetConfig def __init__( self: List[Any] , _lowerCAmelCase: Optional[Any] , **_lowerCAmelCase: Tuple ): super().__init__(**_lowerCAmelCase ) lowercase :List[str] = config lowercase :List[str] = TFRegNetEmbeddings(_lowerCAmelCase , name="embedder" ) lowercase :Optional[Any] = TFRegNetEncoder(_lowerCAmelCase , name="encoder" ) lowercase :Tuple = tf.keras.layers.GlobalAveragePoolingaD(keepdims=_lowerCAmelCase , name="pooler" ) @unpack_inputs def SCREAMING_SNAKE_CASE ( self: int , _lowerCAmelCase: tf.Tensor , _lowerCAmelCase: Optional[bool] = None , _lowerCAmelCase: Optional[bool] = None , _lowerCAmelCase: bool = False , ): lowercase :List[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase :Tuple = return_dict if return_dict is not None else self.config.use_return_dict lowercase :Tuple = self.embedder(_lowerCAmelCase , training=_lowerCAmelCase ) lowercase :Dict = self.encoder( _lowerCAmelCase , output_hidden_states=_lowerCAmelCase , return_dict=_lowerCAmelCase , training=_lowerCAmelCase ) lowercase :str = encoder_outputs[0] lowercase :str = self.pooler(_lowerCAmelCase ) # Change to NCHW output format have uniformity in the modules lowercase :Optional[int] = tf.transpose(_lowerCAmelCase , perm=(0, 3, 1, 2) ) lowercase :int = tf.transpose(_lowerCAmelCase , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: lowercase :Tuple = tuple([tf.transpose(_lowerCAmelCase , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_lowerCAmelCase , pooler_output=_lowerCAmelCase , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class __lowerCAmelCase ( lowerCAmelCase): _a = RegNetConfig _a = '''regnet''' _a = '''pixel_values''' @property def SCREAMING_SNAKE_CASE ( self: Any ): return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_24, 2_24) , dtype=tf.floataa )} _UpperCAmelCase : Optional[int] = r"\n Parameters:\n This model is a Tensorflow\n [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a\n regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and\n behavior.\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights.\n" _UpperCAmelCase : Optional[Any] = r"\n Args:\n pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConveNextImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( '''The bare RegNet model outputting raw features without any specific head on top.''' , lowerCAmelCase , ) class __lowerCAmelCase ( lowerCAmelCase): def __init__( self: Dict , _lowerCAmelCase: RegNetConfig , *_lowerCAmelCase: str , **_lowerCAmelCase: Dict ): super().__init__(_lowerCAmelCase , *_lowerCAmelCase , **_lowerCAmelCase ) lowercase :List[Any] = TFRegNetMainLayer(_lowerCAmelCase , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(_lowerCAmelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_lowerCAmelCase , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def SCREAMING_SNAKE_CASE ( self: int , _lowerCAmelCase: tf.Tensor , _lowerCAmelCase: Optional[bool] = None , _lowerCAmelCase: Optional[bool] = None , _lowerCAmelCase: Any=False , ): lowercase :Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase :Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase :Optional[Any] = self.regnet( pixel_values=_lowerCAmelCase , output_hidden_states=_lowerCAmelCase , return_dict=_lowerCAmelCase , training=_lowerCAmelCase , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( ''' RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. ''' , lowerCAmelCase , ) class __lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase): def __init__( self: List[Any] , _lowerCAmelCase: RegNetConfig , *_lowerCAmelCase: List[Any] , **_lowerCAmelCase: List[Any] ): super().__init__(_lowerCAmelCase , *_lowerCAmelCase , **_lowerCAmelCase ) lowercase :Any = config.num_labels lowercase :int = TFRegNetMainLayer(_lowerCAmelCase , name="regnet" ) # classification head lowercase :List[Any] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(_lowerCAmelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_lowerCAmelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def SCREAMING_SNAKE_CASE ( self: List[Any] , _lowerCAmelCase: tf.Tensor = None , _lowerCAmelCase: tf.Tensor = None , _lowerCAmelCase: bool = None , _lowerCAmelCase: bool = None , _lowerCAmelCase: Tuple=False , ): lowercase :Union[str, Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase :Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase :str = self.regnet( _lowerCAmelCase , output_hidden_states=_lowerCAmelCase , return_dict=_lowerCAmelCase , training=_lowerCAmelCase ) lowercase :str = outputs.pooler_output if return_dict else outputs[1] lowercase :Optional[Any] = self.classifier[0](_lowerCAmelCase ) lowercase :int = self.classifier[1](_lowerCAmelCase ) lowercase :Dict = None if labels is None else self.hf_compute_loss(labels=_lowerCAmelCase , logits=_lowerCAmelCase ) if not return_dict: lowercase :Tuple = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=_lowerCAmelCase , logits=_lowerCAmelCase , hidden_states=outputs.hidden_states )
158
1
import math def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : Any = [] lowerCamelCase : int = 2 lowerCamelCase : Dict = int(math.sqrt(SCREAMING_SNAKE_CASE_ ) ) # Size of every segment lowerCamelCase : Dict = [True] * (end + 1) lowerCamelCase : Optional[int] = [] while start <= end: if temp[start] is True: in_prime.append(SCREAMING_SNAKE_CASE_ ) for i in range(start * start , end + 1 , SCREAMING_SNAKE_CASE_ ): lowerCamelCase : Tuple = False start += 1 prime += in_prime lowerCamelCase : Any = end + 1 lowerCamelCase : Optional[int] = min(2 * end , SCREAMING_SNAKE_CASE_ ) while low <= n: lowerCamelCase : str = [True] * (high - low + 1) for each in in_prime: lowerCamelCase : List[Any] = math.floor(low / each ) * each if t < low: t += each for j in range(SCREAMING_SNAKE_CASE_ , high + 1 , SCREAMING_SNAKE_CASE_ ): lowerCamelCase : str = False for j in range(len(SCREAMING_SNAKE_CASE_ ) ): if temp[j] is True: prime.append(j + low ) lowerCamelCase : Optional[int] = high + 1 lowerCamelCase : Optional[int] = min(high + end , SCREAMING_SNAKE_CASE_ ) return prime print(sieve(10**6))
283
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' __A : Dict = "openai/whisper-base" __A : str = ( "This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the " "transcribed text." ) __A : Any = "transcriber" __A : Any = WhisperProcessor __A : int = WhisperForConditionalGeneration __A : Any = ["audio"] __A : List[str] = ["text"] def _snake_case ( self , __A ): """simple docstring""" return self.pre_processor(__A , return_tensors="pt" ).input_features def _snake_case ( self , __A ): """simple docstring""" return self.model.generate(inputs=__A ) def _snake_case ( self , __A ): """simple docstring""" return self.pre_processor.batch_decode(__A , skip_special_tokens=__A )[0]
283
1
from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize("""repo_id""" , ["""canonical_dataset_name""", """org-name/dataset-name"""] ) @pytest.mark.parametrize("""path""" , ["""filename.csv""", """filename with blanks.csv"""] ) @pytest.mark.parametrize("""revision""" , [None, """v2"""] ) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase = hf_hub_url(repo_id=UpperCamelCase_ , path=UpperCamelCase_ , revision=UpperCamelCase_ ) assert url == f"""https://huggingface.co/datasets/{repo_id}/resolve/{revision or "main"}/{quote(UpperCamelCase_ )}"""
165
def lowercase( UpperCamelCase_ = 1000000 ) -> int: '''simple docstring''' UpperCamelCase = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , UpperCamelCase_ ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
165
1
"""simple docstring""" def UpperCamelCase ( UpperCAmelCase = 200 ) ->int: """simple docstring""" a_ = [1, 2, 5, 10, 20, 50, 100, 200] a_ = [0] * (pence + 1) a_ = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(UpperCAmelCase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
243
"""simple docstring""" import baseaa def UpperCamelCase ( UpperCAmelCase ) ->bytes: """simple docstring""" return baseaa.baaencode(string.encode("utf-8" ) ) def UpperCamelCase ( UpperCAmelCase ) ->str: """simple docstring""" return baseaa.baadecode(UpperCAmelCase ).decode("utf-8" ) if __name__ == "__main__": UpperCamelCase_ = 'Hello World!' UpperCamelCase_ = baseaa_encode(test) print(encoded) UpperCamelCase_ = baseaa_decode(encoded) print(decoded)
243
1
import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class A : def __init__(self , lowerCAmelCase , lowerCAmelCase=1_3 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=True , lowerCAmelCase=9_9 , lowerCAmelCase=3_2 , lowerCAmelCase=5 , lowerCAmelCase=4 , lowerCAmelCase=3_7 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=5_1_2 , lowerCAmelCase=1_6 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=3 , lowerCAmelCase=4 , lowerCAmelCase=None , ): __lowercase= parent __lowercase= batch_size __lowercase= seq_length __lowercase= is_training __lowercase= use_input_mask __lowercase= use_token_type_ids __lowercase= use_labels __lowercase= vocab_size __lowercase= hidden_size __lowercase= num_hidden_layers __lowercase= num_attention_heads __lowercase= intermediate_size __lowercase= hidden_act __lowercase= hidden_dropout_prob __lowercase= attention_probs_dropout_prob __lowercase= max_position_embeddings __lowercase= type_vocab_size __lowercase= type_sequence_label_size __lowercase= initializer_range __lowercase= num_labels __lowercase= num_choices __lowercase= scope def _A (self ): __lowercase= ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase= None if self.use_input_mask: __lowercase= random_attention_mask([self.batch_size, self.seq_length] ) __lowercase= None if self.use_token_type_ids: __lowercase= ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowercase= None __lowercase= None __lowercase= None if self.use_labels: __lowercase= ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase= ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase= ids_tensor([self.batch_size] , self.num_choices ) __lowercase= self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _A (self ): return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__snake_case , initializer_range=self.initializer_range , ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): __lowercase= BioGptModel(config=__snake_case ) model.to(__snake_case ) model.eval() __lowercase= model(__snake_case , attention_mask=__snake_case ) __lowercase= model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ): __lowercase= BioGptForCausalLM(config=__snake_case ) model.to(__snake_case ) model.eval() __lowercase= model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , *lowerCAmelCase ): __lowercase= BioGptModel(config=__snake_case ) model.to(__snake_case ) model.eval() # create attention mask __lowercase= torch.ones(input_ids.shape , dtype=torch.long , device=__snake_case ) __lowercase= self.seq_length // 2 __lowercase= 0 # first forward pass __lowercase= model(__snake_case , attention_mask=__snake_case ).to_tuple() # create hypothetical next token and extent to next_input_ids __lowercase= ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids __lowercase= ids_tensor((1,) , __snake_case ).item() + 1 __lowercase= ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) __lowercase= random_other_next_tokens # append to next input_ids and attn_mask __lowercase= torch.cat([input_ids, next_tokens] , dim=-1 ) __lowercase= torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=__snake_case )] , dim=1 , ) # get two different outputs __lowercase= model(__snake_case , attention_mask=__snake_case )['last_hidden_state'] __lowercase= model(__snake_case , past_key_values=__snake_case , attention_mask=__snake_case )['last_hidden_state'] # select random slice __lowercase= ids_tensor((1,) , output_from_past.shape[-1] ).item() __lowercase= output_from_no_past[:, -1, random_slice_idx].detach() __lowercase= output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__snake_case , __snake_case , atol=1E-3 ) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , *lowerCAmelCase ): __lowercase= BioGptModel(config=__snake_case ).to(__snake_case ).eval() __lowercase= torch.ones(input_ids.shape , dtype=torch.long , device=__snake_case ) # first forward pass __lowercase= model(__snake_case , attention_mask=__snake_case , use_cache=__snake_case ) __lowercase= outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids __lowercase= ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowercase= ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and __lowercase= torch.cat([input_ids, next_tokens] , dim=-1 ) __lowercase= torch.cat([attention_mask, next_attn_mask] , dim=-1 ) __lowercase= model(__snake_case , attention_mask=__snake_case )['last_hidden_state'] __lowercase= model(__snake_case , attention_mask=__snake_case , past_key_values=__snake_case )[ 'last_hidden_state' ] # select random slice __lowercase= ids_tensor((1,) , output_from_past.shape[-1] ).item() __lowercase= output_from_no_past[:, -3:, random_slice_idx].detach() __lowercase= output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__snake_case , __snake_case , atol=1E-3 ) ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , *lowerCAmelCase , lowerCAmelCase=False ): __lowercase= BioGptForCausalLM(__snake_case ) model.to(__snake_case ) if gradient_checkpointing: model.gradient_checkpointing_enable() __lowercase= model(__snake_case , labels=__snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _A (self , lowerCAmelCase , *lowerCAmelCase ): __lowercase= BioGptModel(__snake_case ) __lowercase= model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_01 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , *lowerCAmelCase ): __lowercase= self.num_labels __lowercase= BioGptForTokenClassification(__snake_case ) model.to(__snake_case ) model.eval() __lowercase= model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _A (self ): __lowercase= self.prepare_config_and_inputs() ( __lowercase )= config_and_inputs __lowercase= {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): UpperCamelCase_ : Tuple =( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) UpperCamelCase_ : Dict =(BioGptForCausalLM,) if is_torch_available() else () UpperCamelCase_ : Tuple =( { '''feature-extraction''': BioGptModel, '''text-classification''': BioGptForSequenceClassification, '''text-generation''': BioGptForCausalLM, '''token-classification''': BioGptForTokenClassification, '''zero-shot''': BioGptForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase_ : Optional[int] =False def _A (self ): __lowercase= BioGptModelTester(self ) __lowercase= ConfigTester(self , config_class=__snake_case , hidden_size=3_7 ) def _A (self ): self.config_tester.run_common_tests() def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowercase= type self.model_tester.create_and_check_model(*__snake_case ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*__snake_case ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*__snake_case , gradient_checkpointing=__snake_case ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*__snake_case ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*__snake_case ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*__snake_case ) @slow def _A (self ): __lowercase= BioGptForCausalLM.from_pretrained('microsoft/biogpt' ) model.to(__snake_case ) __lowercase= BioGptTokenizer.from_pretrained('microsoft/biogpt' ) __lowercase= 'left' # Define PAD Token = EOS Token = 50256 __lowercase= tokenizer.eos_token __lowercase= model.config.eos_token_id # use different length sentences to test batching __lowercase= [ 'Hello, my dog is a little', 'Today, I', ] __lowercase= tokenizer(__snake_case , return_tensors='pt' , padding=__snake_case ) __lowercase= inputs['input_ids'].to(__snake_case ) __lowercase= model.generate( input_ids=__snake_case , attention_mask=inputs['attention_mask'].to(__snake_case ) , ) __lowercase= tokenizer(sentences[0] , return_tensors='pt' ).input_ids.to(__snake_case ) __lowercase= model.generate(input_ids=__snake_case ) __lowercase= inputs_non_padded.shape[-1] - inputs['attention_mask'][-1].long().sum().cpu().item() __lowercase= tokenizer(sentences[1] , return_tensors='pt' ).input_ids.to(__snake_case ) __lowercase= model.generate(input_ids=__snake_case , max_length=model.config.max_length - num_paddings ) __lowercase= tokenizer.batch_decode(__snake_case , skip_special_tokens=__snake_case ) __lowercase= tokenizer.decode(output_non_padded[0] , skip_special_tokens=__snake_case ) __lowercase= tokenizer.decode(output_padded[0] , skip_special_tokens=__snake_case ) __lowercase= [ 'Hello, my dog is a little bit bigger than a little bit.', 'Today, I have a good idea of how to use the information', ] self.assertListEqual(__snake_case , __snake_case ) self.assertListEqual(__snake_case , [non_padded_sentence, padded_sentence] ) @slow def _A (self ): for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase= BioGptModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs_for_common() __lowercase= 3 __lowercase= input_dict['input_ids'] __lowercase= input_ids.ne(1 ).to(__snake_case ) __lowercase= ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __lowercase= BioGptForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() __lowercase= model(__snake_case , attention_mask=__snake_case , labels=__snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _A (self ): __lowercase= self.model_tester.prepare_config_and_inputs_for_common() __lowercase= 3 __lowercase= 'multi_label_classification' __lowercase= input_dict['input_ids'] __lowercase= input_ids.ne(1 ).to(__snake_case ) __lowercase= ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __lowercase= BioGptForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() __lowercase= model(__snake_case , attention_mask=__snake_case , labels=__snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class A ( unittest.TestCase ): @slow def _A (self ): __lowercase= BioGptForCausalLM.from_pretrained('microsoft/biogpt' ) __lowercase= torch.tensor([[2, 4_8_0_5, 9, 6_5_6, 2_1]] ) __lowercase= model(__snake_case )[0] __lowercase= 4_2_3_8_4 __lowercase= torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , __snake_case ) __lowercase= torch.tensor( [[[-9.52_36, -9.89_18, 10.45_57], [-11.04_69, -9.64_23, 8.10_22], [-8.86_64, -7.88_26, 5.53_25]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def _A (self ): __lowercase= BioGptTokenizer.from_pretrained('microsoft/biogpt' ) __lowercase= BioGptForCausalLM.from_pretrained('microsoft/biogpt' ) model.to(__snake_case ) torch.manual_seed(0 ) __lowercase= tokenizer('COVID-19 is' , return_tensors='pt' ).to(__snake_case ) __lowercase= model.generate( **__snake_case , min_length=1_0_0 , max_length=1_0_2_4 , num_beams=5 , early_stopping=__snake_case , ) __lowercase= tokenizer.decode(output_ids[0] , skip_special_tokens=__snake_case ) __lowercase= ( 'COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the' ' causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and' ' territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),' ' and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and' ' more than 800,000 deaths.' ) self.assertEqual(__snake_case , __snake_case )
358
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCAmelCase = logging.get_logger(__name__) class A ( A_ ): UpperCamelCase_ : Dict =['''audio_values''', '''audio_mask'''] def __init__(self , lowerCAmelCase=2_0_4_8 , lowerCAmelCase=1 , lowerCAmelCase=[1_6, 1_6] , lowerCAmelCase=1_2_8 , lowerCAmelCase=4_4_1_0_0 , lowerCAmelCase=8_6 , lowerCAmelCase=2_0_4_8 , lowerCAmelCase=0.0 , **lowerCAmelCase , ): super().__init__( feature_size=lowerCAmelCase , sampling_rate=lowerCAmelCase , padding_value=lowerCAmelCase , **lowerCAmelCase , ) __lowercase= spectrogram_length __lowercase= num_channels __lowercase= patch_size __lowercase= feature_size // self.patch_size[1] __lowercase= n_fft __lowercase= sampling_rate // hop_length_to_sampling_rate __lowercase= sampling_rate __lowercase= padding_value __lowercase= mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=lowerCAmelCase , min_frequency=0.0 , max_frequency=2_20_50.0 , sampling_rate=lowerCAmelCase , norm='slaney' , mel_scale='slaney' , ).T def _A (self , lowerCAmelCase ): __lowercase= spectrogram( lowerCAmelCase , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=80.0 , ) __lowercase= log_spec[:, :-1] __lowercase= log_spec - 20.0 __lowercase= np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__(self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = False , lowerCAmelCase = False , **lowerCAmelCase , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled' f' with {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) __lowercase= isinstance(lowerCAmelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'Only mono-channel audio is supported for input to {self}' ) __lowercase= is_batched_numpy or ( isinstance(lowerCAmelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __lowercase= [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase , np.ndarray ): __lowercase= np.asarray(lowerCAmelCase , dtype=np.floataa ) elif isinstance(lowerCAmelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __lowercase= raw_speech.astype(np.floataa ) # always return batch if not is_batched: __lowercase= [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis __lowercase= [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , lowerCAmelCase ): __lowercase= [np.asarray(lowerCAmelCase , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask __lowercase= max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: __lowercase= [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] __lowercase= np.array(lowerCAmelCase ).astype(np.floataa ) # convert into correct format for padding __lowercase= max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch __lowercase= np.ones([len(lowerCAmelCase ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) __lowercase= padded_audio_features * self.padding_value for i in range(len(lowerCAmelCase ) ): __lowercase= audio_features[i] __lowercase= feature # return as BatchFeature if return_attention_mask: __lowercase= {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: __lowercase= {'audio_values': padded_audio_features} __lowercase= BatchFeature(data=lowerCAmelCase , tensor_type=lowerCAmelCase ) return encoded_inputs
304
0
import argparse import math import os import torch from neural_compressor.utils.pytorch import load from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, StableDiffusionPipeline, UNetaDConditionModel def lowerCamelCase__ ( ) -> int: _A: str = argparse.ArgumentParser() parser.add_argument( '''-m''' , '''--pretrained_model_name_or_path''' , type=a , default=a , required=a , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , ) parser.add_argument( '''-c''' , '''--caption''' , type=a , default='''robotic cat with wings''' , help='''Text used to generate images.''' , ) parser.add_argument( '''-n''' , '''--images_num''' , type=a , default=4 , help='''How much images to generate.''' , ) parser.add_argument( '''-s''' , '''--seed''' , type=a , default=42 , help='''Seed for random process.''' , ) parser.add_argument( '''-ci''' , '''--cuda_id''' , type=a , default=0 , help='''cuda_id.''' , ) _A: Dict = parser.parse_args() return args def lowerCamelCase__ ( a , a , a ) -> str: if not len(a ) == rows * cols: raise ValueError('''The specified number of rows and columns are not correct.''' ) _A , _A: Union[str, Any] = imgs[0].size _A: Optional[int] = Image.new('''RGB''' , size=(cols * w, rows * h) ) _A , _A: Tuple = grid.size for i, img in enumerate(a ): grid.paste(a , box=(i % cols * w, i // cols * h) ) return grid def lowerCamelCase__ ( a , a="robotic cat with wings" , a=7.5 , a=50 , a=1 , a=42 , ) -> str: _A: Optional[Any] = torch.Generator(pipeline.device ).manual_seed(a ) _A: List[Any] = pipeline( a , guidance_scale=a , num_inference_steps=a , generator=a , num_images_per_prompt=a , ).images _A: str = int(math.sqrt(a ) ) _A: int = image_grid(a , rows=_rows , cols=num_images_per_prompt // _rows ) return grid, images UpperCAmelCase__ : str = parse_args() # Load models and create wrapper for stable diffusion UpperCAmelCase__ : Dict = CLIPTokenizer.from_pretrained(args.pretrained_model_name_or_path, subfolder='tokenizer') UpperCAmelCase__ : Dict = CLIPTextModel.from_pretrained(args.pretrained_model_name_or_path, subfolder='text_encoder') UpperCAmelCase__ : Union[str, Any] = AutoencoderKL.from_pretrained(args.pretrained_model_name_or_path, subfolder='vae') UpperCAmelCase__ : str = UNetaDConditionModel.from_pretrained(args.pretrained_model_name_or_path, subfolder='unet') UpperCAmelCase__ : List[str] = StableDiffusionPipeline.from_pretrained( args.pretrained_model_name_or_path, text_encoder=text_encoder, vae=vae, unet=unet, tokenizer=tokenizer ) UpperCAmelCase__ : Any = lambda images, clip_input: (images, False) if os.path.exists(os.path.join(args.pretrained_model_name_or_path, 'best_model.pt')): UpperCAmelCase__ : Dict = load(args.pretrained_model_name_or_path, model=unet) unet.eval() setattr(pipeline, 'unet', unet) else: UpperCAmelCase__ : Dict = unet.to(torch.device('cuda', args.cuda_id)) UpperCAmelCase__ : Any = pipeline.to(unet.device) UpperCAmelCase__ ,UpperCAmelCase__ : List[str] = generate_images(pipeline, prompt=args.caption, num_images_per_prompt=args.images_num, seed=args.seed) grid.save(os.path.join(args.pretrained_model_name_or_path, '{}.png'.format('_'.join(args.caption.split())))) UpperCAmelCase__ : int = os.path.join(args.pretrained_model_name_or_path, '_'.join(args.caption.split())) os.makedirs(dirname, exist_ok=True) for idx, image in enumerate(images): image.save(os.path.join(dirname, '{}.png'.format(idx + 1)))
121
import random from .binary_exp_mod import bin_exp_mod def lowerCamelCase__ ( a , a=10_00 ) -> Optional[int]: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd _A: List[Any] = n - 1 _A: Dict = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) _A: List[str] = 0 while count < prec: _A: Optional[int] = random.randint(2 , n - 1 ) _A: Union[str, Any] = bin_exp_mod(a , a , a ) if b != 1: _A: Optional[Any] = True for _ in range(a ): if b == n - 1: _A: int = False break _A: Optional[Any] = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": UpperCAmelCase__ : Dict = abs(int(input('Enter bound : ').strip())) print('Here\'s the list of primes:') print(', '.join(str(i) for i in range(n + 1) if is_prime_big(i)))
121
1
'''simple docstring''' import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=__lowercase ) class _A ( __lowercase ): lowercase__: str = field(default='''automatic-speech-recognition''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) lowercase__: ClassVar[Features] = Features({'''audio''': Audio()} ) lowercase__: ClassVar[Features] = Features({'''transcription''': Value('''string''' )} ) lowercase__: str = "audio" lowercase__: str = "transcription" def lowercase__ ( self : Optional[int] , __magic_name__ : str ) -> int: """simple docstring""" if self.audio_column not in features: raise ValueError(f'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] , __magic_name__ ): raise ValueError(f'''Column {self.audio_column} is not an Audio type.''' ) __snake_case : List[Any] = copy.deepcopy(self ) __snake_case : Any = self.input_schema.copy() __snake_case : Union[str, Any] = features[self.audio_column] __snake_case : int = input_schema return task_template @property def lowercase__ ( self : int ) -> Dict[str, str]: """simple docstring""" return {self.audio_column: "audio", self.transcription_column: "transcription"}
13
'''simple docstring''' # Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union __UpperCamelCase = re.compile(R"^(?P<major>\d+)" R"\.(?P<minor>\d+)" R"\.(?P<patch>\d+)$") @total_ordering @dataclass class _A : lowercase__: str lowercase__: Optional[str] = None lowercase__: Optional[Union[str, int]] = None lowercase__: Optional[Union[str, int]] = None lowercase__: Optional[Union[str, int]] = None def lowercase__ ( self : str ) -> List[str]: """simple docstring""" __snake_case , __snake_case , __snake_case : List[Any] = _str_to_version_tuple(self.version_str ) def __repr__( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" return f'''{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}''' @property def lowercase__ ( self : Tuple ) -> Dict: """simple docstring""" return self.major, self.minor, self.patch def lowercase__ ( self : Any , __magic_name__ : Dict ) -> Optional[int]: """simple docstring""" if isinstance(__magic_name__ , __magic_name__ ): return Version(__magic_name__ ) elif isinstance(__magic_name__ , __magic_name__ ): return other raise TypeError(f'''{other} (type {type(__magic_name__ )}) cannot be compared to version.''' ) def __eq__( self : Optional[Any] , __magic_name__ : Union[str, Any] ) -> List[Any]: """simple docstring""" try: __snake_case : Union[str, Any] = self._validate_operand(__magic_name__ ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self : Union[str, Any] , __magic_name__ : Union[str, Any] ) -> List[str]: """simple docstring""" __snake_case : Union[str, Any] = self._validate_operand(__magic_name__ ) return self.tuple < other.tuple def __hash__( self : Any ) -> Any: """simple docstring""" return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def lowercase__ ( cls : List[str] , __magic_name__ : Tuple ) -> str: """simple docstring""" __snake_case : List[str] = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def lowercase__ ( self : str ) -> str: """simple docstring""" return self.version_str def _a ( _lowerCamelCase ) -> List[str]: """simple docstring""" __snake_case : List[Any] = _VERSION_REG.match(_lowerCamelCase ) if not res: raise ValueError(F'''Invalid version \'{version_str}\'. Format should be x.y.z with {{x,y,z}} being digits.''' ) return tuple(int(_lowerCamelCase ) for v in [res.group("""major""" ), res.group("""minor""" ), res.group("""patch""" )] ) def _a ( _lowerCamelCase ) -> Optional[int]: """simple docstring""" return ".".join(str(_lowerCamelCase ) for v in version_tuple )
13
1
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') UpperCamelCase__ = logging.getLogger(__name__) @dataclass class lowerCamelCase_ : lowerCAmelCase__ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) lowerCAmelCase__ = field( default=_snake_case , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) lowerCAmelCase__ = field( default=_snake_case , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) lowerCAmelCase__ = field( default=_snake_case , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) lowerCAmelCase__ = field( default=_snake_case , metadata={'help': 'Whether to use one of the fast tokenizer (backed by the tokenizers library) or not.'} , ) lowerCAmelCase__ = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) lowerCAmelCase__ = field( default=_snake_case , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) @dataclass class lowerCamelCase_ : lowerCAmelCase__ = field(default=_snake_case , metadata={'help': 'The input training data file (a text file).'} ) lowerCAmelCase__ = field( default=_snake_case , metadata={'help': 'An optional input evaluation data file to evaluate the perplexity on (a text file).'} , ) lowerCAmelCase__ = field( default=_snake_case , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) lowerCAmelCase__ = field( default=_snake_case , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) lowerCAmelCase__ = field( default=_snake_case , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. If passed, sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) lowerCAmelCase__ = field( default=_snake_case , metadata={ 'help': ( 'Whether to pad all samples to the maximum sentence length. ' 'If False, will pad the samples dynamically when batching to the maximum length in the batch. More ' 'efficient on GPU but very bad for TPU.' ) } , ) lowerCAmelCase__ = field( default=_snake_case , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) lowerCAmelCase__ = field( default=_snake_case , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) def lowercase_ ( self : Tuple ): '''simple docstring''' if self.train_file is not None: UpperCAmelCase__ : str = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: UpperCAmelCase__ : Dict = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class lowerCamelCase_ : lowerCAmelCase__ = 42 lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = None def __call__( self : int , _A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = """label""" if """label""" in features[0].keys() else """labels""" UpperCAmelCase__ : List[Any] = [feature.pop(__snake_case ) for feature in features] UpperCAmelCase__ : Union[str, Any] = len(__snake_case ) UpperCAmelCase__ : int = len(features[0]['''input_ids'''] ) UpperCAmelCase__ : List[Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(__snake_case )] for feature in features ] UpperCAmelCase__ : List[Any] = list(chain(*__snake_case ) ) UpperCAmelCase__ : Optional[int] = self.tokenizer.pad( __snake_case , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) # Un-flatten UpperCAmelCase__ : Tuple = {k: v.view(__snake_case , __snake_case , -1 ) for k, v in batch.items()} # Add back labels UpperCAmelCase__ : str = torch.tensor(__snake_case , dtype=torch.intaa ) return batch def a__ ( ) -> Any: UpperCAmelCase__ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCAmelCase__ : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCAmelCase__ : Optional[int] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_swag''' , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() UpperCAmelCase__ : List[str] = training_args.get_process_log_level() logger.setLevel(SCREAMING_SNAKE_CASE__ ) datasets.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE__ ) transformers.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE__ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. UpperCAmelCase__ : Optional[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: UpperCAmelCase__ : List[Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: UpperCAmelCase__ : List[str] = {} if data_args.train_file is not None: UpperCAmelCase__ : Union[str, Any] = data_args.train_file if data_args.validation_file is not None: UpperCAmelCase__ : List[Any] = data_args.validation_file UpperCAmelCase__ : Optional[int] = data_args.train_file.split('''.''' )[-1] UpperCAmelCase__ : Optional[Any] = load_dataset( SCREAMING_SNAKE_CASE__ , data_files=SCREAMING_SNAKE_CASE__ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. UpperCAmelCase__ : List[Any] = load_dataset( '''swag''' , '''regular''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase__ : Optional[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) UpperCAmelCase__ : Tuple = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) UpperCAmelCase__ : int = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE__ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. UpperCAmelCase__ : Any = [F"""ending{i}""" for i in range(4 )] UpperCAmelCase__ : Dict = """sent1""" UpperCAmelCase__ : int = """sent2""" if data_args.max_seq_length is None: UpperCAmelCase__ : Optional[Any] = tokenizer.model_max_length if max_seq_length > 10_24: logger.warning( '''The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value''' ''' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can''' ''' override this default with `--block_size xxx`.''' ) UpperCAmelCase__ : int = 10_24 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" F"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) UpperCAmelCase__ : Optional[Any] = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(lowerCAmelCase__ ): UpperCAmelCase__ : Any = [[context] * 4 for context in examples[context_name]] UpperCAmelCase__ : Dict = examples[question_header_name] UpperCAmelCase__ : Optional[Any] = [ [F"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(SCREAMING_SNAKE_CASE__ ) ] # Flatten out UpperCAmelCase__ : int = list(chain(*SCREAMING_SNAKE_CASE__ ) ) UpperCAmelCase__ : Optional[int] = list(chain(*SCREAMING_SNAKE_CASE__ ) ) # Tokenize UpperCAmelCase__ : Optional[Any] = tokenizer( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , truncation=SCREAMING_SNAKE_CASE__ , max_length=SCREAMING_SNAKE_CASE__ , padding='''max_length''' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(SCREAMING_SNAKE_CASE__ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('''--do_train requires a train dataset''' ) UpperCAmelCase__ : str = raw_datasets["""train"""] if data_args.max_train_samples is not None: UpperCAmelCase__ : List[Any] = min(len(SCREAMING_SNAKE_CASE__ ) , data_args.max_train_samples ) UpperCAmelCase__ : List[Any] = train_dataset.select(range(SCREAMING_SNAKE_CASE__ ) ) with training_args.main_process_first(desc='''train dataset map pre-processing''' ): UpperCAmelCase__ : List[str] = train_dataset.map( SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('''--do_eval requires a validation dataset''' ) UpperCAmelCase__ : Any = raw_datasets["""validation"""] if data_args.max_eval_samples is not None: UpperCAmelCase__ : int = min(len(SCREAMING_SNAKE_CASE__ ) , data_args.max_eval_samples ) UpperCAmelCase__ : List[str] = eval_dataset.select(range(SCREAMING_SNAKE_CASE__ ) ) with training_args.main_process_first(desc='''validation dataset map pre-processing''' ): UpperCAmelCase__ : List[Any] = eval_dataset.map( SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator UpperCAmelCase__ : Dict = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=SCREAMING_SNAKE_CASE__ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(lowerCAmelCase__ ): UpperCAmelCase__ : Optional[Any] = eval_predictions UpperCAmelCase__ : str = np.argmax(SCREAMING_SNAKE_CASE__ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer UpperCAmelCase__ : Tuple = Trainer( model=SCREAMING_SNAKE_CASE__ , args=SCREAMING_SNAKE_CASE__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=SCREAMING_SNAKE_CASE__ , data_collator=SCREAMING_SNAKE_CASE__ , compute_metrics=SCREAMING_SNAKE_CASE__ , ) # Training if training_args.do_train: UpperCAmelCase__ : List[Any] = None if training_args.resume_from_checkpoint is not None: UpperCAmelCase__ : Dict = training_args.resume_from_checkpoint elif last_checkpoint is not None: UpperCAmelCase__ : Dict = last_checkpoint UpperCAmelCase__ : Optional[Any] = trainer.train(resume_from_checkpoint=SCREAMING_SNAKE_CASE__ ) trainer.save_model() # Saves the tokenizer too for easy upload UpperCAmelCase__ : List[Any] = train_result.metrics UpperCAmelCase__ : List[str] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(SCREAMING_SNAKE_CASE__ ) ) UpperCAmelCase__ : List[Any] = min(SCREAMING_SNAKE_CASE__ , len(SCREAMING_SNAKE_CASE__ ) ) trainer.log_metrics('''train''' , SCREAMING_SNAKE_CASE__ ) trainer.save_metrics('''train''' , SCREAMING_SNAKE_CASE__ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''' ) UpperCAmelCase__ : Optional[int] = trainer.evaluate() UpperCAmelCase__ : Dict = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ : List[str] = min(SCREAMING_SNAKE_CASE__ , len(SCREAMING_SNAKE_CASE__ ) ) trainer.log_metrics('''eval''' , SCREAMING_SNAKE_CASE__ ) trainer.save_metrics('''eval''' , SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ : int = { """finetuned_from""": model_args.model_name_or_path, """tasks""": """multiple-choice""", """dataset_tags""": """swag""", """dataset_args""": """regular""", """dataset""": """SWAG""", """language""": """en""", } if training_args.push_to_hub: trainer.push_to_hub(**SCREAMING_SNAKE_CASE__ ) else: trainer.create_model_card(**SCREAMING_SNAKE_CASE__ ) def a__ ( lowerCAmelCase__ ) -> Optional[int]: main() if __name__ == "__main__": main()
181
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class lowercase__ ( unittest.TestCase ): '''simple docstring''' def __init__( self , __snake_case , __snake_case=7 , __snake_case=3 , __snake_case=18 , __snake_case=30 , __snake_case=400 , __snake_case=True , __snake_case=None , __snake_case=True , __snake_case=None , ): _SCREAMING_SNAKE_CASE : Any = size if size is not None else {"""shortest_edge""": 20} _SCREAMING_SNAKE_CASE : int = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} _SCREAMING_SNAKE_CASE : Dict = parent _SCREAMING_SNAKE_CASE : Any = batch_size _SCREAMING_SNAKE_CASE : Any = num_channels _SCREAMING_SNAKE_CASE : Dict = image_size _SCREAMING_SNAKE_CASE : Any = min_resolution _SCREAMING_SNAKE_CASE : Any = max_resolution _SCREAMING_SNAKE_CASE : Union[str, Any] = do_resize _SCREAMING_SNAKE_CASE : Tuple = size _SCREAMING_SNAKE_CASE : Optional[int] = do_center_crop _SCREAMING_SNAKE_CASE : str = crop_size def UpperCAmelCase_ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class lowercase__ ( _snake_case , unittest.TestCase ): '''simple docstring''' A_ : Union[str, Any] = MobileNetVaImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : int = MobileNetVaImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__snake_case , """do_resize""" ) ) self.assertTrue(hasattr(__snake_case , """size""" ) ) self.assertTrue(hasattr(__snake_case , """do_center_crop""" ) ) self.assertTrue(hasattr(__snake_case , """crop_size""" ) ) def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) _SCREAMING_SNAKE_CASE : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) def UpperCAmelCase_ ( self ): pass def UpperCAmelCase_ ( self ): # Initialize image_processing _SCREAMING_SNAKE_CASE : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _SCREAMING_SNAKE_CASE : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , Image.Image ) # Test not batched input _SCREAMING_SNAKE_CASE : Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _SCREAMING_SNAKE_CASE : int = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing _SCREAMING_SNAKE_CASE : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _SCREAMING_SNAKE_CASE : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case , numpify=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , np.ndarray ) # Test not batched input _SCREAMING_SNAKE_CASE : Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _SCREAMING_SNAKE_CASE : Optional[Any] = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ): # Initialize image_processing _SCREAMING_SNAKE_CASE : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _SCREAMING_SNAKE_CASE : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case , torchify=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , torch.Tensor ) # Test not batched input _SCREAMING_SNAKE_CASE : List[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _SCREAMING_SNAKE_CASE : str = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
200
0
"""simple docstring""" import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) UpperCamelCase : Dict = logging.getLogger(__name__) @dataclass(frozen=__SCREAMING_SNAKE_CASE ) class __lowerCAmelCase : lowercase = 42 lowercase = 42 lowercase = None lowercase = None lowercase = None @dataclass(frozen=__SCREAMING_SNAKE_CASE ) class __lowerCAmelCase : lowercase = 42 lowercase = None lowercase = None lowercase = None lowercase = None if is_torch_available(): import torch from torch.utils.data import Dataset class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = 42 def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase=False , __UpperCAmelCase = False , ): '''simple docstring''' __UpperCamelCase = hans_processors[task]() __UpperCamelCase = os.path.join( __UpperCAmelCase , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(__UpperCAmelCase ) , __UpperCAmelCase , ) , ) __UpperCamelCase = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) __UpperCamelCase , __UpperCamelCase = label_list[2], label_list[1] __UpperCamelCase = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __UpperCamelCase = cached_features_file + '.lock' with FileLock(__UpperCAmelCase ): if os.path.exists(__UpperCAmelCase ) and not overwrite_cache: logger.info(F'Loading features from cached file {cached_features_file}' ) __UpperCamelCase = torch.load(__UpperCAmelCase ) else: logger.info(F'Creating features from dataset file at {data_dir}' ) __UpperCamelCase = ( processor.get_dev_examples(__UpperCAmelCase ) if evaluate else processor.get_train_examples(__UpperCAmelCase ) ) logger.info('Training examples: %s' , len(__UpperCAmelCase ) ) __UpperCamelCase = hans_convert_examples_to_features(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) logger.info('Saving features into cached file %s' , __UpperCAmelCase ) torch.save(self.features , __UpperCAmelCase ) def __len__( self ): '''simple docstring''' return len(self.features ) def __getitem__( self , __UpperCAmelCase ): '''simple docstring''' return self.features[i] def UpperCAmelCase ( self ): '''simple docstring''' return self.label_list if is_tf_available(): import tensorflow as tf class __lowerCAmelCase : lowercase = 42 def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = 128 , __UpperCAmelCase=False , __UpperCAmelCase = False , ): '''simple docstring''' __UpperCamelCase = hans_processors[task]() __UpperCamelCase = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) __UpperCamelCase , __UpperCamelCase = label_list[2], label_list[1] __UpperCamelCase = label_list __UpperCamelCase = processor.get_dev_examples(__UpperCAmelCase ) if evaluate else processor.get_train_examples(__UpperCAmelCase ) __UpperCamelCase = hans_convert_examples_to_features(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 1_0000 == 0: logger.info('Writing example %d of %d' % (ex_index, len(__UpperCAmelCase )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) __UpperCamelCase = tf.data.Dataset.from_generator( __UpperCAmelCase , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def UpperCAmelCase ( self ): '''simple docstring''' return self.dataset def __len__( self ): '''simple docstring''' return len(self.features ) def __getitem__( self , __UpperCAmelCase ): '''simple docstring''' return self.features[i] def UpperCAmelCase ( self ): '''simple docstring''' return self.label_list class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(__UpperCAmelCase , 'heuristics_train_set.txt' ) ) , 'train' ) def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(__UpperCAmelCase , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def UpperCAmelCase ( self ): '''simple docstring''' return ["contradiction", "entailment", "neutral"] def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = [] for i, line in enumerate(__UpperCAmelCase ): if i == 0: continue __UpperCamelCase = '%s-%s' % (set_type, line[0]) __UpperCamelCase = line[5] __UpperCamelCase = line[6] __UpperCamelCase = line[7][2:] if line[7].startswith('ex' ) else line[7] __UpperCamelCase = line[0] examples.append(InputExample(guid=__UpperCAmelCase , text_a=__UpperCAmelCase , text_b=__UpperCAmelCase , label=__UpperCAmelCase , pairID=__UpperCAmelCase ) ) return examples def A ( snake_case :List[InputExample] , snake_case :List[str] , snake_case :int , snake_case :PreTrainedTokenizer , ) -> int: __UpperCamelCase = {label: i for i, label in enumerate(snake_case )} __UpperCamelCase = [] for ex_index, example in tqdm.tqdm(enumerate(snake_case ) , desc='convert examples to features' ): if ex_index % 1_0_0_0_0 == 0: logger.info('Writing example %d' % (ex_index) ) __UpperCamelCase = tokenizer( example.text_a , example.text_b , add_special_tokens=snake_case , max_length=snake_case , padding='max_length' , truncation=snake_case , return_overflowing_tokens=snake_case , ) __UpperCamelCase = label_map[example.label] if example.label in label_map else 0 __UpperCamelCase = int(example.pairID ) features.append(InputFeatures(**snake_case , label=snake_case , pairID=snake_case ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(f'guid: {example}' ) logger.info(f'features: {features[i]}' ) return features UpperCamelCase : Any = { "hans": 3, } UpperCamelCase : Union[str, Any] = { "hans": HansProcessor, }
263
"""simple docstring""" import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' super().__init__( __UpperCAmelCase , split=__UpperCAmelCase , features=__UpperCAmelCase , cache_dir=__UpperCAmelCase , keep_in_memory=__UpperCAmelCase , streaming=__UpperCAmelCase , num_proc=__UpperCAmelCase , **__UpperCAmelCase , ) __UpperCamelCase = field __UpperCamelCase = path_or_paths if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else {self.split: path_or_paths} __UpperCamelCase = Json( cache_dir=__UpperCAmelCase , data_files=__UpperCAmelCase , features=__UpperCAmelCase , field=__UpperCAmelCase , **__UpperCAmelCase , ) def UpperCAmelCase ( self ): '''simple docstring''' if self.streaming: __UpperCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None self.builder.download_and_prepare( download_config=__UpperCAmelCase , download_mode=__UpperCAmelCase , verification_mode=__UpperCAmelCase , base_path=__UpperCAmelCase , num_proc=self.num_proc , ) __UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__UpperCAmelCase , in_memory=self.keep_in_memory ) return dataset class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' if num_proc is not None and num_proc <= 0: raise ValueError(F'num_proc {num_proc} must be an integer > 0.' ) __UpperCamelCase = dataset __UpperCamelCase = path_or_buf __UpperCamelCase = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE __UpperCamelCase = num_proc __UpperCamelCase = 'utf-8' __UpperCamelCase = to_json_kwargs def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = self.to_json_kwargs.pop('path_or_buf' , __UpperCAmelCase ) __UpperCamelCase = self.to_json_kwargs.pop('orient' , 'records' ) __UpperCamelCase = self.to_json_kwargs.pop('lines' , True if orient == 'records' else False ) __UpperCamelCase = self.to_json_kwargs.pop('index' , False if orient in ['split', 'table'] else True ) __UpperCamelCase = self.to_json_kwargs.pop('compression' , __UpperCAmelCase ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(F'`datasets` currently does not support {compression} compression' ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , 'wb' , compression=__UpperCAmelCase ) as buffer: __UpperCamelCase = self._write(file_obj=__UpperCAmelCase , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( F'The compression parameter is not supported when writing to a buffer, but compression={compression}' ' was passed. Please provide a local path instead.' ) __UpperCamelCase = self._write( file_obj=self.path_or_buf , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **self.to_json_kwargs ) return written def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = args __UpperCamelCase = query_table( table=self.dataset.data , key=slice(__UpperCAmelCase , offset + self.batch_size ) , indices=self.dataset._indices , ) __UpperCamelCase = batch.to_pandas().to_json( path_or_buf=__UpperCAmelCase , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **__UpperCAmelCase ) if not json_str.endswith('\n' ): json_str += "\n" return json_str.encode(self.encoding ) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ): '''simple docstring''' __UpperCamelCase = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating json from Arrow format' , ): __UpperCamelCase = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(__UpperCAmelCase ) else: __UpperCamelCase , __UpperCamelCase = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , __UpperCAmelCase , __UpperCAmelCase )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating json from Arrow format' , ): written += file_obj.write(__UpperCAmelCase ) return written
263
1
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class __A : def __init__( self : str , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[Any]=2 , UpperCAmelCase_ : List[str]=8 , UpperCAmelCase_ : List[Any]=True , UpperCAmelCase_ : int=True , UpperCAmelCase_ : Union[str, Any]=True , UpperCAmelCase_ : str=True , UpperCAmelCase_ : List[str]=99 , UpperCAmelCase_ : Union[str, Any]=16 , UpperCAmelCase_ : int=5 , UpperCAmelCase_ : Dict=2 , UpperCAmelCase_ : Optional[int]=36 , UpperCAmelCase_ : Dict="gelu" , UpperCAmelCase_ : str=0.0 , UpperCAmelCase_ : Optional[int]=0.0 , UpperCAmelCase_ : Any=512 , UpperCAmelCase_ : Union[str, Any]=16 , UpperCAmelCase_ : Dict=2 , UpperCAmelCase_ : Union[str, Any]=0.02 , UpperCAmelCase_ : Tuple=3 , UpperCAmelCase_ : Optional[int]=4 , UpperCAmelCase_ : Tuple=None , ): lowerCAmelCase : str = parent lowerCAmelCase : Any = batch_size lowerCAmelCase : Optional[Any] = seq_length lowerCAmelCase : Tuple = is_training lowerCAmelCase : Union[str, Any] = use_input_mask lowerCAmelCase : str = use_token_type_ids lowerCAmelCase : int = use_labels lowerCAmelCase : Tuple = vocab_size lowerCAmelCase : Tuple = hidden_size lowerCAmelCase : str = num_hidden_layers lowerCAmelCase : Any = num_attention_heads lowerCAmelCase : Optional[int] = intermediate_size lowerCAmelCase : Tuple = hidden_act lowerCAmelCase : List[str] = hidden_dropout_prob lowerCAmelCase : List[Any] = attention_probs_dropout_prob lowerCAmelCase : Optional[Any] = max_position_embeddings lowerCAmelCase : List[str] = type_vocab_size lowerCAmelCase : Dict = type_sequence_label_size lowerCAmelCase : List[Any] = initializer_range lowerCAmelCase : List[Any] = num_labels lowerCAmelCase : str = num_choices lowerCAmelCase : Tuple = scope def lowercase__ ( self : Optional[int] ): lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase : Dict = None if self.use_input_mask: lowerCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase : int = None if self.use_token_type_ids: lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase : List[str] = None lowerCAmelCase : List[str] = None lowerCAmelCase : Tuple = None if self.use_labels: lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase : Tuple = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase__ ( self : Any ): return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_UpperCamelCase , initializer_range=self.initializer_range , ) def lowercase__ ( self : Dict ): lowerCAmelCase : Dict = self.get_config() lowerCAmelCase : Optional[Any] = 300 return config def lowercase__ ( self : Tuple ): ( ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ) : Dict = self.prepare_config_and_inputs() lowerCAmelCase : Union[str, Any] = True lowerCAmelCase : Optional[int] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def lowercase__ ( self : Dict , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Dict , UpperCAmelCase_ : str , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : List[Any] ): lowerCAmelCase : Tuple = MraModel(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase : Union[str, Any] = model(_UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase ) lowerCAmelCase : Optional[Any] = model(_UpperCamelCase , token_type_ids=_UpperCamelCase ) lowerCAmelCase : Optional[Any] = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self : Any , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : str , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : int , UpperCAmelCase_ : Any , ): lowerCAmelCase : Union[str, Any] = True lowerCAmelCase : Dict = MraModel(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase : Optional[int] = model( _UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase , encoder_hidden_states=_UpperCamelCase , encoder_attention_mask=_UpperCamelCase , ) lowerCAmelCase : Tuple = model( _UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase , encoder_hidden_states=_UpperCamelCase , ) lowerCAmelCase : int = model(_UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self : List[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any] ): lowerCAmelCase : List[Any] = MraForMaskedLM(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase : Optional[int] = model(_UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase__ ( self : Union[str, Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Any , UpperCAmelCase_ : int , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int ): lowerCAmelCase : Dict = MraForQuestionAnswering(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase : Any = model( _UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase , start_positions=_UpperCamelCase , end_positions=_UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase__ ( self : List[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : Any , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : str ): lowerCAmelCase : List[str] = self.num_labels lowerCAmelCase : List[str] = MraForSequenceClassification(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase : Any = model(_UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase__ ( self : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Dict ): lowerCAmelCase : Optional[Any] = self.num_labels lowerCAmelCase : Tuple = MraForTokenClassification(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase : Optional[int] = model(_UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase__ ( self : Optional[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : Any , UpperCAmelCase_ : str , UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[int] ): lowerCAmelCase : List[str] = self.num_choices lowerCAmelCase : Optional[int] = MraForMultipleChoice(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() lowerCAmelCase : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase : str = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase : Any = model( _UpperCamelCase , attention_mask=_UpperCamelCase , token_type_ids=_UpperCamelCase , labels=_UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase__ ( self : List[Any] ): lowerCAmelCase : Tuple = self.prepare_config_and_inputs() ( ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ) : Any = config_and_inputs lowerCAmelCase : Union[str, Any] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class __A ( __a , unittest.TestCase ): lowerCAmelCase_ : int = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) lowerCAmelCase_ : List[Any] = False lowerCAmelCase_ : int = False lowerCAmelCase_ : Tuple = False lowerCAmelCase_ : Dict = False lowerCAmelCase_ : str = () def lowercase__ ( self : Optional[Any] ): lowerCAmelCase : int = MraModelTester(self ) lowerCAmelCase : List[Any] = ConfigTester(self , config_class=_UpperCamelCase , hidden_size=37 ) def lowercase__ ( self : List[Any] ): self.config_tester.run_common_tests() def lowercase__ ( self : int ): lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCamelCase ) def lowercase__ ( self : str ): lowerCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCAmelCase : int = type self.model_tester.create_and_check_model(*_UpperCamelCase ) def lowercase__ ( self : str ): lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_UpperCamelCase ) def lowercase__ ( self : Optional[Any] ): lowerCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_UpperCamelCase ) def lowercase__ ( self : Tuple ): lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_UpperCamelCase ) def lowercase__ ( self : Union[str, Any] ): lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_UpperCamelCase ) def lowercase__ ( self : Tuple ): lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_UpperCamelCase ) @slow def lowercase__ ( self : Union[str, Any] ): for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase : Union[str, Any] = MraModel.from_pretrained(_UpperCamelCase ) self.assertIsNotNone(_UpperCamelCase ) @unittest.skip(reason='MRA does not output attentions' ) def lowercase__ ( self : List[str] ): return @require_torch class __A ( unittest.TestCase ): @slow def lowercase__ ( self : str ): lowerCAmelCase : Optional[int] = MraModel.from_pretrained('uw-madison/mra-base-512-4' ) lowerCAmelCase : List[str] = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): lowerCAmelCase : int = model(_UpperCamelCase )[0] lowerCAmelCase : Dict = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , _UpperCamelCase ) lowerCAmelCase : Optional[Any] = torch.tensor( [[[-0.01_40, 0.08_30, -0.03_81], [0.15_46, 0.14_02, 0.02_20], [0.11_62, 0.08_51, 0.01_65]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _UpperCamelCase , atol=1E-4 ) ) @slow def lowercase__ ( self : Optional[Any] ): lowerCAmelCase : List[Any] = MraForMaskedLM.from_pretrained('uw-madison/mra-base-512-4' ) lowerCAmelCase : List[Any] = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): lowerCAmelCase : List[str] = model(_UpperCamelCase )[0] lowerCAmelCase : int = 50265 lowerCAmelCase : int = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , _UpperCamelCase ) lowerCAmelCase : int = torch.tensor( [[[9.25_95, -3.60_38, 11.88_19], [9.38_69, -3.26_93, 11.09_56], [11.85_24, -3.49_38, 13.12_10]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _UpperCamelCase , atol=1E-4 ) ) @slow def lowercase__ ( self : Tuple ): lowerCAmelCase : Dict = MraForMaskedLM.from_pretrained('uw-madison/mra-base-4096-8-d3' ) lowerCAmelCase : Dict = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): lowerCAmelCase : str = model(_UpperCamelCase )[0] lowerCAmelCase : Any = 50265 lowerCAmelCase : List[Any] = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , _UpperCamelCase ) lowerCAmelCase : Dict = torch.tensor( [[[5.47_89, -2.35_64, 7.50_64], [7.90_67, -1.33_69, 9.96_68], [9.07_12, -1.81_06, 7.03_80]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _UpperCamelCase , atol=1E-4 ) )
138
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _A = { "configuration_distilbert": [ "DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DistilBertConfig", "DistilBertOnnxConfig", ], "tokenization_distilbert": ["DistilBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["DistilBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "DistilBertForMaskedLM", "DistilBertForMultipleChoice", "DistilBertForQuestionAnswering", "DistilBertForSequenceClassification", "DistilBertForTokenClassification", "DistilBertModel", "DistilBertPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDistilBertForMaskedLM", "TFDistilBertForMultipleChoice", "TFDistilBertForQuestionAnswering", "TFDistilBertForSequenceClassification", "TFDistilBertForTokenClassification", "TFDistilBertMainLayer", "TFDistilBertModel", "TFDistilBertPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "FlaxDistilBertForMaskedLM", "FlaxDistilBertForMultipleChoice", "FlaxDistilBertForQuestionAnswering", "FlaxDistilBertForSequenceClassification", "FlaxDistilBertForTokenClassification", "FlaxDistilBertModel", "FlaxDistilBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys _A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
231
0
'''simple docstring''' from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { """google/efficientnet-b7""": """https://huggingface.co/google/efficientnet-b7/resolve/main/config.json""", } class _UpperCamelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase__ = "efficientnet" def __init__( self : Dict , _lowerCAmelCase : int = 3 , _lowerCAmelCase : int = 6_0_0 , _lowerCAmelCase : float = 2.0 , _lowerCAmelCase : float = 3.1 , _lowerCAmelCase : int = 8 , _lowerCAmelCase : List[int] = [3, 3, 5, 3, 5, 5, 3] , _lowerCAmelCase : List[int] = [3_2, 1_6, 2_4, 4_0, 8_0, 1_1_2, 1_9_2] , _lowerCAmelCase : List[int] = [1_6, 2_4, 4_0, 8_0, 1_1_2, 1_9_2, 3_2_0] , _lowerCAmelCase : List[int] = [] , _lowerCAmelCase : List[int] = [1, 2, 2, 2, 1, 2, 1] , _lowerCAmelCase : List[int] = [1, 2, 2, 3, 3, 4, 1] , _lowerCAmelCase : List[int] = [1, 6, 6, 6, 6, 6, 6] , _lowerCAmelCase : float = 0.25 , _lowerCAmelCase : str = "swish" , _lowerCAmelCase : int = 2_5_6_0 , _lowerCAmelCase : str = "mean" , _lowerCAmelCase : float = 0.02 , _lowerCAmelCase : float = 0.001 , _lowerCAmelCase : float = 0.99 , _lowerCAmelCase : float = 0.5 , _lowerCAmelCase : float = 0.2 , **_lowerCAmelCase : List[str] , ): '''simple docstring''' super().__init__(**lowerCamelCase_) __lowercase =num_channels __lowercase =image_size __lowercase =width_coefficient __lowercase =depth_coefficient __lowercase =depth_divisor __lowercase =kernel_sizes __lowercase =in_channels __lowercase =out_channels __lowercase =depthwise_padding __lowercase =strides __lowercase =num_block_repeats __lowercase =expand_ratios __lowercase =squeeze_expansion_ratio __lowercase =hidden_act __lowercase =hidden_dim __lowercase =pooling_type __lowercase =initializer_range __lowercase =batch_norm_eps __lowercase =batch_norm_momentum __lowercase =dropout_rate __lowercase =drop_connect_rate __lowercase =sum(lowerCamelCase_) * 4 class _UpperCamelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase__ = version.parse("""1.11""" ) @property def __lowerCamelCase ( self : Any): '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ]) @property def __lowerCamelCase ( self : List[str]): '''simple docstring''' return 1e-5
364
'''simple docstring''' import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } lowerCamelCase = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" for attribute in key.split('.' ): __lowercase =getattr(_lowerCAmelCase , _lowerCAmelCase ) if weight_type is not None: __lowercase =getattr(_lowerCAmelCase , _lowerCAmelCase ).shape else: __lowercase =hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowercase =value elif weight_type == "weight_g": __lowercase =value elif weight_type == "weight_v": __lowercase =value elif weight_type == "bias": __lowercase =value else: __lowercase =value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =[] __lowercase =fairseq_model.state_dict() __lowercase =hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight __lowercase =None for name, value in fairseq_dict.items(): __lowercase =False if "conv_layers" in name: load_conv_layer( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) __lowercase =True elif name.split('.' )[0] == "proj": __lowercase =fairseq_model.proj __lowercase =True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: __lowercase =True if "*" in mapped_key: __lowercase =name.split(_lowerCAmelCase )[0].split('.' )[-2] __lowercase =mapped_key.replace('*' , _lowerCAmelCase ) if "weight_g" in name: __lowercase ='weight_g' elif "weight_v" in name: __lowercase ='weight_v' elif "bias" in name: __lowercase ='bias' elif "weight" in name: __lowercase ='weight' else: __lowercase =None set_recursively(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) continue if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(f"""Unused weights: {unused_weights}""" ) return proj_weight def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =full_name.split('conv_layers.' )[-1] __lowercase =name.split('.' ) __lowercase =int(items[0] ) __lowercase =int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowercase =value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowercase =value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __lowercase =value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowercase =value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCAmelCase ) def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase , __lowercase =emb.weight.shape __lowercase =nn.Linear(_lowerCAmelCase , _lowerCAmelCase , bias=_lowerCAmelCase ) __lowercase =emb.weight.data return lin_layer def _A ( _lowerCAmelCase ): """simple docstring""" with open(_lowerCAmelCase , 'r' , encoding='utf-8' ) as f: __lowercase =f.readlines() __lowercase =[line.split(' ' )[0] for line in lines] __lowercase =len(_lowerCAmelCase ) __lowercase ={ '<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3, } vocab_dict.update(dict(zip(_lowerCAmelCase , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ): """simple docstring""" __lowercase =WavaVecaConfig.from_pretrained(_lowerCAmelCase ) __lowercase =SpeechaTextaConfig.from_pretrained( _lowerCAmelCase , vocab_size=_lowerCAmelCase , decoder_layers=_lowerCAmelCase , do_stable_layer_norm=_lowerCAmelCase ) __lowercase =WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) __lowercase , __lowercase , __lowercase =fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) __lowercase =model[0].eval() # set weights for wav2vec2 encoder __lowercase =WavaVecaModel(_lowerCAmelCase ) __lowercase =recursively_load_weights_wavaveca(model.encoder , _lowerCAmelCase ) __lowercase =SpeechaTextaForCausalLM(_lowerCAmelCase ) __lowercase , __lowercase =hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=_lowerCAmelCase ) # set output linear layer unexpected_keys.remove('embed_out' ) __lowercase =nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(f"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(f"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) __lowercase =SpeechEncoderDecoderModel(encoder=_lowerCAmelCase , decoder=_lowerCAmelCase ) __lowercase =False # add projection layer __lowercase =nn.Parameter(projection_layer.weight ) __lowercase =nn.Parameter(projection_layer.bias ) __lowercase =create_vocab_dict(_lowerCAmelCase ) with open(os.path.join(_lowerCAmelCase , 'vocab.json' ) , 'w' ) as fp: json.dump(_lowerCAmelCase , _lowerCAmelCase ) __lowercase =SpeechaTextaTokenizer(os.path.join(_lowerCAmelCase , 'vocab.json' ) ) tokenizer.save_pretrained(_lowerCAmelCase ) __lowercase =hf_wavavec.config.to_dict() __lowercase =tokenizer.pad_token_id __lowercase =tokenizer.bos_token_id __lowercase =tokenizer.eos_token_id __lowercase ='speech_to_text_2' __lowercase ='wav2vec2' __lowercase =SpeechEncoderDecoderConfig.from_dict(_lowerCAmelCase ) hf_wavavec.save_pretrained(_lowerCAmelCase ) feature_extractor.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument( """--encoder_config_path""", default="""facebook/wav2vec2-large-lv60""", type=str, help="""Path to hf encoder wav2vec2 checkpoint config""", ) parser.add_argument( """--decoder_config_path""", default="""facebook/s2t-small-mustc-en-fr-st""", type=str, help="""Path to hf decoder s2t checkpoint config""", ) parser.add_argument("""--vocab_size""", default=1_0224, type=int, help="""Vocab size of decoder""") parser.add_argument("""--num_decoder_layers""", default=7, type=int, help="""Number of decoder layers""") lowerCamelCase = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
48
0
import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def a ( lowerCamelCase_ ): '''simple docstring''' lowercase__ = SwinConfig(image_size=192 ) if "base" in model_name: lowercase__ = 6 lowercase__ = 128 lowercase__ = (2, 2, 18, 2) lowercase__ = (4, 8, 16, 32) elif "large" in model_name: lowercase__ = 12 lowercase__ = 192 lowercase__ = (2, 2, 18, 2) lowercase__ = (6, 12, 24, 48) else: raise ValueError('''Model not supported, only supports base and large variants''' ) lowercase__ = window_size lowercase__ = embed_dim lowercase__ = depths lowercase__ = num_heads return config def a ( lowerCamelCase_ ): '''simple docstring''' if "encoder.mask_token" in name: lowercase__ = name.replace('''encoder.mask_token''' , '''embeddings.mask_token''' ) if "encoder.patch_embed.proj" in name: lowercase__ = name.replace('''encoder.patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "encoder.patch_embed.norm" in name: lowercase__ = name.replace('''encoder.patch_embed.norm''' , '''embeddings.norm''' ) if "attn.proj" in name: lowercase__ = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: lowercase__ = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: lowercase__ = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: lowercase__ = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: lowercase__ = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: lowercase__ = name.replace('''mlp.fc2''' , '''output.dense''' ) if name == "encoder.norm.weight": lowercase__ = "layernorm.weight" if name == "encoder.norm.bias": lowercase__ = "layernorm.bias" if "decoder" in name: pass else: lowercase__ = "swin." + name return name def a ( lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' for key in orig_state_dict.copy().keys(): lowercase__ = orig_state_dict.pop(SCREAMING_SNAKE_CASE_ ) if "attn_mask" in key: pass elif "qkv" in key: lowercase__ = key.split('''.''' ) lowercase__ = int(key_split[2] ) lowercase__ = int(key_split[4] ) lowercase__ = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowercase__ = val[:dim, :] lowercase__ = val[ dim : dim * 2, : ] lowercase__ = val[-dim:, :] else: lowercase__ = val[ :dim ] lowercase__ = val[ dim : dim * 2 ] lowercase__ = val[ -dim: ] else: lowercase__ = val return orig_state_dict def a ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' lowercase__ = torch.load(SCREAMING_SNAKE_CASE_ , map_location='''cpu''' )["model"] lowercase__ = get_swin_config(SCREAMING_SNAKE_CASE_ ) lowercase__ = SwinForMaskedImageModeling(SCREAMING_SNAKE_CASE_ ) model.eval() lowercase__ = convert_state_dict(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) model.load_state_dict(SCREAMING_SNAKE_CASE_ ) lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = ViTImageProcessor(size={'''height''': 192, '''width''': 192} ) lowercase__ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) lowercase__ = image_processor(images=SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ) with torch.no_grad(): lowercase__ = model(**SCREAMING_SNAKE_CASE_ ).logits print(outputs.keys() ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: print(F"""Pushing model and image processor for {model_name} to hub""" ) model.push_to_hub(F"""microsoft/{model_name}""" ) image_processor.push_to_hub(F"""microsoft/{model_name}""" ) if __name__ == "__main__": A__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='swin-base-simmim-window6-192', type=str, choices=['swin-base-simmim-window6-192', 'swin-large-simmim-window12-192'], help='Name of the Swin SimMIM model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default='/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth', type=str, help='Path to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) A__ : Dict = parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
207
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' lowerCamelCase : Tuple = "backbone." if is_semantic else "" lowerCamelCase : int = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""{prefix}blocks.{i}.norm1.weight""", f"""beit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""{prefix}blocks.{i}.norm1.bias""", f"""beit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append( (f"""{prefix}blocks.{i}.attn.proj.weight""", f"""beit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append( (f"""{prefix}blocks.{i}.attn.proj.bias""", f"""beit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""{prefix}blocks.{i}.norm2.weight""", f"""beit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""{prefix}blocks.{i}.norm2.bias""", f"""beit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""{prefix}blocks.{i}.mlp.fc1.weight""", f"""beit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""{prefix}blocks.{i}.mlp.fc1.bias""", f"""beit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""{prefix}blocks.{i}.mlp.fc2.weight""", f"""beit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""{prefix}blocks.{i}.mlp.fc2.bias""", f"""beit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ (f"""{prefix}cls_token""", "beit.embeddings.cls_token"), (f"""{prefix}patch_embed.proj.weight""", "beit.embeddings.patch_embeddings.projection.weight"), (f"""{prefix}patch_embed.proj.bias""", "beit.embeddings.patch_embeddings.projection.bias"), (f"""{prefix}pos_embed""", "beit.embeddings.position_embeddings"), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("mask_token", "beit.embeddings.mask_token"), ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) else: # layernorm + classification head rename_keys.extend( [ ("fc_norm.weight", "beit.pooler.layernorm.weight"), ("fc_norm.bias", "beit.pooler.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): lowerCamelCase : Optional[Any] = "backbone." if is_semantic else "" # queries, keys and values lowerCamelCase : Optional[Any] = state_dict.pop(f"""{prefix}blocks.{i}.attn.qkv.weight""" ) lowerCamelCase : Optional[Any] = state_dict.pop(f"""{prefix}blocks.{i}.attn.q_bias""" ) lowerCamelCase : Tuple = state_dict.pop(f"""{prefix}blocks.{i}.attn.v_bias""" ) lowerCamelCase : str = in_proj_weight[ : config.hidden_size, : ] lowerCamelCase : Any = q_bias lowerCamelCase : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase : int = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained lowerCamelCase : Any = state_dict.pop(f"""{prefix}blocks.{i}.gamma_1""" ) lowerCamelCase : Any = state_dict.pop(f"""{prefix}blocks.{i}.gamma_2""" ) lowerCamelCase : int = gamma_a lowerCamelCase : Optional[Any] = gamma_a def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : Optional[Any] = dct.pop(SCREAMING_SNAKE_CASE_ ) lowerCamelCase : List[Any] = val def lowercase_( ): '''simple docstring''' lowerCamelCase : Dict = "http://images.cocodataset.org/val2017/000000039769.jpg" lowerCamelCase : Optional[Any] = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) return im @torch.no_grad() def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' lowerCamelCase : List[Any] = False if "rvlcdip" in checkpoint_url else True lowerCamelCase : str = BeitConfig(use_absolute_position_embeddings=SCREAMING_SNAKE_CASE_ , use_mask_token=SCREAMING_SNAKE_CASE_ ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: lowerCamelCase : Union[str, Any] = 1024 lowerCamelCase : Any = 4096 lowerCamelCase : str = 24 lowerCamelCase : List[Any] = 16 # labels if "rvlcdip" in checkpoint_url: lowerCamelCase : Optional[Any] = 16 lowerCamelCase : Tuple = "huggingface/label-files" lowerCamelCase : List[str] = "rvlcdip-id2label.json" lowerCamelCase : str = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) lowerCamelCase : Any = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} lowerCamelCase : Tuple = idalabel lowerCamelCase : Dict = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys lowerCamelCase : int = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE_ , map_location="cpu" )["model"] lowerCamelCase : Tuple = create_rename_keys(SCREAMING_SNAKE_CASE_ , has_lm_head=SCREAMING_SNAKE_CASE_ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) read_in_q_k_v(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , has_lm_head=SCREAMING_SNAKE_CASE_ ) # load HuggingFace model lowerCamelCase : List[Any] = BeitForMaskedImageModeling(SCREAMING_SNAKE_CASE_ ) if has_lm_head else BeitForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # Check outputs on an image lowerCamelCase : str = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE_ ) lowerCamelCase : Any = prepare_img() lowerCamelCase : Optional[int] = image_processor(images=SCREAMING_SNAKE_CASE_ , return_tensors="pt" ) lowerCamelCase : Optional[Any] = encoding["pixel_values"] lowerCamelCase : Optional[Any] = model(SCREAMING_SNAKE_CASE_ ) lowerCamelCase : Dict = outputs.logits # verify logits lowerCamelCase : List[Any] = [1, 16] if "rvlcdip" in checkpoint_url else [1, 196, 8192] assert logits.shape == torch.Size(SCREAMING_SNAKE_CASE_ ), "Shape of logits not as expected" Path(SCREAMING_SNAKE_CASE_ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: if has_lm_head: lowerCamelCase : Optional[Any] = "dit-base" if "base" in checkpoint_url else "dit-large" else: lowerCamelCase : Dict = "dit-base-finetuned-rvlcdip" if "dit-b" in checkpoint_url else "dit-large-finetuned-rvlcdip" image_processor.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=SCREAMING_SNAKE_CASE_ , ) model.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=SCREAMING_SNAKE_CASE_ , ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_url''', default='''https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth''', type=str, help='''URL to the original PyTorch checkpoint (.pth file).''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) _snake_case = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
283
0
import string import numpy def A (__A : int , __A : int ) -> int: """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , __A ) class __snake_case : UpperCAmelCase__ : Union[str, Any] = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) UpperCAmelCase__ : Union[str, Any] = numpy.vectorize(lambda a : x % 3_6 ) UpperCAmelCase__ : List[str] = numpy.vectorize(a ) def __init__( self : str , _snake_case : numpy.ndarray): """simple docstring""" UpperCAmelCase_ = self.modulus(_snake_case) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key UpperCAmelCase_ = encrypt_key.shape[0] def lowerCamelCase ( self : Tuple , _snake_case : str): """simple docstring""" return self.key_string.index(_snake_case) def lowerCamelCase ( self : str , _snake_case : int): """simple docstring""" return self.key_string[round(_snake_case)] def lowerCamelCase ( self : Tuple): """simple docstring""" UpperCAmelCase_ = round(numpy.linalg.det(self.encrypt_key)) if det < 0: UpperCAmelCase_ = det % len(self.key_string) UpperCAmelCase_ = len(self.key_string) if greatest_common_divisor(_snake_case , len(self.key_string)) != 1: UpperCAmelCase_ = ( F"""determinant modular {req_l} of encryption key({det}) """ F"""is not co prime w.r.t {req_l}.\nTry another key.""" ) raise ValueError(_snake_case) def lowerCamelCase ( self : Optional[int] , _snake_case : str): """simple docstring""" UpperCAmelCase_ = [char for char in text.upper() if char in self.key_string] UpperCAmelCase_ = chars[-1] while len(_snake_case) % self.break_key != 0: chars.append(_snake_case) return "".join(_snake_case) def lowerCamelCase ( self : List[Any] , _snake_case : str): """simple docstring""" UpperCAmelCase_ = self.process_text(text.upper()) UpperCAmelCase_ = '''''' for i in range(0 , len(_snake_case) - self.break_key + 1 , self.break_key): UpperCAmelCase_ = text[i : i + self.break_key] UpperCAmelCase_ = [self.replace_letters(_snake_case) for char in batch] UpperCAmelCase_ = numpy.array([vec]).T UpperCAmelCase_ = self.modulus(self.encrypt_key.dot(_snake_case)).T.tolist()[ 0 ] UpperCAmelCase_ = ''''''.join( self.replace_digits(_snake_case) for num in batch_encrypted) encrypted += encrypted_batch return encrypted def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = round(numpy.linalg.det(self.encrypt_key)) if det < 0: UpperCAmelCase_ = det % len(self.key_string) UpperCAmelCase_ = None for i in range(len(self.key_string)): if (det * i) % len(self.key_string) == 1: UpperCAmelCase_ = i break UpperCAmelCase_ = ( det_inv * numpy.linalg.det(self.encrypt_key) * numpy.linalg.inv(self.encrypt_key) ) return self.to_int(self.modulus(_snake_case)) def lowerCamelCase ( self : Dict , _snake_case : str): """simple docstring""" UpperCAmelCase_ = self.make_decrypt_key() UpperCAmelCase_ = self.process_text(text.upper()) UpperCAmelCase_ = '''''' for i in range(0 , len(_snake_case) - self.break_key + 1 , self.break_key): UpperCAmelCase_ = text[i : i + self.break_key] UpperCAmelCase_ = [self.replace_letters(_snake_case) for char in batch] UpperCAmelCase_ = numpy.array([vec]).T UpperCAmelCase_ = self.modulus(decrypt_key.dot(_snake_case)).T.tolist()[0] UpperCAmelCase_ = ''''''.join( self.replace_digits(_snake_case) for num in batch_decrypted) decrypted += decrypted_batch return decrypted def A () -> None: """simple docstring""" UpperCAmelCase_ = int(input('''Enter the order of the encryption key: ''' ) ) UpperCAmelCase_ = [] print('''Enter each row of the encryption key with space separated integers''' ) for _ in range(__A ): UpperCAmelCase_ = [int(__A ) for x in input().split()] hill_matrix.append(__A ) UpperCAmelCase_ = HillCipher(numpy.array(__A ) ) print('''Would you like to encrypt or decrypt some text? (1 or 2)''' ) UpperCAmelCase_ = input('''\n1. Encrypt\n2. Decrypt\n''' ) if option == "1": UpperCAmelCase_ = input('''What text would you like to encrypt?: ''' ) print('''Your encrypted text is:''' ) print(hc.encrypt(__A ) ) elif option == "2": UpperCAmelCase_ = input('''What text would you like to decrypt?: ''' ) print('''Your decrypted text is:''' ) print(hc.decrypt(__A ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
7
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
7
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''edbeeching/decision-transformer-gym-hopper-medium''': ( '''https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json''' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class lowerCamelCase_ (snake_case__ ): '''simple docstring''' __UpperCamelCase: Tuple = "decision_transformer" __UpperCamelCase: Any = ["past_key_values"] __UpperCamelCase: Dict = { "max_position_embeddings": "n_positions", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Optional[int] , A : Optional[Any]=17 , A : str=4 , A : Dict=128 , A : int=4096 , A : List[Any]=True , A : Dict=1 , A : Union[str, Any]=1024 , A : Union[str, Any]=3 , A : Any=1 , A : List[Any]=None , A : Dict="relu" , A : Optional[Any]=0.1 , A : str=0.1 , A : List[str]=0.1 , A : List[Any]=1E-5 , A : List[Any]=0.02 , A : Tuple=True , A : List[str]=True , A : int=50256 , A : int=50256 , A : Tuple=False , A : Any=False , **A : List[Any] , ): _UpperCAmelCase : Optional[int] = state_dim _UpperCAmelCase : int = act_dim _UpperCAmelCase : List[str] = hidden_size _UpperCAmelCase : Any = max_ep_len _UpperCAmelCase : Any = action_tanh _UpperCAmelCase : Optional[Any] = vocab_size _UpperCAmelCase : Optional[Any] = n_positions _UpperCAmelCase : Dict = n_layer _UpperCAmelCase : int = n_head _UpperCAmelCase : Tuple = n_inner _UpperCAmelCase : Union[str, Any] = activation_function _UpperCAmelCase : str = resid_pdrop _UpperCAmelCase : Union[str, Any] = embd_pdrop _UpperCAmelCase : List[Any] = attn_pdrop _UpperCAmelCase : Any = layer_norm_epsilon _UpperCAmelCase : Any = initializer_range _UpperCAmelCase : str = scale_attn_weights _UpperCAmelCase : Optional[Any] = use_cache _UpperCAmelCase : List[Any] = scale_attn_by_inverse_layer_idx _UpperCAmelCase : List[Any] = reorder_and_upcast_attn _UpperCAmelCase : Optional[int] = bos_token_id _UpperCAmelCase : int = eos_token_id super().__init__(bos_token_id=A , eos_token_id=A , **A )
31
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase _lowerCamelCase : List[Any] = logging.get_logger(__name__) _lowerCamelCase : Union[str, Any] = { '''allenai/longformer-base-4096''': '''https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json''', '''allenai/longformer-large-4096''': '''https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json''', '''allenai/longformer-large-4096-finetuned-triviaqa''': ( '''https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json''' ), '''allenai/longformer-base-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json''' ), '''allenai/longformer-large-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : Dict = "longformer" def __init__( self : Optional[Any] , lowercase : Union[List[int], int] = 512 , lowercase : int = 2 , lowercase : int = 1 , lowercase : int = 0 , lowercase : int = 2 , lowercase : int = 30_522 , lowercase : int = 768 , lowercase : int = 12 , lowercase : int = 12 , lowercase : int = 3_072 , lowercase : str = "gelu" , lowercase : float = 0.1 , lowercase : float = 0.1 , lowercase : int = 512 , lowercase : int = 2 , lowercase : float = 0.02 , lowercase : float = 1E-12 , lowercase : bool = False , **lowercase : Optional[Any] , ): '''simple docstring''' super().__init__(pad_token_id=lowercase , **lowercase ) _snake_case = attention_window _snake_case = sep_token_id _snake_case = bos_token_id _snake_case = eos_token_id _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = onnx_export class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : int , lowercase : "PretrainedConfig" , lowercase : str = "default" , lowercase : "List[PatchingSpec]" = None ): '''simple docstring''' super().__init__(lowercase , lowercase , lowercase ) _snake_case = True @property def A ( self : Union[str, Any] ): '''simple docstring''' if self.task == "multiple-choice": _snake_case = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _snake_case = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('global_attention_mask', dynamic_axis), ] ) @property def A ( self : int ): '''simple docstring''' _snake_case = super().outputs if self.task == "default": _snake_case = {0: 'batch'} return outputs @property def A ( self : List[Any] ): '''simple docstring''' return 1E-4 @property def A ( self : List[str] ): '''simple docstring''' return max(super().default_onnx_opset , 14 ) def A ( self : str , lowercase : "PreTrainedTokenizerBase" , lowercase : int = -1 , lowercase : int = -1 , lowercase : bool = False , lowercase : Optional[TensorType] = None , ): '''simple docstring''' _snake_case = super().generate_dummy_inputs( preprocessor=lowercase , batch_size=lowercase , seq_length=lowercase , is_pair=lowercase , framework=lowercase ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly _snake_case = torch.zeros_like(inputs['input_ids'] ) # make every second token global _snake_case = 1 return inputs
282
0
import random from .binary_exp_mod import bin_exp_mod def snake_case__ ( SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : int=1_000 ): '''simple docstring''' if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd lowercase__ : str = n - 1 lowercase__ : Dict = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) lowercase__ : List[str] = 0 while count < prec: lowercase__ : Union[str, Any] = random.randint(2 , n - 1 ) lowercase__ : str = bin_exp_mod(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if b != 1: lowercase__ : Optional[int] = True for _ in range(SCREAMING_SNAKE_CASE_ ): if b == n - 1: lowercase__ : List[Any] = False break lowercase__ : Optional[Any] = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": snake_case_ = abs(int(input('''Enter bound : ''').strip())) print('''Here\'s the list of primes:''') print(''', '''.join(str(i) for i in range(n + 1) if is_prime_big(i)))
216
def snake_case__ ( SCREAMING_SNAKE_CASE_ : int = 50_000_000 ): '''simple docstring''' lowercase__ : List[Any] = set() lowercase__ : Any = int((limit - 24) ** (1 / 2) ) lowercase__ : Any = set(range(3 , prime_square_limit + 1 , 2 ) ) primes.add(2 ) for p in range(3 , prime_square_limit + 1 , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , prime_square_limit + 1 , SCREAMING_SNAKE_CASE_ ) ) ) for primea in primes: lowercase__ : Any = primea * primea for primea in primes: lowercase__ : Optional[int] = primea * primea * primea if square + cube >= limit - 16: break for primea in primes: lowercase__ : Dict = primea * primea * primea * primea lowercase__ : List[Any] = square + cube + tetr if total >= limit: break ret.add(SCREAMING_SNAKE_CASE_ ) return len(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": print(F'''{solution() = }''')
216
1
"""simple docstring""" def lowercase__ ( snake_case_ :list ): if any(not isinstance(snake_case_ , snake_case_ ) or x < 0 for x in sequence ): raise TypeError('''Sequence must be list of non-negative integers''' ) for _ in range(len(snake_case_ ) ): for i, (rod_upper, rod_lower) in enumerate(zip(snake_case_ , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
332
"""simple docstring""" from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class _UpperCAmelCase : a__ : int a__ : Node | None = None a__ : Node | None = None def lowercase__ ( ): __UpperCAmelCase = Node(1 ) __UpperCAmelCase = Node(2 ) __UpperCAmelCase = Node(3 ) __UpperCAmelCase = Node(4 ) __UpperCAmelCase = Node(5 ) return tree def lowercase__ ( snake_case_ :Node | None ): return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def lowercase__ ( snake_case_ :Node | None ): return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def lowercase__ ( snake_case_ :Node | None ): return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def lowercase__ ( snake_case_ :Node | None ): return (max(height(root.left ) , height(root.right ) ) + 1) if root else 0 def lowercase__ ( snake_case_ :Node | None ): __UpperCAmelCase = [] if root is None: return output __UpperCAmelCase = deque([root] ) while process_queue: __UpperCAmelCase = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def lowercase__ ( snake_case_ :Node | None , snake_case_ :int ): __UpperCAmelCase = [] def populate_output(snake_case_ :Node | None , snake_case_ :int ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left , level - 1 ) populate_output(root.right , level - 1 ) populate_output(snake_case_ , snake_case_ ) return output def lowercase__ ( snake_case_ :Node | None , snake_case_ :int ): __UpperCAmelCase = [] def populate_output(snake_case_ :Node | None , snake_case_ :int ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right , level - 1 ) populate_output(root.left , level - 1 ) populate_output(snake_case_ , snake_case_ ) return output def lowercase__ ( snake_case_ :Node | None ): if root is None: return [] __UpperCAmelCase = [] __UpperCAmelCase = 0 __UpperCAmelCase = height(snake_case_ ) for h in range(1 , height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(snake_case_ , snake_case_ ) ) __UpperCAmelCase = 1 else: output.append(get_nodes_from_right_to_left(snake_case_ , snake_case_ ) ) __UpperCAmelCase = 0 return output def lowercase__ ( ): # Main function for testing. __UpperCAmelCase = make_tree() print(F'''In-order Traversal: {inorder(snake_case_ )}''' ) print(F'''Pre-order Traversal: {preorder(snake_case_ )}''' ) print(F'''Post-order Traversal: {postorder(snake_case_ )}''' , '''\n''' ) print(F'''Height of Tree: {height(snake_case_ )}''' , '''\n''' ) print('''Complete Level Order Traversal: ''' ) print(level_order(snake_case_ ) , '''\n''' ) print('''Level-wise order Traversal: ''' ) for level in range(1 , height(snake_case_ ) + 1 ): print(F'''Level {level}:''' , get_nodes_from_left_to_right(snake_case_ , level=snake_case_ ) ) print('''\nZigZag order Traversal: ''' ) print(zigzag(snake_case_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
332
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class __SCREAMING_SNAKE_CASE ( __lowercase): _SCREAMING_SNAKE_CASE : Union[List[PIL.Image.Image], np.ndarray] _SCREAMING_SNAKE_CASE : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.26.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(""">=""", """0.0.12""") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class __SCREAMING_SNAKE_CASE ( __lowercase): _SCREAMING_SNAKE_CASE : np.ndarray _SCREAMING_SNAKE_CASE : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
364
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __SCREAMING_SNAKE_CASE ( __lowercase , unittest.TestCase): _SCREAMING_SNAKE_CASE : int = KandinskyVaaInpaintPipeline _SCREAMING_SNAKE_CASE : int = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''mask_image'''] _SCREAMING_SNAKE_CASE : Optional[Any] = [ '''image_embeds''', '''negative_image_embeds''', '''image''', '''mask_image''', ] _SCREAMING_SNAKE_CASE : List[Any] = [ '''generator''', '''height''', '''width''', '''latents''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] _SCREAMING_SNAKE_CASE : Optional[Any] = False @property def UpperCamelCase__ ( self ): """simple docstring""" return 32 @property def UpperCamelCase__ ( self ): """simple docstring""" return 32 @property def UpperCamelCase__ ( self ): """simple docstring""" return self.time_input_dim @property def UpperCamelCase__ ( self ): """simple docstring""" return self.time_input_dim * 4 @property def UpperCamelCase__ ( self ): """simple docstring""" return 1_00 @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } lowerCAmelCase__ = UNetaDConditionModel(**_UpperCamelCase ) return model @property def UpperCamelCase__ ( self ): """simple docstring""" return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ = VQModel(**self.dummy_movq_kwargs ) return model def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.dummy_unet lowerCAmelCase__ = self.dummy_movq lowerCAmelCase__ = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=_UpperCamelCase , set_alpha_to_one=_UpperCamelCase , steps_offset=1 , prediction_type='epsilon' , thresholding=_UpperCamelCase , ) lowerCAmelCase__ = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase=0 ): """simple docstring""" lowerCAmelCase__ = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) lowerCAmelCase__ = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _UpperCamelCase ) # create init_image lowerCAmelCase__ = floats_tensor((1, 3, 64, 64) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) lowerCAmelCase__ = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase__ = Image.fromarray(np.uinta(_UpperCamelCase ) ).convert('RGB' ).resize((2_56, 2_56) ) # create mask lowerCAmelCase__ = np.ones((64, 64) , dtype=np.floataa ) lowerCAmelCase__ = 0 if str(_UpperCamelCase ).startswith('mps' ): lowerCAmelCase__ = torch.manual_seed(_UpperCamelCase ) else: lowerCAmelCase__ = torch.Generator(device=_UpperCamelCase ).manual_seed(_UpperCamelCase ) lowerCAmelCase__ = { 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = 'cpu' lowerCAmelCase__ = self.get_dummy_components() lowerCAmelCase__ = self.pipeline_class(**_UpperCamelCase ) lowerCAmelCase__ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase__ = pipe(**self.get_dummy_inputs(_UpperCamelCase ) ) lowerCAmelCase__ = output.images lowerCAmelCase__ = pipe( **self.get_dummy_inputs(_UpperCamelCase ) , return_dict=_UpperCamelCase , )[0] lowerCAmelCase__ = image[0, -3:, -3:, -1] lowerCAmelCase__ = image_from_tuple[0, -3:, -3:, -1] print(F"image.shape {image.shape}" ) assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ = np.array( [0.50_77_59_03, 0.49_52_71_95, 0.48_82_45_43, 0.50_19_22_37, 0.48_64_49_06, 0.49_37_38_14, 0.4_78_05_98, 0.47_23_48_27, 0.48_32_78_48] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" def UpperCamelCase__ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase): def UpperCamelCase__ ( self ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy' ) lowerCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) lowerCAmelCase__ = np.ones((7_68, 7_68) , dtype=np.floataa ) lowerCAmelCase__ = 0 lowerCAmelCase__ = 'a hat' lowerCAmelCase__ = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(_UpperCamelCase ) lowerCAmelCase__ = KandinskyVaaInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-decoder-inpaint' , torch_dtype=torch.floataa ) lowerCAmelCase__ = pipeline.to(_UpperCamelCase ) pipeline.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) lowerCAmelCase__ , lowerCAmelCase__ = pipe_prior( _UpperCamelCase , generator=_UpperCamelCase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() lowerCAmelCase__ = pipeline( image=_UpperCamelCase , mask_image=_UpperCamelCase , image_embeds=_UpperCamelCase , negative_image_embeds=_UpperCamelCase , generator=_UpperCamelCase , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type='np' , ) lowerCAmelCase__ = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(_UpperCamelCase , _UpperCamelCase )
122
0
import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration _lowerCamelCase =pytest.mark.integration _lowerCamelCase ={"""comet"""} _lowerCamelCase =importlib.util.find_spec("""fairseq""") is not None _lowerCamelCase ={"""code_eval"""} _lowerCamelCase =os.name == """nt""" _lowerCamelCase ={"""bertscore""", """frugalscore""", """perplexity"""} _lowerCamelCase =importlib.util.find_spec("""transformers""") is not None def _a ( lowerCamelCase ): @wraps(lowerCamelCase ) def wrapper(self, lowerCamelCase ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest("""\"test requires Fairseq\"""" ) else: test_case(self, lowerCamelCase ) return wrapper def _a ( lowerCamelCase ): @wraps(lowerCamelCase ) def wrapper(self, lowerCamelCase ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest("""\"test requires transformers\"""" ) else: test_case(self, lowerCamelCase ) return wrapper def _a ( lowerCamelCase ): @wraps(lowerCamelCase ) def wrapper(self, lowerCamelCase ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest("""\"test not supported on Windows\"""" ) else: test_case(self, lowerCamelCase ) return wrapper def _a ( ): lowerCamelCase : Union[str, Any] = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob("""./metrics/*/""" )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names()) @for_all_test_methods( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) @local class A__ ( parameterized.TestCase): _UpperCAmelCase : Optional[Any] = {} _UpperCAmelCase : List[Any] = None @pytest.mark.filterwarnings("""ignore:metric_module_factory is deprecated:FutureWarning""" ) @pytest.mark.filterwarnings("""ignore:load_metric is deprecated:FutureWarning""" ) def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : List[Any] = """[...]""" lowerCamelCase : Any = importlib.import_module( datasets.load.metric_module_factory(os.path.join("""metrics""" , __magic_name__ ) ).module_path ) lowerCamelCase : Optional[Any] = datasets.load.import_main_class(metric_module.__name__ , dataset=__magic_name__ ) # check parameters lowerCamelCase : Optional[int] = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(__magic_name__ , metric_module.__name__ ): with self.use_local_metrics(): try: lowerCamelCase : List[str] = doctest.testmod(__magic_name__ , verbose=__magic_name__ , raise_on_error=__magic_name__ ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : str = """[...]""" lowerCamelCase : int = importlib.import_module( datasets.load.metric_module_factory(os.path.join("""metrics""" , __magic_name__ ) ).module_path ) # run doctest with self.use_local_metrics(): lowerCamelCase : str = doctest.testmod(__magic_name__ , verbose=__magic_name__ , raise_on_error=__magic_name__ ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](__magic_name__ ): yield else: yield @contextmanager def UpperCamelCase__ ( self ): def load_local_metric(__magic_name__ , *__magic_name__ , **__magic_name__ ): return load_metric(os.path.join("""metrics""" , __magic_name__ ) , *__magic_name__ , **__magic_name__ ) with patch("""datasets.load_metric""" ) as mock_load_metric: lowerCamelCase : List[Any] = load_local_metric yield @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): def wrapper(__magic_name__ ): lowerCamelCase : int = contextmanager(__magic_name__ ) lowerCamelCase : Optional[int] = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher("""bleurt""" ) def _a ( lowerCamelCase ): import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string("""sv""", """""", """""" ) # handle pytest cli flags class A__ ( __SCREAMING_SNAKE_CASE): def UpperCamelCase__ ( self , __magic_name__ ): assert len(input_dict["""input_ids"""] ) == 2 return np.array([1.03, 1.04] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch("""bleurt.score._create_predictor""" ) as mock_create_predictor: lowerCamelCase : List[str] = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher("""bertscore""" ) def _a ( lowerCamelCase ): import torch def bert_cos_score_idf(lowerCamelCase, lowerCamelCase, *lowerCamelCase, **lowerCamelCase ): return torch.tensor([[1.0, 1.0, 1.0]] * len(lowerCamelCase ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch("""bert_score.scorer.get_model""" ), patch( """bert_score.scorer.bert_cos_score_idf""" ) as mock_bert_cos_score_idf: lowerCamelCase : List[Any] = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher("""comet""" ) def _a ( lowerCamelCase ): def load_from_checkpoint(lowerCamelCase ): class A__ : def UpperCamelCase__ ( self , __magic_name__ , *__magic_name__ , **__magic_name__ ): assert len(__magic_name__ ) == 2 lowerCamelCase : Tuple = [0.19, 0.92] return scores, sum(__magic_name__ ) / len(__magic_name__ ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch("""comet.download_model""" ) as mock_download_model: lowerCamelCase : Optional[int] = None with patch("""comet.load_from_checkpoint""" ) as mock_load_from_checkpoint: lowerCamelCase : Dict = load_from_checkpoint yield def _a ( ): lowerCamelCase : int = load_metric(os.path.join("""metrics""", """seqeval""" ) ) lowerCamelCase : int = """ERROR""" lowerCamelCase : List[Any] = F'''Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}''' with pytest.raises(lowerCamelCase, match=re.escape(lowerCamelCase ) ): metric.compute(predictions=[], references=[], scheme=lowerCamelCase )
287
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """nvidia/segformer-b0-finetuned-ade-512-512""": ( """https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json""" ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """segformer""" def __init__( self , __magic_name__=3 , __magic_name__=4 , __magic_name__=[2, 2, 2, 2] , __magic_name__=[8, 4, 2, 1] , __magic_name__=[3_2, 6_4, 1_6_0, 2_5_6] , __magic_name__=[7, 3, 3, 3] , __magic_name__=[4, 2, 2, 2] , __magic_name__=[1, 2, 5, 8] , __magic_name__=[4, 4, 4, 4] , __magic_name__="gelu" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=0.02 , __magic_name__=0.1 , __magic_name__=1e-6 , __magic_name__=2_5_6 , __magic_name__=2_5_5 , **__magic_name__ , ): super().__init__(**__magic_name__ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , __magic_name__ , ) lowerCamelCase : Optional[Any] = num_channels lowerCamelCase : str = num_encoder_blocks lowerCamelCase : Any = depths lowerCamelCase : List[Any] = sr_ratios lowerCamelCase : int = hidden_sizes lowerCamelCase : Union[str, Any] = patch_sizes lowerCamelCase : Optional[Any] = strides lowerCamelCase : Dict = mlp_ratios lowerCamelCase : str = num_attention_heads lowerCamelCase : Any = hidden_act lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase : Dict = classifier_dropout_prob lowerCamelCase : Tuple = initializer_range lowerCamelCase : Dict = drop_path_rate lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : Any = decoder_hidden_size lowerCamelCase : str = kwargs.get("""reshape_last_stage""" , __magic_name__ ) lowerCamelCase : Dict = semantic_loss_ignore_index class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-4 @property def UpperCamelCase__ ( self ): return 1_2
287
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from ...utils.dataclasses import ( ComputeEnvironment, DistributedType, DynamoBackend, PrecisionType, SageMakerDistributedType, ) from ..menu import BulletMenu _lowerCamelCase : Optional[int] = [ "EAGER", "AOT_EAGER", "INDUCTOR", "NVFUSER", "AOT_NVFUSER", "AOT_CUDAGRAPHS", "OFI", "FX2TRT", "ONNXRT", "IPEX", ] def _UpperCAmelCase (UpperCamelCase_ : List[str] , UpperCamelCase_ : List[Any]=None , UpperCamelCase_ : List[Any]=None , UpperCamelCase_ : int=None ): '''simple docstring''' _lowerCAmelCase : Dict = True while ask_again: _lowerCAmelCase : Tuple = input(lowerCamelCase_ ) try: if default is not None and len(lowerCamelCase_ ) == 0: return default return convert_value(lowerCamelCase_ ) if convert_value is not None else result except Exception: if error_message is not None: print(lowerCamelCase_ ) def _UpperCAmelCase (UpperCamelCase_ : List[Any] , UpperCamelCase_ : int=[] , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : Optional[int]=0 ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = BulletMenu(lowerCamelCase_ , lowerCamelCase_ ) _lowerCAmelCase : Any = menu.run(default_choice=lowerCamelCase_ ) return convert_value(lowerCamelCase_ ) if convert_value is not None else result def _UpperCAmelCase (UpperCamelCase_ : int ): '''simple docstring''' _lowerCAmelCase : str = int(lowerCamelCase_ ) return ComputeEnvironment(["""LOCAL_MACHINE""", """AMAZON_SAGEMAKER"""][value] ) def _UpperCAmelCase (UpperCamelCase_ : Dict ): '''simple docstring''' _lowerCAmelCase : Optional[int] = int(lowerCamelCase_ ) return DistributedType(["""NO""", """MULTI_CPU""", """MULTI_XPU""", """MULTI_GPU""", """MULTI_NPU""", """TPU"""][value] ) def _UpperCAmelCase (UpperCamelCase_ : Union[str, Any] ): '''simple docstring''' _lowerCAmelCase : str = int(lowerCamelCase_ ) return DynamoBackend(DYNAMO_BACKENDS[value] ).value def _UpperCAmelCase (UpperCamelCase_ : Union[str, Any] ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = int(lowerCamelCase_ ) return PrecisionType(["""no""", """fp16""", """bf16""", """fp8"""][value] ) def _UpperCAmelCase (UpperCamelCase_ : Union[str, Any] ): '''simple docstring''' _lowerCAmelCase : Any = int(lowerCamelCase_ ) return SageMakerDistributedType(["""NO""", """DATA_PARALLEL""", """MODEL_PARALLEL"""][value] ) def _UpperCAmelCase (UpperCamelCase_ : Dict ): '''simple docstring''' return {"yes": True, "no": False}[value.lower()] class __snake_case (argparse.RawDescriptionHelpFormatter ): def SCREAMING_SNAKE_CASE ( self : Optional[int] , _UpperCAmelCase : int , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : str ) -> Optional[int]: '''simple docstring''' _lowerCAmelCase : int = super()._format_usage(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : Dict = usage.replace("""<command> [<args>] """ , """""" ) return usage
368
from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def _UpperCAmelCase (UpperCamelCase_ : Sequence[float] , UpperCamelCase_ : int , UpperCamelCase_ : int ): '''simple docstring''' if not arr: return None, None, 0 if low == high: return low, high, arr[low] _lowerCAmelCase : List[str] = (low + high) // 2 _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Union[str, Any] = max_subarray(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : List[Any] = max_subarray(UpperCamelCase_ , mid + 1 , UpperCamelCase_ ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[int] = max_cross_sum(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def _UpperCAmelCase (UpperCamelCase_ : Sequence[float] , UpperCamelCase_ : int , UpperCamelCase_ : int , UpperCamelCase_ : int ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : Optional[int] = float("""-inf""" ), -1 _lowerCAmelCase , _lowerCAmelCase : str = float("""-inf""" ), -1 _lowerCAmelCase : int | float = 0 for i in range(UpperCamelCase_ , low - 1 , -1 ): summ += arr[i] if summ > left_sum: _lowerCAmelCase : Any = summ _lowerCAmelCase : Tuple = i _lowerCAmelCase : int = 0 for i in range(mid + 1 , high + 1 ): summ += arr[i] if summ > right_sum: _lowerCAmelCase : List[Any] = summ _lowerCAmelCase : str = i return max_left, max_right, (left_sum + right_sum) def _UpperCAmelCase (UpperCamelCase_ : int ): '''simple docstring''' _lowerCAmelCase : str = [randint(1 , UpperCamelCase_ ) for _ in range(UpperCamelCase_ )] _lowerCAmelCase : str = time.time() max_subarray(UpperCamelCase_ , 0 , input_size - 1 ) _lowerCAmelCase : Any = time.time() return end - start def _UpperCAmelCase (): '''simple docstring''' _lowerCAmelCase : Any = [10, 100, 1000, 10000, 50000, 100000, 200000, 300000, 400000, 500000] _lowerCAmelCase : Any = [time_max_subarray(UpperCamelCase_ ) for input_size in input_sizes] print("""No of Inputs\t\tTime Taken""" ) for input_size, runtime in zip(UpperCamelCase_ , UpperCamelCase_ ): print(UpperCamelCase_ , """\t\t""" , UpperCamelCase_ ) plt.plot(UpperCamelCase_ , UpperCamelCase_ ) plt.xlabel("""Number of Inputs""" ) plt.ylabel("""Time taken in seconds""" ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
159
0
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings _lowerCAmelCase = R''' [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `" / "`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `" // "`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `"wiki_dpr"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `"train"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `"compressed"`) The index name of the index associated with the `dataset`. One can choose between `"legacy"`, `"exact"` and `"compressed"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a "dummy" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. ''' @add_start_docstrings(SCREAMING_SNAKE_CASE_ ) class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' __lowercase : str = '''rag''' __lowercase : Any = True def __init__( self ,__UpperCAmelCase=None ,__UpperCAmelCase=True ,__UpperCAmelCase=None ,__UpperCAmelCase=None ,__UpperCAmelCase=None ,__UpperCAmelCase=None ,__UpperCAmelCase=None ,__UpperCAmelCase=" / " ,__UpperCAmelCase=" // " ,__UpperCAmelCase=5 ,__UpperCAmelCase=300 ,__UpperCAmelCase=768 ,__UpperCAmelCase=8 ,__UpperCAmelCase="wiki_dpr" ,__UpperCAmelCase="train" ,__UpperCAmelCase="compressed" ,__UpperCAmelCase=None ,__UpperCAmelCase=None ,__UpperCAmelCase=False ,__UpperCAmelCase=False ,__UpperCAmelCase=0.0 ,__UpperCAmelCase=True ,__UpperCAmelCase=False ,__UpperCAmelCase=False ,__UpperCAmelCase=False ,__UpperCAmelCase=True ,__UpperCAmelCase=None ,**__UpperCAmelCase ,) -> Optional[int]: super().__init__( bos_token_id=__UpperCAmelCase ,pad_token_id=__UpperCAmelCase ,eos_token_id=__UpperCAmelCase ,decoder_start_token_id=__UpperCAmelCase ,forced_eos_token_id=__UpperCAmelCase ,is_encoder_decoder=__UpperCAmelCase ,prefix=__UpperCAmelCase ,vocab_size=__UpperCAmelCase ,**__UpperCAmelCase ,) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" lowerCAmelCase__ : List[str] = kwargs.pop("""question_encoder""" ) lowerCAmelCase__ : Dict = question_encoder_config.pop("""model_type""" ) lowerCAmelCase__ : Optional[Any] = kwargs.pop("""generator""" ) lowerCAmelCase__ : Any = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig lowerCAmelCase__ : Optional[int] = AutoConfig.for_model(__UpperCAmelCase ,**__UpperCAmelCase ) lowerCAmelCase__ : Union[str, Any] = AutoConfig.for_model(__UpperCAmelCase ,**__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = reduce_loss lowerCAmelCase__ : Dict = label_smoothing lowerCAmelCase__ : List[Any] = exclude_bos_score lowerCAmelCase__ : Union[str, Any] = do_marginalize lowerCAmelCase__ : Tuple = title_sep lowerCAmelCase__ : Union[str, Any] = doc_sep lowerCAmelCase__ : Union[str, Any] = n_docs lowerCAmelCase__ : Union[str, Any] = max_combined_length lowerCAmelCase__ : str = dataset lowerCAmelCase__ : List[Any] = dataset_split lowerCAmelCase__ : Optional[Any] = index_name lowerCAmelCase__ : Dict = retrieval_vector_size lowerCAmelCase__ : Tuple = retrieval_batch_size lowerCAmelCase__ : Optional[Any] = passages_path lowerCAmelCase__ : Union[str, Any] = index_path lowerCAmelCase__ : Dict = use_dummy_dataset lowerCAmelCase__ : Optional[int] = output_retrieved lowerCAmelCase__ : int = do_deduplication lowerCAmelCase__ : Optional[Any] = use_cache if self.forced_eos_token_id is None: lowerCAmelCase__ : Tuple = getattr(self.generator ,"""forced_eos_token_id""" ,__UpperCAmelCase ) @classmethod def UpperCAmelCase_ ( cls ,__UpperCAmelCase ,__UpperCAmelCase ,**__UpperCAmelCase ) -> PretrainedConfig: return cls(question_encoder=question_encoder_config.to_dict() ,generator=generator_config.to_dict() ,**__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> str: lowerCAmelCase__ : Optional[int] = copy.deepcopy(self.__dict__ ) lowerCAmelCase__ : Optional[Any] = self.question_encoder.to_dict() lowerCAmelCase__ : List[Any] = self.generator.to_dict() lowerCAmelCase__ : Tuple = self.__class__.model_type return output
37
from __future__ import annotations def lowerCAmelCase__( lowercase : str , lowercase : list[str] | None = None ) -> list[list[str]]: __snake_case : List[str] = word_bank or [] # create a table __snake_case : int = len(lowercase ) + 1 __snake_case : list[list[list[str]]] = [] for _ in range(lowercase ): table.append([] ) # seed value __snake_case : Optional[int] = [[]] # because empty string has empty combination # iterate through the indices for i in range(lowercase ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(lowercase )] == word: __snake_case : list[list[str]] = [ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(lowercase )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(lowercase )]: combination.reverse() return table[len(lowercase )] if __name__ == "__main__": print(all_construct('''jwajalapa''', ['''jwa''', '''j''', '''w''', '''a''', '''la''', '''lapa'''])) print(all_construct('''rajamati''', ['''s''', '''raj''', '''amat''', '''raja''', '''ma''', '''i''', '''t'''])) print( all_construct( '''hexagonosaurus''', ['''h''', '''ex''', '''hex''', '''ag''', '''ago''', '''ru''', '''auru''', '''rus''', '''go''', '''no''', '''o''', '''s'''], ) )
326
0
'''simple docstring''' from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging lowerCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a =['pixel_values'] def __init__( self : Optional[int] , __a : bool = True , __a : Dict[str, int] = None , __a : PILImageResampling = PILImageResampling.BICUBIC , __a : bool = True , __a : Dict[str, int] = None , __a : bool = True , __a : Union[int, float] = 1 / 2_55 , __a : bool = True , __a : Optional[Union[float, Iterable[float]]] = IMAGENET_DEFAULT_MEAN , __a : Optional[Union[float, Iterable[float]]] = IMAGENET_DEFAULT_STD , **__a : str , ): super().__init__(**__UpperCAmelCase ) _a = size if size is not None else {"shortest_edge": 2_24} _a = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) _a = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} _a = get_size_dict(__UpperCAmelCase , param_name="crop_size" ) _a = do_resize _a = size _a = resample _a = do_center_crop _a = crop_size _a = do_rescale _a = rescale_factor _a = do_normalize _a = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN _a = image_std if image_std is not None else IMAGENET_DEFAULT_STD def UpperCamelCase__ ( self : Tuple , __a : np.ndarray , __a : Dict[str, int] , __a : PILImageResampling = PILImageResampling.BICUBIC , __a : Optional[Union[str, ChannelDimension]] = None , **__a : Tuple , ): _a = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: _a = int((2_56 / 2_24) * size["shortest_edge"] ) _a = get_resize_output_image_size(__UpperCAmelCase , size=__UpperCAmelCase , default_to_square=__UpperCAmelCase ) _a = {"height": output_size[0], "width": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( f'Size dict must have keys \'height\' and \'width\' or \'shortest_edge\'. Got {size_dict.keys()}' ) return resize( __UpperCAmelCase , size=(size_dict["height"], size_dict["width"]) , resample=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def UpperCamelCase__ ( self : Tuple , __a : np.ndarray , __a : Dict[str, int] , __a : Optional[Union[str, ChannelDimension]] = None , **__a : Optional[int] , ): _a = get_size_dict(__UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'Size dict must have keys \'height\' and \'width\'. Got {size.keys()}' ) return center_crop(__UpperCAmelCase , size=(size["height"], size["width"]) , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def UpperCamelCase__ ( self : Union[str, Any] , __a : np.ndarray , __a : Union[int, float] , __a : Optional[Union[str, ChannelDimension]] = None , **__a : Optional[Any] , ): return rescale(__UpperCAmelCase , scale=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def UpperCamelCase__ ( self : str , __a : np.ndarray , __a : Union[float, List[float]] , __a : Union[float, List[float]] , __a : Optional[Union[str, ChannelDimension]] = None , **__a : Union[str, Any] , ): return normalize(__UpperCAmelCase , mean=__UpperCAmelCase , std=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def UpperCamelCase__ ( self : Optional[Any] , __a : ImageInput , __a : Optional[bool] = None , __a : Optional[Dict[str, int]] = None , __a : PILImageResampling = None , __a : Optional[bool] = None , __a : Optional[Dict[str, int]] = None , __a : Optional[bool] = None , __a : Optional[float] = None , __a : Optional[bool] = None , __a : Optional[Union[float, Iterable[float]]] = None , __a : Optional[Union[float, Iterable[float]]] = None , __a : Optional[TensorType] = None , __a : ChannelDimension = ChannelDimension.FIRST , **__a : List[Any] , ): _a = do_resize if do_resize is not None else self.do_resize _a = resample if resample is not None else self.resample _a = do_center_crop if do_center_crop is not None else self.do_center_crop _a = do_rescale if do_rescale is not None else self.do_rescale _a = rescale_factor if rescale_factor is not None else self.rescale_factor _a = do_normalize if do_normalize is not None else self.do_normalize _a = image_mean if image_mean is not None else self.image_mean _a = image_std if image_std is not None else self.image_std _a = size if size is not None else self.size _a = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) _a = crop_size if crop_size is not None else self.crop_size _a = get_size_dict(__UpperCAmelCase , param_name="crop_size" ) _a = make_list_of_images(__UpperCAmelCase ) if not valid_images(__UpperCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _a = [to_numpy_array(__UpperCAmelCase ) for image in images] if do_resize: _a = [self.resize(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) for image in images] if do_center_crop: _a = [self.center_crop(__UpperCAmelCase , __UpperCAmelCase ) for image in images] if do_rescale: _a = [self.rescale(__UpperCAmelCase , __UpperCAmelCase ) for image in images] if do_normalize: _a = [self.normalize(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) for image in images] _a = [to_channel_dimension_format(__UpperCAmelCase , __UpperCAmelCase ) for image in images] _a = {"pixel_values": images} return BatchFeature(data=__UpperCAmelCase , tensor_type=__UpperCAmelCase )
354
'''simple docstring''' import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient lowerCAmelCase_ : Tuple = WebClient(token=os.environ['CI_SLACK_BOT_TOKEN']) def _lowerCamelCase ( lowercase : List[Any] ) -> Optional[int]: _a = test_results.split(" " ) _a = 0 _a = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. _a = expressions[-2] if "=" in expressions[-1] else expressions[-1] for i, expression in enumerate(lowercase ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def _lowerCamelCase ( lowercase : str ) -> Optional[Any]: _a = {} _a = None _a = False for line in failures_short_lines.split("\n" ): if re.search(r"_ \[doctest\]" , lowercase ): _a = True _a = line.split(" " )[2] elif in_error and not line.split(" " )[0].isdigit(): _a = line _a = False return failures class __SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Tuple , __a : str , __a : Dict ): _a = title _a = doc_test_results["time_spent"].split("," )[0] _a = doc_test_results["success"] _a = doc_test_results["failures"] _a = self.n_success + self.n_failures # Failures and success of the modeling tests _a = doc_test_results @property def UpperCamelCase__ ( self : int ): _a = [self._time_spent] _a = 0 for time in time_spent: _a = time.split(":" ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(__a ) == 1: _a = [0, 0, time_parts[0]] _a , _a , _a = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 36_00 + minutes * 60 + seconds _a , _a , _a = total_secs // 36_00, (total_secs % 36_00) // 60, total_secs % 60 return f'{int(__a )}h{int(__a )}m{int(__a )}s' @property def UpperCamelCase__ ( self : Optional[Any] ): return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def UpperCamelCase__ ( self : Optional[Any] ): return { "type": "section", "text": { "type": "plain_text", "text": f'🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.', "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f'https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}', }, } @property def UpperCamelCase__ ( self : List[str] ): return { "type": "section", "text": { "type": "plain_text", "text": ( f'There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in' f' {self.time}.' ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f'https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}', }, } @property def UpperCamelCase__ ( self : str ): _a = 40 _a = {k: v["failed"] for k, v in doc_test_results.items() if isinstance(__a , __a )} _a = "" for category, failures in category_failures.items(): if len(__a ) == 0: continue if report != "": report += "\n\n" report += f'*{category} failures*:'.ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(__a ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": f'The following examples had failures:\n\n\n{report}\n', }, } @property def UpperCamelCase__ ( self : List[str] ): _a = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(__a ) @staticmethod def UpperCamelCase__ ( ): _a = [ { "type": "section", "text": { "type": "plain_text", "text": "There was an issue running the tests.", }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f'https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}', }, } ] print("Sending the following payload" ) print(json.dumps({"blocks": json.loads(__a )} ) ) client.chat_postMessage( channel=os.environ["CI_SLACK_CHANNEL_ID_DAILY"] , text="There was an issue running the tests." , blocks=__a , ) def UpperCamelCase__ ( self : Tuple ): print("Sending the following payload" ) print(json.dumps({"blocks": json.loads(self.payload )} ) ) _a = f'{self.n_failures} failures out of {self.n_tests} tests,' if self.n_failures else "All tests passed." _a = client.chat_postMessage( channel=os.environ["CI_SLACK_CHANNEL_ID_DAILY"] , blocks=self.payload , text=__a , ) def UpperCamelCase__ ( self : Dict , __a : List[str] , __a : List[Any] , __a : Tuple , __a : int ): _a = "" for key, value in failures.items(): _a = value[:2_00] + " [Truncated]" if len(__a ) > 2_50 else value failures_text += f'*{key}*\n_{value}_\n\n' _a = job_name _a = {"type": "section", "text": {"type": "mrkdwn", "text": text}} if job_link is not None: _a = { "type": "button", "text": {"type": "plain_text", "text": "GitHub Action job", "emoji": True}, "url": job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def UpperCamelCase__ ( self : str ): if self.thread_ts is None: raise ValueError("Can only post reply if a post has been made." ) _a = self.doc_test_results.pop("job_link" ) self.doc_test_results.pop("failures" ) self.doc_test_results.pop("success" ) self.doc_test_results.pop("time_spent" ) _a = sorted(self.doc_test_results.items() , key=lambda __a : t[0] ) for job, job_result in sorted_dict: if len(job_result["failures"] ): _a = f'*Num failures* :{len(job_result["failed"] )} \n' _a = job_result["failures"] _a = self.get_reply_blocks(__a , __a , __a , text=__a ) print("Sending the following reply" ) print(json.dumps({"blocks": blocks} ) ) client.chat_postMessage( channel=os.environ["CI_SLACK_CHANNEL_ID_DAILY"] , text=f'Results for {job}' , blocks=__a , thread_ts=self.thread_ts["ts"] , ) time.sleep(1 ) def _lowerCamelCase ( ) -> Any: _a = os.environ["GITHUB_RUN_ID"] _a = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100' _a = requests.get(lowercase ).json() _a = {} try: jobs.update({job["name"]: job["html_url"] for job in result["jobs"]} ) _a = math.ceil((result["total_count"] - 100) / 100 ) for i in range(lowercase ): _a = requests.get(url + F'&page={i + 2}' ).json() jobs.update({job["name"]: job["html_url"] for job in result["jobs"]} ) return jobs except Exception as e: print("Unknown error, could not fetch links." , lowercase ) return {} def _lowerCamelCase ( lowercase : str ) -> Dict: _a = {} if os.path.exists(lowercase ): _a = os.listdir(lowercase ) for file in files: try: with open(os.path.join(lowercase , lowercase ) , encoding="utf-8" ) as f: _a = f.read() except UnicodeDecodeError as e: raise ValueError(F'Could not open {os.path.join(lowercase , lowercase )}.' ) from e return _artifact def _lowerCamelCase ( ) -> str: class __SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Dict , __a : str ): _a = name _a = [] def __str__( self : List[str] ): return self.name def UpperCamelCase__ ( self : str , __a : str ): self.paths.append({"name": self.name, "path": path} ) _a = {} _a = filter(os.path.isdir , os.listdir() ) for directory in directories: _a = directory if artifact_name not in _available_artifacts: _a = Artifact(lowercase ) _available_artifacts[artifact_name].add_path(lowercase ) return _available_artifacts if __name__ == "__main__": lowerCAmelCase_ : List[Any] = get_job_links() lowerCAmelCase_ : Any = retrieve_available_artifacts() lowerCAmelCase_ : List[str] = collections.OrderedDict( [ ('*.py', 'API Examples'), ('*.md', 'MD Examples'), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' lowerCAmelCase_ : Optional[Any] = { v: { 'failed': [], 'failures': {}, } for v in docs.values() } # Link to the GitHub Action job lowerCAmelCase_ : int = github_actions_job_links.get('run_doctests') lowerCAmelCase_ : Union[str, Any] = available_artifacts['doc_tests_gpu_test_reports'].paths[0] lowerCAmelCase_ : List[str] = retrieve_artifact(artifact_path['name']) if "stats" in artifact: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = handle_test_results(artifact['stats']) lowerCAmelCase_ : List[str] = failed lowerCAmelCase_ : Optional[Any] = success lowerCAmelCase_ : Tuple = time_spent[1:-1] + ', ' lowerCAmelCase_ : List[Any] = extract_first_line_failure(artifact['failures_short']) for line in artifact["summary_short"].split('\n'): if re.search('FAILED', line): lowerCAmelCase_ : int = line.replace('FAILED ', '') lowerCAmelCase_ : Optional[int] = line.split()[0].replace('\n', '') if "::" in line: lowerCAmelCase_ , lowerCAmelCase_ : str = line.split('::') else: lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): lowerCAmelCase_ : Union[str, Any] = docs[file_regex] doc_test_results[category]["failed"].append(test) lowerCAmelCase_ : List[str] = all_failures[test] if test in all_failures else 'N/A' lowerCAmelCase_ : Optional[Any] = failure break lowerCAmelCase_ : Tuple = Message('🤗 Results of the doc tests.', doc_test_results) message.post() message.post_reply()
346
0
'''simple docstring''' def lowercase__ ( __lowercase : int = 50000000 ) -> int: """simple docstring""" __UpperCamelCase = set() __UpperCamelCase = int((limit - 24) ** (1 / 2) ) __UpperCamelCase = set(range(3 , prime_square_limit + 1 , 2 ) ) primes.add(2 ) for p in range(3 , prime_square_limit + 1 , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , prime_square_limit + 1 , __lowercase ) ) ) for primea in primes: __UpperCamelCase = primea * primea for primea in primes: __UpperCamelCase = primea * primea * primea if square + cube >= limit - 16: break for primea in primes: __UpperCamelCase = primea * primea * primea * primea __UpperCamelCase = square + cube + tetr if total >= limit: break ret.add(__lowercase ) return len(__lowercase ) if __name__ == "__main__": print(f'{solution() = }')
53
'''simple docstring''' import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowercase__ ( __lowercase : Features ) -> Optional[int]: """simple docstring""" __UpperCamelCase = np.inf def set_batch_size(__lowercase : FeatureType ) -> None: nonlocal batch_size if isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__lowercase , __lowercase ) and feature.dtype == "binary": __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__lowercase , __lowercase ) return None if batch_size is np.inf else batch_size class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : List[str] , __A : NestedDataStructureLike[PathLike] , __A : Optional[NamedSplit] = None , __A : Optional[Features] = None , __A : str = None , __A : bool = False , __A : bool = False , __A : Optional[int] = None , **__A : Dict , ): super().__init__( __A , split=__A , features=__A , cache_dir=__A , keep_in_memory=__A , streaming=__A , num_proc=__A , **__A , ) __UpperCamelCase = path_or_paths if isinstance(__A , __A ) else {self.split: path_or_paths} __UpperCamelCase = _PACKAGED_DATASETS_MODULES['parquet'][1] __UpperCamelCase = Parquet( cache_dir=__A , data_files=__A , features=__A , hash=__A , **__A , ) def _lowerCamelCase ( self : Optional[int] ): # Build iterable dataset if self.streaming: __UpperCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None self.builder.download_and_prepare( download_config=__A , download_mode=__A , verification_mode=__A , base_path=__A , num_proc=self.num_proc , ) __UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__A , in_memory=self.keep_in_memory ) return dataset class snake_case : """simple docstring""" def __init__( self : List[str] , __A : Dataset , __A : Union[PathLike, BinaryIO] , __A : Optional[int] = None , **__A : Dict , ): __UpperCamelCase = dataset __UpperCamelCase = path_or_buf __UpperCamelCase = batch_size or get_writer_batch_size(dataset.features ) __UpperCamelCase = parquet_writer_kwargs def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with open(self.path_or_buf , 'wb+' ) as buffer: __UpperCamelCase = self._write(file_obj=__A , batch_size=__A , **self.parquet_writer_kwargs ) else: __UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__A , **self.parquet_writer_kwargs ) return written def _lowerCamelCase ( self : List[str] , __A : BinaryIO , __A : int , **__A : List[str] ): __UpperCamelCase = 0 __UpperCamelCase = parquet_writer_kwargs.pop('path_or_buf' , __A ) __UpperCamelCase = self.dataset.features.arrow_schema __UpperCamelCase = pq.ParquetWriter(__A , schema=__A , **__A ) for offset in logging.tqdm( range(0 , len(self.dataset ) , __A ) , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating parquet from Arrow format' , ): __UpperCamelCase = query_table( table=self.dataset._data , key=slice(__A , offset + batch_size ) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__A ) written += batch.nbytes writer.close() return written
53
1
def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ ): if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
194
import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): # Load configuration defined in the metadata file with open(SCREAMING_SNAKE_CASE__ ) as metadata_file: __lowerCamelCase : List[str] = json.load(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : int = LukeConfig(use_entity_aware_attention=SCREAMING_SNAKE_CASE__ , **metadata['model_config'] ) # Load in the weights from the checkpoint_path __lowerCamelCase : Union[str, Any] = torch.load(SCREAMING_SNAKE_CASE__ , map_location='cpu' ) # Load the entity vocab file __lowerCamelCase : List[Any] = load_entity_vocab(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Optional[int] = RobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks __lowerCamelCase : str = AddedToken('<ent>' , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : Dict = AddedToken('<ent2>' , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(f'Saving tokenizer to {pytorch_dump_folder_path}' ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE__ ) with open(os.path.join(SCREAMING_SNAKE_CASE__ , LukeTokenizer.vocab_files_names['entity_vocab_file'] ) , 'w' ) as f: json.dump(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase : str = LukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) # Initialize the embeddings of the special tokens __lowerCamelCase : Union[str, Any] = state_dict['embeddings.word_embeddings.weight'] __lowerCamelCase : Tuple = word_emb[tokenizer.convert_tokens_to_ids(['@'] )[0]].unsqueeze(0 ) __lowerCamelCase : Any = word_emb[tokenizer.convert_tokens_to_ids(['#'] )[0]].unsqueeze(0 ) __lowerCamelCase : List[str] = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: __lowerCamelCase : Optional[int] = f'encoder.layer.{layer_index}.attention.self.' __lowerCamelCase : Dict = state_dict[prefix + matrix_name] __lowerCamelCase : List[Any] = state_dict[prefix + matrix_name] __lowerCamelCase : Union[str, Any] = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks __lowerCamelCase : Optional[int] = state_dict['entity_embeddings.entity_embeddings.weight'] __lowerCamelCase : Union[str, Any] = entity_emb[entity_vocab['[MASK]']] __lowerCamelCase : Optional[Any] = LukeModel(config=SCREAMING_SNAKE_CASE__ ).eval() __lowerCamelCase , __lowerCamelCase : List[Any] = model.load_state_dict(SCREAMING_SNAKE_CASE__ , strict=SCREAMING_SNAKE_CASE__ ) if not (len(SCREAMING_SNAKE_CASE__ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(f'Missing keys {", ".join(SCREAMING_SNAKE_CASE__ )}. Expected only missing embeddings.position_ids' ) if not (all(key.startswith('entity_predictions' ) or key.startswith('lm_head' ) for key in unexpected_keys )): raise ValueError( 'Unexpected keys' f' {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}' ) # Check outputs __lowerCamelCase : Optional[Any] = LukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ , task='entity_classification' ) __lowerCamelCase : Dict = ( 'Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the' ' new world number one avoid a humiliating second- round exit at Wimbledon .' ) __lowerCamelCase : Union[str, Any] = (39, 42) __lowerCamelCase : Optional[Any] = tokenizer(SCREAMING_SNAKE_CASE__ , entity_spans=[span] , add_prefix_space=SCREAMING_SNAKE_CASE__ , return_tensors='pt' ) __lowerCamelCase : List[str] = model(**SCREAMING_SNAKE_CASE__ ) # Verify word hidden states if model_size == "large": __lowerCamelCase : Dict = torch.Size((1, 42, 1_024) ) __lowerCamelCase : int = torch.tensor( [[0.0_133, 0.0_865, 0.0_095], [0.3_093, -0.2_576, -0.7_418], [-0.1_720, -0.2_117, -0.2_869]] ) else: # base __lowerCamelCase : Union[str, Any] = torch.Size((1, 42, 768) ) __lowerCamelCase : Tuple = torch.tensor([[0.0_037, 0.1_368, -0.0_091], [0.1_099, 0.3_329, -0.1_095], [0.0_765, 0.5_335, 0.1_179]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( f'Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": __lowerCamelCase : Union[str, Any] = torch.Size((1, 1, 1_024) ) __lowerCamelCase : Dict = torch.tensor([[0.0_466, -0.0_106, -0.0_179]] ) else: # base __lowerCamelCase : int = torch.Size((1, 1, 768) ) __lowerCamelCase : Dict = torch.tensor([[0.1_457, 0.1_044, 0.0_174]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( f'Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is' f' {expected_shape}' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(SCREAMING_SNAKE_CASE__ ) ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : Tuple = {} with open(SCREAMING_SNAKE_CASE__ , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(SCREAMING_SNAKE_CASE__ ): __lowerCamelCase , __lowerCamelCase : List[Any] = line.rstrip().split('\t' ) __lowerCamelCase : Any = index return entity_vocab if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument('--checkpoint_path', type=str, help='Path to a pytorch_model.bin file.') parser.add_argument( '--metadata_path', default=None, type=str, help='Path to a metadata.json file, defining the configuration.' ) parser.add_argument( '--entity_vocab_path', default=None, type=str, help='Path to an entity_vocab.tsv file, containing the entity vocabulary.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to where to dump the output PyTorch model.' ) parser.add_argument( '--model_size', default='base', type=str, choices=['base', 'large'], help='Size of the model to be converted.' ) lowercase_ = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
194
1
def A_ ( snake_case : str ) -> bool: '''simple docstring''' __UpperCamelCase = [int(snake_case ) for i in ip_va_address.split('''.''' ) if i.isdigit()] return len(snake_case ) == 4 and all(0 <= int(snake_case ) <= 254 for octet in octets ) if __name__ == "__main__": lowercase__ : int = input().strip() lowercase__ : Tuple = "valid" if is_ip_va_address_valid(ip) else "invalid" print(F"{ip} is a {valid_or_invalid} IP v4 address.")
328
import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal lowercase__ : Optional[int] = datasets.utils.logging.get_logger(__name__) lowercase__ : Optional[Any] = ["names", "prefix"] lowercase__ : List[Any] = ["warn_bad_lines", "error_bad_lines", "mangle_dupe_cols"] lowercase__ : Optional[Any] = ["encoding_errors", "on_bad_lines"] lowercase__ : List[str] = ["date_format"] @dataclass class SCREAMING_SNAKE_CASE__ ( datasets.BuilderConfig ): """simple docstring""" _snake_case = "," _snake_case = None _snake_case = "infer" _snake_case = None _snake_case = None _snake_case = None _snake_case = None _snake_case = None _snake_case = True _snake_case = None _snake_case = None _snake_case = None _snake_case = None _snake_case = False _snake_case = None _snake_case = None _snake_case = None _snake_case = True _snake_case = True _snake_case = False _snake_case = True _snake_case = None _snake_case = "." _snake_case = None _snake_case = '"' _snake_case = 0 _snake_case = None _snake_case = None _snake_case = None _snake_case = None _snake_case = True _snake_case = True _snake_case = 0 _snake_case = True _snake_case = False _snake_case = None _snake_case = 10000 _snake_case = None _snake_case = "strict" _snake_case = "error" _snake_case = None def A__ ( self )-> Any: '''simple docstring''' if self.delimiter is not None: __UpperCamelCase = self.delimiter if self.column_names is not None: __UpperCamelCase = self.column_names @property def A__ ( self )-> Any: '''simple docstring''' __UpperCamelCase = { '''sep''': self.sep, '''header''': self.header, '''names''': self.names, '''index_col''': self.index_col, '''usecols''': self.usecols, '''prefix''': self.prefix, '''mangle_dupe_cols''': self.mangle_dupe_cols, '''engine''': self.engine, '''converters''': self.converters, '''true_values''': self.true_values, '''false_values''': self.false_values, '''skipinitialspace''': self.skipinitialspace, '''skiprows''': self.skiprows, '''nrows''': self.nrows, '''na_values''': self.na_values, '''keep_default_na''': self.keep_default_na, '''na_filter''': self.na_filter, '''verbose''': self.verbose, '''skip_blank_lines''': self.skip_blank_lines, '''thousands''': self.thousands, '''decimal''': self.decimal, '''lineterminator''': self.lineterminator, '''quotechar''': self.quotechar, '''quoting''': self.quoting, '''escapechar''': self.escapechar, '''comment''': self.comment, '''encoding''': self.encoding, '''dialect''': self.dialect, '''error_bad_lines''': self.error_bad_lines, '''warn_bad_lines''': self.warn_bad_lines, '''skipfooter''': self.skipfooter, '''doublequote''': self.doublequote, '''memory_map''': self.memory_map, '''float_precision''': self.float_precision, '''chunksize''': self.chunksize, '''encoding_errors''': self.encoding_errors, '''on_bad_lines''': self.on_bad_lines, '''date_format''': self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , SCREAMING_SNAKE_CASE_ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class SCREAMING_SNAKE_CASE__ ( datasets.ArrowBasedBuilder ): """simple docstring""" _snake_case = CsvConfig def A__ ( self )-> Any: '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def A__ ( self , SCREAMING_SNAKE_CASE_ )-> Optional[int]: '''simple docstring''' if not self.config.data_files: raise ValueError(F"At least one data file must be specified, but got data_files={self.config.data_files}" ) __UpperCamelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(SCREAMING_SNAKE_CASE_ , (str, list, tuple) ): __UpperCamelCase = data_files if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): __UpperCamelCase = [files] __UpperCamelCase = [dl_manager.iter_files(SCREAMING_SNAKE_CASE_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] __UpperCamelCase = [] for split_name, files in data_files.items(): if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): __UpperCamelCase = [files] __UpperCamelCase = [dl_manager.iter_files(SCREAMING_SNAKE_CASE_ ) for file in files] splits.append(datasets.SplitGenerator(name=SCREAMING_SNAKE_CASE_ , gen_kwargs={'''files''': files} ) ) return splits def A__ ( self , SCREAMING_SNAKE_CASE_ )-> pa.Table: '''simple docstring''' if self.config.features is not None: __UpperCamelCase = self.config.features.arrow_schema if all(not require_storage_cast(SCREAMING_SNAKE_CASE_ ) for feature in self.config.features.values() ): # cheaper cast __UpperCamelCase = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=SCREAMING_SNAKE_CASE_ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example __UpperCamelCase = table_cast(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return pa_table def A__ ( self , SCREAMING_SNAKE_CASE_ )-> str: '''simple docstring''' __UpperCamelCase = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str __UpperCamelCase = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(SCREAMING_SNAKE_CASE_ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(SCREAMING_SNAKE_CASE_ ) ): __UpperCamelCase = pd.read_csv(SCREAMING_SNAKE_CASE_ , iterator=SCREAMING_SNAKE_CASE_ , dtype=SCREAMING_SNAKE_CASE_ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(SCREAMING_SNAKE_CASE_ ): __UpperCamelCase = pa.Table.from_pandas(SCREAMING_SNAKE_CASE_ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(SCREAMING_SNAKE_CASE_ ) except ValueError as e: logger.error(F"Failed to read file '{file}' with error {type(SCREAMING_SNAKE_CASE_ )}: {e}" ) raise
328
1
'''simple docstring''' import argparse import os import re import packaging.version _UpperCAmelCase : Optional[int] = """examples/""" _UpperCAmelCase : Dict = { """examples""": (re.compile(r"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(r"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(r"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), r"""\1version=\"VERSION\","""), """doc""": (re.compile(r"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } _UpperCAmelCase : str = { """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } _UpperCAmelCase : Optional[Any] = """README.md""" def __magic_name__( lowerCamelCase, lowerCamelCase, lowerCamelCase): with open(lowerCAmelCase__, '''r''', encoding='''utf-8''', newline='''\n''') as f: __lowerCAmelCase = f.read() __lowerCAmelCase = REPLACE_PATTERNS[pattern] __lowerCAmelCase = replace.replace('''VERSION''', lowerCAmelCase__) __lowerCAmelCase = re_pattern.sub(lowerCAmelCase__, lowerCAmelCase__) with open(lowerCAmelCase__, '''w''', encoding='''utf-8''', newline='''\n''') as f: f.write(lowerCAmelCase__) def __magic_name__( lowerCamelCase): for folder, directories, fnames in os.walk(lowerCAmelCase__): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''') if "legacy" in directories: directories.remove('''legacy''') for fname in fnames: if fname.endswith('''.py'''): update_version_in_file(os.path.join(lowerCAmelCase__, lowerCAmelCase__), lowerCAmelCase__, pattern='''examples''') def __magic_name__( lowerCamelCase, lowerCamelCase=False): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__) if not patch: update_version_in_examples(lowerCAmelCase__) def __magic_name__( ): __lowerCAmelCase = '''🤗 Transformers currently provides the following architectures''' __lowerCAmelCase = '''1. Want to contribute a new model?''' with open(lowerCAmelCase__, '''r''', encoding='''utf-8''', newline='''\n''') as f: __lowerCAmelCase = f.readlines() # Find the start of the list. __lowerCAmelCase = 0 while not lines[start_index].startswith(_start_prompt): start_index += 1 start_index += 1 __lowerCAmelCase = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt): if lines[index].startswith('''1.'''): __lowerCAmelCase = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''', '''https://huggingface.co/docs/transformers/model_doc''', ) index += 1 with open(lowerCAmelCase__, '''w''', encoding='''utf-8''', newline='''\n''') as f: f.writelines(lowerCAmelCase__) def __magic_name__( ): with open(REPLACE_FILES['''init'''], '''r''') as f: __lowerCAmelCase = f.read() __lowerCAmelCase = REPLACE_PATTERNS['''init'''][0].search(lowerCAmelCase__).groups()[0] return packaging.version.parse(lowerCAmelCase__) def __magic_name__( lowerCamelCase=False): __lowerCAmelCase = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''') if default_version.is_devrelease: __lowerCAmelCase = default_version.base_version elif patch: __lowerCAmelCase = F"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: __lowerCAmelCase = F"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. __lowerCAmelCase = input(F"""Which version are you releasing? [{default_version}]""") if len(lowerCAmelCase__) == 0: __lowerCAmelCase = default_version print(F"""Updating version to {version}.""") global_version_update(lowerCAmelCase__, patch=lowerCAmelCase__) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''') clean_main_ref_in_model_list() def __magic_name__( ): __lowerCAmelCase = get_version() __lowerCAmelCase = F"""{current_version.major}.{current_version.minor + 1}.0.dev0""" __lowerCAmelCase = current_version.base_version # Check with the user we got that right. __lowerCAmelCase = input(F"""Which version are we developing now? [{dev_version}]""") if len(lowerCAmelCase__) == 0: __lowerCAmelCase = dev_version print(F"""Updating version to {version}.""") global_version_update(lowerCAmelCase__) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''') clean_main_ref_in_model_list() if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") _UpperCAmelCase : str = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
351
'''simple docstring''' from __future__ import annotations import math def __magic_name__( lowerCamelCase, lowerCamelCase): if len(lowerCamelCase) != 2 or len(a[0]) != 2 or len(lowerCamelCase) != 2 or len(b[0]) != 2: raise Exception('''Matrices are not 2x2''') __lowerCAmelCase = [ [a[0][0] * b[0][0] + a[0][1] * b[1][0], a[0][0] * b[0][1] + a[0][1] * b[1][1]], [a[1][0] * b[0][0] + a[1][1] * b[1][0], a[1][0] * b[0][1] + a[1][1] * b[1][1]], ] return new_matrix def __magic_name__( lowerCamelCase, lowerCamelCase): return [ [matrix_a[row][col] + matrix_b[row][col] for col in range(len(matrix_a[row]))] for row in range(len(lowerCamelCase)) ] def __magic_name__( lowerCamelCase, lowerCamelCase): return [ [matrix_a[row][col] - matrix_b[row][col] for col in range(len(matrix_a[row]))] for row in range(len(lowerCamelCase)) ] def __magic_name__( lowerCamelCase): if len(lowerCamelCase) % 2 != 0 or len(a[0]) % 2 != 0: raise Exception('''Odd matrices are not supported!''') __lowerCAmelCase = len(lowerCamelCase) __lowerCAmelCase = matrix_length // 2 __lowerCAmelCase = [[a[i][j] for j in range(lowerCamelCase, lowerCamelCase)] for i in range(lowerCamelCase)] __lowerCAmelCase = [ [a[i][j] for j in range(lowerCamelCase, lowerCamelCase)] for i in range(lowerCamelCase, lowerCamelCase) ] __lowerCAmelCase = [[a[i][j] for j in range(lowerCamelCase)] for i in range(lowerCamelCase)] __lowerCAmelCase = [[a[i][j] for j in range(lowerCamelCase)] for i in range(lowerCamelCase, lowerCamelCase)] return top_left, top_right, bot_left, bot_right def __magic_name__( lowerCamelCase): return len(lowerCamelCase), len(matrix[0]) def __magic_name__( lowerCamelCase): print('''\n'''.join(str(lowerCamelCase) for line in matrix)) def __magic_name__( lowerCamelCase, lowerCamelCase): if matrix_dimensions(lowerCamelCase) == (2, 2): return default_matrix_multiplication(lowerCamelCase, lowerCamelCase) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = split_matrix(lowerCamelCase) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = split_matrix(lowerCamelCase) __lowerCAmelCase = actual_strassen(lowerCamelCase, matrix_subtraction(lowerCamelCase, lowerCamelCase)) __lowerCAmelCase = actual_strassen(matrix_addition(lowerCamelCase, lowerCamelCase), lowerCamelCase) __lowerCAmelCase = actual_strassen(matrix_addition(lowerCamelCase, lowerCamelCase), lowerCamelCase) __lowerCAmelCase = actual_strassen(lowerCamelCase, matrix_subtraction(lowerCamelCase, lowerCamelCase)) __lowerCAmelCase = actual_strassen(matrix_addition(lowerCamelCase, lowerCamelCase), matrix_addition(lowerCamelCase, lowerCamelCase)) __lowerCAmelCase = actual_strassen(matrix_subtraction(lowerCamelCase, lowerCamelCase), matrix_addition(lowerCamelCase, lowerCamelCase)) __lowerCAmelCase = actual_strassen(matrix_subtraction(lowerCamelCase, lowerCamelCase), matrix_addition(lowerCamelCase, lowerCamelCase)) __lowerCAmelCase = matrix_addition(matrix_subtraction(matrix_addition(lowerCamelCase, lowerCamelCase), lowerCamelCase), lowerCamelCase) __lowerCAmelCase = matrix_addition(lowerCamelCase, lowerCamelCase) __lowerCAmelCase = matrix_addition(lowerCamelCase, lowerCamelCase) __lowerCAmelCase = matrix_subtraction(matrix_subtraction(matrix_addition(lowerCamelCase, lowerCamelCase), lowerCamelCase), lowerCamelCase) # construct the new matrix from our 4 quadrants __lowerCAmelCase = [] for i in range(len(lowerCamelCase)): new_matrix.append(top_left[i] + top_right[i]) for i in range(len(lowerCamelCase)): new_matrix.append(bot_left[i] + bot_right[i]) return new_matrix def __magic_name__( lowerCamelCase, lowerCamelCase): if matrix_dimensions(lowerCamelCase)[1] != matrix_dimensions(lowerCamelCase)[0]: __lowerCAmelCase = ( '''Unable to multiply these matrices, please check the dimensions.\n''' F"""Matrix A: {matrixa}\n""" F"""Matrix B: {matrixa}""" ) raise Exception(lowerCamelCase) __lowerCAmelCase = matrix_dimensions(lowerCamelCase) __lowerCAmelCase = matrix_dimensions(lowerCamelCase) if dimensiona[0] == dimensiona[1] and dimensiona[0] == dimensiona[1]: return [matrixa, matrixa] __lowerCAmelCase = max(*lowerCamelCase, *lowerCamelCase) __lowerCAmelCase = int(math.pow(2, math.ceil(math.loga(lowerCamelCase)))) __lowerCAmelCase = matrixa __lowerCAmelCase = matrixa # Adding zeros to the matrices so that the arrays dimensions are the same and also # power of 2 for i in range(0, lowerCamelCase): if i < dimensiona[0]: for _ in range(dimensiona[1], lowerCamelCase): new_matrixa[i].append(0) else: new_matrixa.append([0] * maxim) if i < dimensiona[0]: for _ in range(dimensiona[1], lowerCamelCase): new_matrixa[i].append(0) else: new_matrixa.append([0] * maxim) __lowerCAmelCase = actual_strassen(lowerCamelCase, lowerCamelCase) # Removing the additional zeros for i in range(0, lowerCamelCase): if i < dimensiona[0]: for _ in range(dimensiona[1], lowerCamelCase): final_matrix[i].pop() else: final_matrix.pop() return final_matrix if __name__ == "__main__": _UpperCAmelCase : List[str] = [ [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 2, 3, 1], ] _UpperCAmelCase : Optional[Any] = [[0, 2, 1, 1], [1_6, 2, 3, 3], [2, 2, 7, 7], [1_3, 1_1, 2_2, 4]] print(strassen(matrixa, matrixa))
9
0
"""simple docstring""" from typing import Dict from .base import GenericTensor, Pipeline class lowerCAmelCase_ (a__ ): """simple docstring""" def __magic_name__ (self , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , **SCREAMING_SNAKE_CASE__ ) -> Any: """simple docstring""" if tokenize_kwargs is None: SCREAMING_SNAKE_CASE__ : Optional[Any] = {} if truncation is not None: if "truncation" in tokenize_kwargs: raise ValueError( """truncation parameter defined twice (given as keyword argument as well as in tokenize_kwargs)""" ) SCREAMING_SNAKE_CASE__ : List[str] = truncation SCREAMING_SNAKE_CASE__ : List[str] = tokenize_kwargs SCREAMING_SNAKE_CASE__ : Any = {} if return_tensors is not None: SCREAMING_SNAKE_CASE__ : int = return_tensors return preprocess_params, {}, postprocess_params def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) -> Dict[str, GenericTensor]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = self.framework SCREAMING_SNAKE_CASE__ : Tuple = self.tokenizer(SCREAMING_SNAKE_CASE__ , return_tensors=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) return model_inputs def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = self.model(**SCREAMING_SNAKE_CASE__ ) return model_outputs def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ) -> Optional[Any]: """simple docstring""" if return_tensors: return model_outputs[0] if self.framework == "pt": return model_outputs[0].tolist() elif self.framework == "tf": return model_outputs[0].numpy().tolist() def __call__(self , *SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) -> Any: """simple docstring""" return super().__call__(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ )
25
'''simple docstring''' a_ : Any = """0.21.0""" from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
55
0
'''simple docstring''' from dataclasses import dataclass, field from typing import Optional @dataclass class a__ : '''simple docstring''' lowercase__ : Optional[str] = field( default="codeparrot/codeparrot" , metadata={"help": "Model name or path of model to be trained."} ) lowercase__ : Optional[str] = field( default="./" , metadata={"help": "Save dir where model repo is cloned and models updates are saved to."} ) lowercase__ : Optional[str] = field( default="codeparrot/codeparrot-clean-train" , metadata={"help": "Name or path of training dataset."} ) lowercase__ : Optional[str] = field( default="codeparrot/codeparrot-clean-valid" , metadata={"help": "Name or path of validation dataset."} ) lowercase__ : Optional[int] = field(default=2 , metadata={"help": "Batch size for training."} ) lowercase__ : Optional[int] = field(default=2 , metadata={"help": "Batch size for evaluation."} ) lowercase__ : Optional[float] = field(default=0.1 , metadata={"help": "Value of weight decay."} ) lowercase__ : Optional[int] = field( default=1_0_0_0_0 , metadata={"help": "Size of buffer used to shuffle streaming dataset."} ) lowercase__ : Optional[float] = field(default=2e-4 , metadata={"help": "Learning rate fo training."} ) lowercase__ : Optional[str] = field(default="cosine" , metadata={"help": "Learning rate."} ) lowercase__ : Optional[int] = field( default=7_5_0 , metadata={"help": "Number of warmup steps in the learning rate schedule."} ) lowercase__ : Optional[int] = field( default=1_6 , metadata={"help": "Number of gradient accumulation steps."} ) lowercase__ : Optional[bool] = field( default=a__ , metadata={"help": "Use gradient checkpointing to reduce memory footprint."} ) lowercase__ : Optional[int] = field(default=5_0_0_0_0 , metadata={"help": "Maximum number of training steps."} ) lowercase__ : Optional[int] = field( default=-1 , metadata={"help": "Maximum number of evaluation steps. If -1 the full dataset is evaluated."} ) lowercase__ : Optional[int] = field(default=1_0_2_4 , metadata={"help": "Sequence lengths used for training."} ) lowercase__ : Optional[int] = field(default=1 , metadata={"help": "Training seed."} ) lowercase__ : Optional[int] = field( default=1_0_2_4 , metadata={"help": "Interval to save checkpoints. Measured as number of forward passes not training steps."} , ) lowercase__ : Optional[str] = field( default=a__ , metadata={"help": "States path if the training should continue from a checkpoint folder."} ) lowercase__ : Optional[bool] = field(default=a__ , metadata={"help": "If True the data is pretokenized."} ) @dataclass class a__ : '''simple docstring''' lowercase__ : Optional[str] = field( default="codeparrot/codeparrot" , metadata={"help": "Model name or path of model to be evaluated."} ) lowercase__ : Optional[str] = field( default="codeparrot/codeparrot-clean-valid" , metadata={"help": "Name or path of validation dataset."} ) lowercase__ : Optional[int] = field(default=2 , metadata={"help": "Batch size used for evaluation."} ) lowercase__ : Optional[int] = field( default=-1 , metadata={"help": "Maximum number of evaluation steps. If -1 the full dataset is evaluated."} ) lowercase__ : Optional[int] = field(default=1_0_2_4 , metadata={"help": "Length of sequences to be evaluated."} ) lowercase__ : Optional[int] = field(default=1 , metadata={"help": "Random seed used for evaluation."} ) @dataclass class a__ : '''simple docstring''' lowercase__ : Optional[str] = field( default="codeparrot/codeparrot" , metadata={"help": "Model name or path of model to be evaluated."} ) lowercase__ : Optional[int] = field(default=a__ , metadata={"help": "Number of workers used for code evaluation."} ) lowercase__ : Optional[int] = field( default=a__ , metadata={"help": "The number of human-eval tasks to run. If not included all tasks are evaluated."} , ) lowercase__ : Optional[bool] = field( default=a__ , metadata={"help": "Sample from the language model's output distribution."} ) lowercase__ : Optional[float] = field(default=0.2 , metadata={"help": "Sampling temperature used for generation."} ) lowercase__ : Optional[int] = field(default=2_5_6 , metadata={"help": "Maximum number of newly generated tokens."} ) lowercase__ : Optional[int] = field(default=0 , metadata={"help": "Top-k parameter used for generation."} ) lowercase__ : Optional[float] = field(default=0.9_5 , metadata={"help": "Top-p parameter used for nucleus sampling."} ) lowercase__ : Optional[int] = field(default=1_0 , metadata={"help": "Number of generations to run in parallel."} ) lowercase__ : Optional[int] = field( default=2_0_0 , metadata={"help": "Number of completions to generate for each sample."} ) lowercase__ : Optional[int] = field(default=1 , metadata={"help": "Random seed used for evaluation."} ) lowercase__ : Optional[str] = field( default="eval_results.json" , metadata={"help": "Random seed used for evaluation."} ) lowercase__ : Optional[str] = field( default="0" , metadata={"help": "Allow `code_eval` to execute Python code on machine"} ) lowercase__ : Optional[int] = field( default=-1 , metadata={ "help": ( "Determine which device to run the `text-generation` Pipeline on. -1 is CPU and any zero or positive" " number corresponds to which GPU device id to run on." ) } , ) @dataclass class a__ : '''simple docstring''' lowercase__ : Optional[int] = field( default=a__ , metadata={ "help": "The number of CPU cores to use for parallel preprocessing. Default uses the maximum available." } , ) lowercase__ : Optional[str] = field( default="transformersbook/codeparrot" , metadata={"help": "Folder or name of dataset to process."} ) lowercase__ : Optional[str] = field( default="codeparrot-clean" , metadata={"help": "Folder to save processed processed dataset."} ) lowercase__ : Optional[int] = field( default=1_0_0_0_0_0 , metadata={"help": "Number of files to save per JSON output file."} ) lowercase__ : Optional[str] = field(default="content" , metadata={"help": "Column containing text data to process."} ) lowercase__ : Optional[float] = field( default=1_0_0_0 , metadata={"help": "Maximum line length in file, otherwise file is filtered."} ) lowercase__ : Optional[float] = field( default=1_0_0 , metadata={"help": "Maximum mean line length in file, otherwise file is filtered."} ) lowercase__ : Optional[float] = field( default=0.2_5 , metadata={"help": "Maximum fraction of non-alphanumeric characters, otherwise file is filtered."} ) lowercase__ : Optional[float] = field( default=1.5 , metadata={"help": "Minimum character token ratio for the file, otherwise file is filtered."} ) lowercase__ : Optional[float] = field( default=0.7 , metadata={"help": "Probability for filtering config, test and uncommon files."} ) lowercase__ : Optional[str] = field( default="codeparrot/codeparrot" , metadata={"help": "Name or path to the tokenizer."} , ) lowercase__ : Optional[bool] = field( default=a__ , metadata={"help": "If True, near-duplicate samples are removed."} ) lowercase__ : Optional[float] = field( default=0.8_5 , metadata={"help": "Jaccard threshold for near-duplicate samples."} ) @dataclass class a__ : '''simple docstring''' lowercase__ : Optional[str] = field( default="gpt2" , metadata={"help": "Base tokenizer to build new tokenizer from."} ) lowercase__ : Optional[str] = field( default="transformersbook/codeparrot-train" , metadata={"help": "Dataset to train tokenizer on."} ) lowercase__ : Optional[str] = field(default="content" , metadata={"help": "Column containing text data to process."} ) lowercase__ : Optional[int] = field(default=2_0_0_0_0_0 , metadata={"help": "Number of examples to train tokenizer on."} ) lowercase__ : Optional[int] = field( default=3_2_7_6_8 , metadata={"help": "Number of examples to train the tokenizer on."} ) lowercase__ : Optional[str] = field(default="codeparrot" , metadata={"help": "Name of new tokenizer."} ) lowercase__ : Optional[bool] = field(default=a__ , metadata={"help": "Push saved tokenizer to the hub."} ) @dataclass class a__ : '''simple docstring''' lowercase__ : Optional[str] = field( default="codeparrot/codeparrot" , metadata={"help": "Name or path to the tokenizer."} ) lowercase__ : Optional[str] = field( default="codeparrot/codeparrot-clean-train" , metadata={"help": "Name or path to the dataset to pretokenize."} ) lowercase__ : Optional[str] = field( default="tokenized-codeparrot-train" , metadata={"help": "Repo name of the pretokenized data."} ) lowercase__ : Optional[int] = field(default=a__ , metadata={"help": "Number of workers used for code evaluation."} ) @dataclass class a__ : '''simple docstring''' lowercase__ : Optional[str] = field( default="gpt2-large" , metadata={"help": "Configuration to use for model initialization."} ) lowercase__ : Optional[str] = field( default="codeparrot/codeparrot" , metadata={"help": "Tokenizer attached to model."} ) lowercase__ : Optional[str] = field(default="codeparrot" , metadata={"help": "Name of the created model."} ) lowercase__ : Optional[bool] = field(default=a__ , metadata={"help": "Push saved tokenizer to the hub."} )
367
'''simple docstring''' from __future__ import annotations import math from collections.abc import Callable def _snake_case ( A , A , A , A = 100 , ) -> float: lowerCAmelCase__ = x_start lowerCAmelCase__ = fnc(A ) lowerCAmelCase__ = 0.0 for _ in range(A ): # Approximates curve as a sequence of linear lines and sums their length lowerCAmelCase__ = (x_end - x_start) / steps + xa lowerCAmelCase__ = fnc(A ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowerCAmelCase__ = xa lowerCAmelCase__ = fxa return length if __name__ == "__main__": def _snake_case ( A ) -> List[Any]: return math.sin(10 * x ) print('''f(x) = sin(10 * x)''') print('''The length of the curve from x = -10 to x = 10 is:''') __UpperCAmelCase = 10 while i <= 100_000: print(f"""With {i} steps: {line_length(f, -10, 10, i)}""") i *= 10
228
0
import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller snake_case_ = 3 def snake_case__ ( SCREAMING_SNAKE_CASE_ : int ): '''simple docstring''' print('Generating primitive root of p' ) while True: lowercase__ : str = random.randrange(3 , SCREAMING_SNAKE_CASE_ ) if pow(SCREAMING_SNAKE_CASE_ , 2 , SCREAMING_SNAKE_CASE_ ) == 1: continue if pow(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) == 1: continue return g def snake_case__ ( SCREAMING_SNAKE_CASE_ : int ): '''simple docstring''' print('Generating prime p...' ) lowercase__ : List[str] = rabin_miller.generate_large_prime(SCREAMING_SNAKE_CASE_ ) # select large prime number. lowercase__ : Tuple = primitive_root(SCREAMING_SNAKE_CASE_ ) # one primitive root on modulo p. lowercase__ : Tuple = random.randrange(3 , SCREAMING_SNAKE_CASE_ ) # private_key -> have to be greater than 2 for safety. lowercase__ : Optional[Any] = cryptomath.find_mod_inverse(pow(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) lowercase__ : List[Any] = (key_size, e_a, e_a, p) lowercase__ : List[Any] = (key_size, d) return public_key, private_key def snake_case__ ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : int ): '''simple docstring''' if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print('\nWARNING:' ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" 'Use a different name or delete these files and re-run this program.' ) sys.exit() lowercase__ , lowercase__ : int = generate_key(SCREAMING_SNAKE_CASE_ ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , 'w' ) as fo: fo.write(f"""{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , 'w' ) as fo: fo.write(f"""{private_key[0]},{private_key[1]}""" ) def snake_case__ ( ): '''simple docstring''' print('Making key files...' ) make_key_files('elgamal' , 2_048 ) print('Key files generation successful' ) if __name__ == "__main__": main()
214
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case_ = logging.get_logger(__name__) snake_case_ = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ (__snake_case ): __lowerCamelCase : Any = """distilbert""" __lowerCamelCase : List[Any] = { """hidden_size""": """dim""", """num_attention_heads""": """n_heads""", """num_hidden_layers""": """n_layers""", } def __init__( self , a=3_0522 , a=512 , a=False , a=6 , a=12 , a=768 , a=4 * 768 , a=0.1 , a=0.1 , a="gelu" , a=0.02 , a=0.1 , a=0.2 , a=0 , **a , ): lowercase__ : int = vocab_size lowercase__ : int = max_position_embeddings lowercase__ : List[str] = sinusoidal_pos_embds lowercase__ : Tuple = n_layers lowercase__ : Optional[int] = n_heads lowercase__ : List[Any] = dim lowercase__ : Any = hidden_dim lowercase__ : Union[str, Any] = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : List[Any] = activation lowercase__ : Union[str, Any] = initializer_range lowercase__ : Dict = qa_dropout lowercase__ : Tuple = seq_classif_dropout super().__init__(**a , pad_token_id=a) class SCREAMING_SNAKE_CASE__ (__snake_case ): @property def snake_case_ ( self): if self.task == "multiple-choice": lowercase__ : List[str] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowercase__ : Tuple = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
214
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a : List[str] = { 'configuration_maskformer': ['MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MaskFormerConfig'], 'configuration_maskformer_swin': ['MaskFormerSwinConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : int = ['MaskFormerFeatureExtractor'] a : int = ['MaskFormerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ 'MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'MaskFormerForInstanceSegmentation', 'MaskFormerModel', 'MaskFormerPreTrainedModel', ] a : str = [ 'MaskFormerSwinBackbone', 'MaskFormerSwinModel', 'MaskFormerSwinPreTrainedModel', ] if TYPE_CHECKING: from .configuration_maskformer import MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskFormerConfig from .configuration_maskformer_swin import MaskFormerSwinConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_maskformer import MaskFormerFeatureExtractor from .image_processing_maskformer import MaskFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskformer import ( MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskFormerForInstanceSegmentation, MaskFormerModel, MaskFormerPreTrainedModel, ) from .modeling_maskformer_swin import ( MaskFormerSwinBackbone, MaskFormerSwinModel, MaskFormerSwinPreTrainedModel, ) else: import sys a : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure)
72
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins a : int = ['tests.fixtures.files', 'tests.fixtures.hub', 'tests.fixtures.fsspec'] def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Tuple: '''simple docstring''' for item in items: if any(marker in item.keywords for marker in ['''integration''', '''unit'''] ): continue item.add_marker(pytest.mark.unit ) def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' config.addinivalue_line('''markers''', '''torchaudio_latest: mark test to run with torchaudio>=0.12''' ) @pytest.fixture(autouse=__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> List[Any]: '''simple docstring''' snake_case_ = tmp_path_factory.getbasetemp() / '''cache''' snake_case_ = test_hf_cache_home / '''datasets''' snake_case_ = test_hf_cache_home / '''metrics''' snake_case_ = test_hf_cache_home / '''modules''' monkeypatch.setattr('''datasets.config.HF_DATASETS_CACHE''', str(__UpperCAmelCase ) ) monkeypatch.setattr('''datasets.config.HF_METRICS_CACHE''', str(__UpperCAmelCase ) ) monkeypatch.setattr('''datasets.config.HF_MODULES_CACHE''', str(__UpperCAmelCase ) ) snake_case_ = test_hf_datasets_cache / '''downloads''' monkeypatch.setattr('''datasets.config.DOWNLOADED_DATASETS_PATH''', str(__UpperCAmelCase ) ) snake_case_ = test_hf_datasets_cache / '''downloads''' / '''extracted''' monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''', str(__UpperCAmelCase ) ) @pytest.fixture(autouse=__UpperCAmelCase, scope='''session''' ) def __magic_name__ ( ) -> List[Any]: '''simple docstring''' datasets.disable_progress_bar() @pytest.fixture(autouse=__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' monkeypatch.setattr('''datasets.config.HF_UPDATE_DOWNLOAD_COUNTS''', __UpperCAmelCase ) @pytest.fixture def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' monkeypatch.setattr('''sqlalchemy.util.deprecations.SILENCE_UBER_WARNING''', __UpperCAmelCase )
72
1
from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCamelCase_ ( _a = "isbn/0140328726" ): """simple docstring""" lowerCAmelCase__ : Union[str, Any] = olid.strip().strip('''/''' ) # Remove leading/trailing whitespace & slashes if new_olid.count('''/''' ) != 1: lowerCAmelCase__ : Optional[Any] = f'{olid} is not a valid Open Library olid' raise ValueError(_a ) return requests.get(f'https://openlibrary.org/{new_olid}.json' ).json() def lowerCamelCase_ ( _a ): """simple docstring""" lowerCAmelCase__ : Dict = { '''title''': '''Title''', '''publish_date''': '''Publish date''', '''authors''': '''Authors''', '''number_of_pages''': '''Number of pages:''', '''first_sentence''': '''First sentence''', '''isbn_10''': '''ISBN (10)''', '''isbn_13''': '''ISBN (13)''', } lowerCAmelCase__ : Tuple = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} lowerCAmelCase__ : Union[str, Any] = [ get_openlibrary_data(author['''key'''] )['''name'''] for author in data['''Authors'''] ] lowerCAmelCase__ : Optional[Any] = data['''First sentence''']['''value'''] for key, value in data.items(): if isinstance(_a , _a ): lowerCAmelCase__ : List[Any] = ''', '''.join(_a ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: lowerCamelCase = input('''\nEnter the ISBN code to search (or \'quit\' to stop): ''').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(f'''Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.''') continue print(f'''\nSearching Open Library for ISBN: {isbn}...\n''') try: lowerCamelCase = summarize_book(get_openlibrary_data(f'''isbn/{isbn}''')) print('''\n'''.join(f'''{key}: {value}''' for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(f'''Sorry, there are no results for ISBN: {isbn}.''')
131
import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa lowerCamelCase = logging.getLogger(__name__) class _a ( _lowercase): _a : Dict = '''summarization''' _a : int = ['''loss'''] _a : Optional[Any] = ROUGE_KEYS _a : Any = '''rouge2''' def __init__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : List[str] , **_SCREAMING_SNAKE_CASE : Dict )-> int: if hparams.sortish_sampler and hparams.gpus > 1: lowerCAmelCase__ : int = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError('''Dynamic Batch size does not work for multi-gpu training''' ) if hparams.sortish_sampler: raise ValueError('''--sortish_sampler and --max_tokens_per_batch may not be used simultaneously''' ) super().__init__(_SCREAMING_SNAKE_CASE , num_labels=_SCREAMING_SNAKE_CASE , mode=self.mode , **_SCREAMING_SNAKE_CASE ) use_task_specific_params(self.model , '''summarization''' ) save_git_info(self.hparams.output_dir ) lowerCAmelCase__ : List[Any] = Path(self.output_dir ) / '''metrics.json''' lowerCAmelCase__ : int = Path(self.output_dir ) / '''hparams.pkl''' pickle_save(self.hparams , self.hparams_save_path ) lowerCAmelCase__ : Tuple = 0 lowerCAmelCase__ : List[Any] = defaultdict(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Any = self.config.model_type lowerCAmelCase__ : Any = self.config.tgt_vocab_size if self.model_type == '''fsmt''' else self.config.vocab_size lowerCAmelCase__ : dict = { "data_dir": self.hparams.data_dir, "max_source_length": self.hparams.max_source_length, "prefix": self.model.config.prefix or "", } lowerCAmelCase__ : List[Any] = { '''train''': self.hparams.n_train, '''val''': self.hparams.n_val, '''test''': self.hparams.n_test, } lowerCAmelCase__ : Union[str, Any] = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} lowerCAmelCase__ : Tuple = { '''train''': self.hparams.max_target_length, '''val''': self.hparams.val_max_target_length, '''test''': self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], F'target_lens: {self.target_lens}' assert self.target_lens["train"] <= self.target_lens["test"], F'target_lens: {self.target_lens}' if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) lowerCAmelCase__ : List[str] = get_git_info()['''repo_sha'''] lowerCAmelCase__ : int = hparams.num_workers lowerCAmelCase__ : Optional[int] = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : Optional[int] = self.tokenizer.lang_code_to_id[hparams.tgt_lang] lowerCAmelCase__ : List[Any] = self.decoder_start_token_id lowerCAmelCase__ : int = ( SeqaSeqDataset if hasattr(self.tokenizer , '''prepare_seq2seq_batch''' ) else LegacySeqaSeqDataset ) lowerCAmelCase__ : Union[str, Any] = False lowerCAmelCase__ : List[str] = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: lowerCAmelCase__ : int = self.hparams.eval_max_gen_length else: lowerCAmelCase__ : Tuple = self.model.config.max_length lowerCAmelCase__ : List[str] = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def UpperCAmelCase__( self : Dict , _SCREAMING_SNAKE_CASE : Dict[str, torch.Tensor] )-> Dict[str, List[str]]: lowerCAmelCase__ : Any = { k: self.tokenizer.batch_decode(v.tolist() ) if '''mask''' not in k else v.shape for k, v in batch.items() } save_json(_SCREAMING_SNAKE_CASE , Path(self.output_dir ) / '''text_batch.json''' ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / '''tok_batch.json''' ) lowerCAmelCase__ : Union[str, Any] = True return readable_batch def UpperCAmelCase__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : Tuple , **_SCREAMING_SNAKE_CASE : Tuple )-> Tuple: return self.model(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Optional[int] , _SCREAMING_SNAKE_CASE : List[int] )-> Tuple: lowerCAmelCase__ : int = self.tokenizer.batch_decode( _SCREAMING_SNAKE_CASE , skip_special_tokens=_SCREAMING_SNAKE_CASE , clean_up_tokenization_spaces=_SCREAMING_SNAKE_CASE ) return lmap(str.strip , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : dict )-> Tuple: lowerCAmelCase__ : Any = self.tokenizer.pad_token_id lowerCAmelCase__ , lowerCAmelCase__ : int = batch['''input_ids'''], batch['''attention_mask'''] lowerCAmelCase__ : Optional[Any] = batch['''labels'''] if isinstance(self.model , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : List[str] = self.model._shift_right(_SCREAMING_SNAKE_CASE ) else: lowerCAmelCase__ : Any = shift_tokens_right(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero lowerCAmelCase__ : Any = decoder_input_ids self.save_readable_batch(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Tuple = self(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE , use_cache=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = outputs['''logits'''] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id lowerCAmelCase__ : Optional[int] = nn.CrossEntropyLoss(ignore_index=_SCREAMING_SNAKE_CASE ) assert lm_logits.shape[-1] == self.vocab_size lowerCAmelCase__ : Optional[Any] = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: lowerCAmelCase__ : Union[str, Any] = nn.functional.log_softmax(_SCREAMING_SNAKE_CASE , dim=-1 ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = label_smoothed_nll_loss( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , self.hparams.label_smoothing , ignore_index=_SCREAMING_SNAKE_CASE ) return (loss,) @property def UpperCAmelCase__( self : Optional[Any] )-> int: return self.tokenizer.pad_token_id def UpperCAmelCase__( self : Dict , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : List[str] )-> Dict: lowerCAmelCase__ : List[Any] = self._step(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = dict(zip(self.loss_names , _SCREAMING_SNAKE_CASE ) ) # tokens per batch lowerCAmelCase__ : Dict = batch['''input_ids'''].ne(self.pad ).sum() + batch['''labels'''].ne(self.pad ).sum() lowerCAmelCase__ : Optional[int] = batch['''input_ids'''].shape[0] lowerCAmelCase__ : List[Any] = batch['''input_ids'''].eq(self.pad ).sum() lowerCAmelCase__ : List[str] = batch['''input_ids'''].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def UpperCAmelCase__( self : Tuple , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : Optional[int] )-> Dict: return self._generative_step(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : List[Any] , _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : Tuple="val" )-> Dict: self.step_count += 1 lowerCAmelCase__ : Any = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} lowerCAmelCase__ : Union[str, Any] = losses['''loss'''] lowerCAmelCase__ : Union[str, Any] = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ['''gen_time''', '''gen_len'''] } lowerCAmelCase__ : Optional[Any] = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) lowerCAmelCase__ : torch.FloatTensor = torch.tensor(_SCREAMING_SNAKE_CASE ).type_as(_SCREAMING_SNAKE_CASE ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Dict = {F'{prefix}_avg_{k}': x for k, x in losses.items()} lowerCAmelCase__ : Optional[Any] = self.step_count self.metrics[prefix].append(_SCREAMING_SNAKE_CASE ) # callback writes this to self.metrics_save_path lowerCAmelCase__ : Optional[int] = flatten_list([x['''preds'''] for x in outputs] ) return { "log": all_metrics, "preds": preds, F'{prefix}_loss': loss, F'{prefix}_{self.val_metric}': metric_tensor, } def UpperCAmelCase__( self : Union[str, Any] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : List[Any] )-> Dict: return calculate_rouge(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Tuple , _SCREAMING_SNAKE_CASE : dict )-> dict: lowerCAmelCase__ : Any = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') lowerCAmelCase__ : str = self.model.generate( batch['''input_ids'''] , attention_mask=batch['''attention_mask'''] , use_cache=_SCREAMING_SNAKE_CASE , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) lowerCAmelCase__ : List[str] = (time.time() - ta) / batch['''input_ids'''].shape[0] lowerCAmelCase__ : List[str] = self.ids_to_clean_text(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = self.ids_to_clean_text(batch['''labels'''] ) lowerCAmelCase__ : Any = self._step(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = dict(zip(self.loss_names , _SCREAMING_SNAKE_CASE ) ) lowerCAmelCase__ : Dict = self.calc_generative_metrics(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : str = np.mean(lmap(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) base_metrics.update(gen_time=_SCREAMING_SNAKE_CASE , gen_len=_SCREAMING_SNAKE_CASE , preds=_SCREAMING_SNAKE_CASE , target=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) return base_metrics def UpperCAmelCase__( self : Any , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : Optional[Any] )-> Optional[int]: return self._generative_step(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Tuple , _SCREAMING_SNAKE_CASE : int )-> str: return self.validation_epoch_end(_SCREAMING_SNAKE_CASE , prefix='''test''' ) def UpperCAmelCase__( self : Dict , _SCREAMING_SNAKE_CASE : int )-> SeqaSeqDataset: lowerCAmelCase__ : Dict = self.n_obs[type_path] lowerCAmelCase__ : Dict = self.target_lens[type_path] lowerCAmelCase__ : int = self.dataset_class( self.tokenizer , type_path=_SCREAMING_SNAKE_CASE , n_obs=_SCREAMING_SNAKE_CASE , max_target_length=_SCREAMING_SNAKE_CASE , **self.dataset_kwargs , ) return dataset def UpperCAmelCase__( self : List[str] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : bool = False )-> DataLoader: lowerCAmelCase__ : str = self.get_dataset(_SCREAMING_SNAKE_CASE ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": lowerCAmelCase__ : Any = dataset.make_sortish_sampler(_SCREAMING_SNAKE_CASE , distributed=self.hparams.gpus > 1 ) return DataLoader( _SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE , collate_fn=dataset.collate_fn , shuffle=_SCREAMING_SNAKE_CASE , num_workers=self.num_workers , sampler=_SCREAMING_SNAKE_CASE , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": lowerCAmelCase__ : Optional[int] = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( _SCREAMING_SNAKE_CASE , batch_sampler=_SCREAMING_SNAKE_CASE , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( _SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE , collate_fn=dataset.collate_fn , shuffle=_SCREAMING_SNAKE_CASE , num_workers=self.num_workers , sampler=_SCREAMING_SNAKE_CASE , ) def UpperCAmelCase__( self : List[str] )-> DataLoader: lowerCAmelCase__ : Dict = self.get_dataloader('''train''' , batch_size=self.hparams.train_batch_size , shuffle=_SCREAMING_SNAKE_CASE ) return dataloader def UpperCAmelCase__( self : List[str] )-> DataLoader: return self.get_dataloader('''val''' , batch_size=self.hparams.eval_batch_size ) def UpperCAmelCase__( self : List[Any] )-> DataLoader: return self.get_dataloader('''test''' , batch_size=self.hparams.eval_batch_size ) @staticmethod def UpperCAmelCase__( _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : str )-> List[str]: BaseTransformer.add_model_specific_args(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) add_generic_args(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) parser.add_argument( '''--max_source_length''' , default=1024 , type=_SCREAMING_SNAKE_CASE , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--max_target_length''' , default=56 , type=_SCREAMING_SNAKE_CASE , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--val_max_target_length''' , default=142 , type=_SCREAMING_SNAKE_CASE , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--test_max_target_length''' , default=142 , type=_SCREAMING_SNAKE_CASE , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument('''--freeze_encoder''' , action='''store_true''' ) parser.add_argument('''--freeze_embeds''' , action='''store_true''' ) parser.add_argument('''--sortish_sampler''' , action='''store_true''' , default=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--overwrite_output_dir''' , action='''store_true''' , default=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--max_tokens_per_batch''' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--logger_name''' , type=_SCREAMING_SNAKE_CASE , choices=['''default''', '''wandb''', '''wandb_shared'''] , default='''default''' ) parser.add_argument('''--n_train''' , type=_SCREAMING_SNAKE_CASE , default=-1 , required=_SCREAMING_SNAKE_CASE , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_val''' , type=_SCREAMING_SNAKE_CASE , default=500 , required=_SCREAMING_SNAKE_CASE , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_test''' , type=_SCREAMING_SNAKE_CASE , default=-1 , required=_SCREAMING_SNAKE_CASE , help='''# examples. -1 means use all.''' ) parser.add_argument( '''--task''' , type=_SCREAMING_SNAKE_CASE , default='''summarization''' , required=_SCREAMING_SNAKE_CASE , help='''# examples. -1 means use all.''' ) parser.add_argument('''--label_smoothing''' , type=_SCREAMING_SNAKE_CASE , default=0.0 , required=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--src_lang''' , type=_SCREAMING_SNAKE_CASE , default='''''' , required=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--tgt_lang''' , type=_SCREAMING_SNAKE_CASE , default='''''' , required=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--eval_beams''' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE ) parser.add_argument( '''--val_metric''' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE , choices=['''bleu''', '''rouge2''', '''loss''', None] ) parser.add_argument('''--eval_max_gen_length''' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , help='''never generate more than n tokens''' ) parser.add_argument('''--save_top_k''' , type=_SCREAMING_SNAKE_CASE , default=1 , required=_SCREAMING_SNAKE_CASE , help='''How many checkpoints to save''' ) parser.add_argument( '''--early_stopping_patience''' , type=_SCREAMING_SNAKE_CASE , default=-1 , required=_SCREAMING_SNAKE_CASE , help=( '''-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So''' ''' val_check_interval will effect it.''' ) , ) return parser class _a ( _lowercase): _a : Tuple = '''translation''' _a : Union[str, Any] = ['''loss'''] _a : int = ['''bleu'''] _a : List[str] = '''bleu''' def __init__( self : Optional[int] , _SCREAMING_SNAKE_CASE : int , **_SCREAMING_SNAKE_CASE : Any )-> Optional[int]: super().__init__(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Dict = hparams.src_lang lowerCAmelCase__ : str = hparams.tgt_lang def UpperCAmelCase__( self : int , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Tuple )-> dict: return calculate_bleu(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def lowerCamelCase_ ( _a , _a=None ): """simple docstring""" Path(args.output_dir ).mkdir(exist_ok=_a ) check_output_dir(_a , expected_items=3 ) if model is None: if "summarization" in args.task: lowerCAmelCase__ : SummarizationModule = SummarizationModule(_a ) else: lowerCAmelCase__ : SummarizationModule = TranslationModule(_a ) lowerCAmelCase__ : Optional[int] = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith('''/tmp''' ) or str(args.output_dir ).startswith('''/var''' ) ): lowerCAmelCase__ : Any = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger lowerCAmelCase__ : Optional[int] = os.environ.get('''WANDB_PROJECT''' , _a ) lowerCAmelCase__ : Union[str, Any] = WandbLogger(name=model.output_dir.name , project=_a ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger lowerCAmelCase__ : List[Any] = WandbLogger(name=model.output_dir.name , project=f'hf_{dataset}' ) if args.early_stopping_patience >= 0: lowerCAmelCase__ : Optional[Any] = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: lowerCAmelCase__ : int = False lowerCAmelCase__ : str = args.val_metric == '''loss''' lowerCAmelCase__ : pl.Trainer = generic_train( _a , _a , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , _a ) , early_stopping_callback=_a , logger=_a , ) pickle_save(model.hparams , model.output_dir / '''hparams.pkl''' ) if not args.do_predict: return model lowerCAmelCase__ : Any = '''''' lowerCAmelCase__ : Tuple = sorted(glob.glob(os.path.join(args.output_dir , '''*.ckpt''' ) , recursive=_a ) ) if checkpoints: lowerCAmelCase__ : List[str] = checkpoints[-1] lowerCAmelCase__ : Any = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() lowerCamelCase = pl.Trainer.add_argparse_args(parser) lowerCamelCase = SummarizationModule.add_model_specific_args(parser, os.getcwd()) lowerCamelCase = parser.parse_args() main(args)
131
1
from __future__ import annotations import pandas as pd def lowercase (SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : int ) -> list[int]: SCREAMING_SNAKE_CASE = [0] * no_of_processes SCREAMING_SNAKE_CASE = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(_A ): SCREAMING_SNAKE_CASE = burst_time[i] SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 9_99_99_99_99 SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = False # Process until all processes are completed while complete != no_of_processes: for j in range(_A ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: SCREAMING_SNAKE_CASE = remaining_time[j] SCREAMING_SNAKE_CASE = j SCREAMING_SNAKE_CASE = True if not check: increment_time += 1 continue remaining_time[short] -= 1 SCREAMING_SNAKE_CASE = remaining_time[short] if minm == 0: SCREAMING_SNAKE_CASE = 9_99_99_99_99 if remaining_time[short] == 0: complete += 1 SCREAMING_SNAKE_CASE = False # Find finish time of current process SCREAMING_SNAKE_CASE = increment_time + 1 # Calculate waiting time SCREAMING_SNAKE_CASE = finish_time - arrival_time[short] SCREAMING_SNAKE_CASE = finar - burst_time[short] if waiting_time[short] < 0: SCREAMING_SNAKE_CASE = 0 # Increment time increment_time += 1 return waiting_time def lowercase (SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : list[int] ) -> list[int]: SCREAMING_SNAKE_CASE = [0] * no_of_processes for i in range(_A ): SCREAMING_SNAKE_CASE = burst_time[i] + waiting_time[i] return turn_around_time def lowercase (SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : int ) -> None: SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 0 for i in range(_A ): SCREAMING_SNAKE_CASE = total_waiting_time + waiting_time[i] SCREAMING_SNAKE_CASE = total_turn_around_time + turn_around_time[i] print(F'Average waiting time = {total_waiting_time / no_of_processes:.5f}' ) print('Average turn around time =' , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print('''Enter how many process you want to analyze''') __UpperCamelCase = int(input()) __UpperCamelCase = [0] * no_of_processes __UpperCamelCase = [0] * no_of_processes __UpperCamelCase = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print('''Enter the arrival time and burst time for process:--''' + str(i + 1)) __UpperCamelCase = map(int, input().split()) __UpperCamelCase = calculate_waitingtime(arrival_time, burst_time, no_of_processes) __UpperCamelCase = burst_time __UpperCamelCase = no_of_processes __UpperCamelCase = waiting_time __UpperCamelCase = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) __UpperCamelCase = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ '''Process''', '''BurstTime''', '''ArrivalTime''', '''WaitingTime''', '''TurnAroundTime''', ], ) # Printing the dataFrame pd.set_option('''display.max_rows''', fcfs.shape[0] + 1) print(fcfs)
354
"""simple docstring""" import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() __UpperCamelCase = logging.get_logger(__name__) def lowercase (SCREAMING_SNAKE_CASE_ : List[str] ) -> Optional[Any]: SCREAMING_SNAKE_CASE = torch.load(SCREAMING_SNAKE_CASE_ , map_location='cpu' ) if "model" in sd.keys(): SCREAMING_SNAKE_CASE = torch.load(SCREAMING_SNAKE_CASE_ , map_location='cpu' )['model'] # pop unnecessary weights SCREAMING_SNAKE_CASE = [ 'decoder.version', 'decoder.output_projection.weight', ] for key in keys_to_delete: if key in sd: sd.pop(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = { 'decoder.project_in_dim.weight': 'decoder.project_in.weight', 'decoder.project_out_dim.weight': 'decoder.project_out.weight', 'decoder.layer_norm.weight': 'decoder.final_layer_norm.weight', 'decoder.layer_norm.bias': 'decoder.final_layer_norm.bias', } for old_key, new_key in keys_to_rename.items(): if old_key in sd: SCREAMING_SNAKE_CASE = sd.pop(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: SCREAMING_SNAKE_CASE = sd[key] # We split QKV in separate Q,K,V SCREAMING_SNAKE_CASE = key.replace('.qkv_proj.' , '.q_proj.' ) SCREAMING_SNAKE_CASE = key.replace('.qkv_proj.' , '.k_proj.' ) SCREAMING_SNAKE_CASE = key.replace('.qkv_proj.' , '.v_proj.' ) SCREAMING_SNAKE_CASE = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = torch.split(SCREAMING_SNAKE_CASE_ , depth // 3 , dim=0 ) SCREAMING_SNAKE_CASE = q SCREAMING_SNAKE_CASE = k SCREAMING_SNAKE_CASE = v del sd[key] return sd @torch.no_grad() def lowercase (SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Optional[int]=None ) -> List[Any]: SCREAMING_SNAKE_CASE = load_checkpoint(SCREAMING_SNAKE_CASE_ ) if config is not None: SCREAMING_SNAKE_CASE = OPTConfig.from_pretrained(SCREAMING_SNAKE_CASE_ ) else: SCREAMING_SNAKE_CASE = OPTConfig() SCREAMING_SNAKE_CASE = OPTModel(SCREAMING_SNAKE_CASE_ ).half().eval() model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # Check results Path(SCREAMING_SNAKE_CASE_ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') __UpperCamelCase = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
38
0
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : str = logging.get_logger(__name__) UpperCAmelCase_ : int = { '''microsoft/markuplm-base''': '''https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json''', '''microsoft/markuplm-large''': '''https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json''', } class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : Tuple = """markuplm""" def __init__( self : Union[str, Any] , __lowerCamelCase : List[Any]=30_522 , __lowerCamelCase : Union[str, Any]=768 , __lowerCamelCase : Dict=12 , __lowerCamelCase : Optional[int]=12 , __lowerCamelCase : Optional[int]=3_072 , __lowerCamelCase : Optional[int]="gelu" , __lowerCamelCase : Tuple=0.1 , __lowerCamelCase : str=0.1 , __lowerCamelCase : Any=512 , __lowerCamelCase : int=2 , __lowerCamelCase : Tuple=0.02 , __lowerCamelCase : Any=1E-12 , __lowerCamelCase : Any=0 , __lowerCamelCase : Union[str, Any]=0 , __lowerCamelCase : Union[str, Any]=2 , __lowerCamelCase : List[str]=256 , __lowerCamelCase : Union[str, Any]=1_024 , __lowerCamelCase : Any=216 , __lowerCamelCase : Optional[Any]=1_001 , __lowerCamelCase : Tuple=32 , __lowerCamelCase : Dict=50 , __lowerCamelCase : Optional[int]="absolute" , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : int=None , **__lowerCamelCase : str , ): super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , **__lowerCamelCase , ) UpperCamelCase :str = vocab_size UpperCamelCase :List[str] = hidden_size UpperCamelCase :int = num_hidden_layers UpperCamelCase :Dict = num_attention_heads UpperCamelCase :int = hidden_act UpperCamelCase :Optional[int] = intermediate_size UpperCamelCase :List[Any] = hidden_dropout_prob UpperCamelCase :Any = attention_probs_dropout_prob UpperCamelCase :Union[str, Any] = max_position_embeddings UpperCamelCase :Optional[Any] = type_vocab_size UpperCamelCase :Union[str, Any] = initializer_range UpperCamelCase :List[str] = layer_norm_eps UpperCamelCase :Optional[Any] = position_embedding_type UpperCamelCase :Optional[Any] = use_cache UpperCamelCase :Any = classifier_dropout # additional properties UpperCamelCase :Dict = max_depth UpperCamelCase :List[Any] = max_xpath_tag_unit_embeddings UpperCamelCase :List[Any] = max_xpath_subs_unit_embeddings UpperCamelCase :Dict = tag_pad_id UpperCamelCase :Optional[Any] = subs_pad_id UpperCamelCase :Any = xpath_unit_hidden_size
38
"""simple docstring""" import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _lowerCamelCase : List[Any] = logging.get_logger(__name__) _lowerCamelCase : List[str] = { 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/config.json', # See all BART models at https://huggingface.co/models?filter=bart } class lowercase ( __UpperCAmelCase): __lowerCAmelCase : Union[str, Any] = """bart""" __lowerCAmelCase : Optional[int] = ["""past_key_values"""] __lowerCAmelCase : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[Any] , _lowerCamelCase : List[Any]=5_02_65 , _lowerCamelCase : Optional[Any]=10_24 , _lowerCamelCase : Dict=12 , _lowerCamelCase : Dict=40_96 , _lowerCamelCase : Tuple=16 , _lowerCamelCase : Optional[Any]=12 , _lowerCamelCase : Tuple=40_96 , _lowerCamelCase : List[str]=16 , _lowerCamelCase : List[str]=0.0 , _lowerCamelCase : List[str]=0.0 , _lowerCamelCase : int="gelu" , _lowerCamelCase : Any=10_24 , _lowerCamelCase : Union[str, Any]=0.1 , _lowerCamelCase : Tuple=0.0 , _lowerCamelCase : List[Any]=0.0 , _lowerCamelCase : Optional[int]=0.02 , _lowerCamelCase : int=0.0 , _lowerCamelCase : Any=False , _lowerCamelCase : List[Any]=True , _lowerCamelCase : int=3 , _lowerCamelCase : Tuple=1 , _lowerCamelCase : int=0 , _lowerCamelCase : Optional[int]=2 , _lowerCamelCase : Any=True , _lowerCamelCase : str=2 , _lowerCamelCase : str=2 , **_lowerCamelCase : str , ): """simple docstring""" A_ : Dict = vocab_size A_ : Union[str, Any] = max_position_embeddings A_ : Union[str, Any] = d_model A_ : Optional[int] = encoder_ffn_dim A_ : Optional[Any] = encoder_layers A_ : Union[str, Any] = encoder_attention_heads A_ : List[str] = decoder_ffn_dim A_ : List[str] = decoder_layers A_ : Any = decoder_attention_heads A_ : List[Any] = dropout A_ : Optional[int] = attention_dropout A_ : List[Any] = activation_dropout A_ : Tuple = activation_function A_ : Any = init_std A_ : Union[str, Any] = encoder_layerdrop A_ : Optional[Any] = decoder_layerdrop A_ : Tuple = classifier_dropout A_ : Tuple = use_cache A_ : List[str] = encoder_layers A_ : Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=_lowerCamelCase , pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , is_encoder_decoder=_lowerCamelCase , decoder_start_token_id=_lowerCamelCase , forced_eos_token_id=_lowerCamelCase , **_lowerCamelCase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , _lowerCamelCase ): A_ : int = self.bos_token_id warnings.warn( F"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ '''The config can simply be saved and uploaded again to be fixed.''' ) class lowercase ( __UpperCAmelCase): @property def a_ ( self : Optional[Any] ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: A_ : Tuple = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: A_ : Tuple = {0: '''batch'''} A_ : Dict = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: A_ : Optional[Any] = {0: '''batch''', 1: '''decoder_sequence'''} A_ : List[Any] = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(_lowerCamelCase , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. A_ : List[Any] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: A_ , A_ : Any = self.num_layers for i in range(_lowerCamelCase ): A_ : List[Any] = {0: '''batch''', 2: '''past_sequence + sequence'''} A_ : Optional[Any] = {0: '''batch''', 2: '''past_sequence + sequence'''} else: A_ : List[Any] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property def a_ ( self : str ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: A_ : Optional[Any] = super().outputs else: A_ : List[Any] = super(_lowerCamelCase , self ).outputs if self.use_past: A_ , A_ : int = self.num_layers for i in range(_lowerCamelCase ): A_ : Any = {0: '''batch''', 2: '''past_sequence + sequence'''} A_ : Dict = {0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def a_ ( self : Union[str, Any] , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : int = -1 , _lowerCamelCase : int = -1 , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[TensorType] = None , ): """simple docstring""" A_ : Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Generate decoder inputs A_ : Tuple = seq_length if not self.use_past else 1 A_ : List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = {F"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} A_ : Dict = dict(**_lowerCamelCase , **_lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch A_ , A_ : Union[str, Any] = common_inputs['''input_ids'''].shape A_ : Any = common_inputs['''decoder_input_ids'''].shape[1] A_ , A_ : Optional[Any] = self.num_attention_heads A_ : Union[str, Any] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) A_ : Optional[Any] = decoder_seq_length + 3 A_ : Dict = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) A_ : Optional[int] = torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(_lowerCamelCase , _lowerCamelCase )] , dim=1 ) A_ : Optional[Any] = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered A_ , A_ : Optional[Any] = self.num_layers A_ : Optional[Any] = min(_lowerCamelCase , _lowerCamelCase ) A_ : Tuple = max(_lowerCamelCase , _lowerCamelCase ) - min_num_layers A_ : Tuple = '''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(_lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), ) ) # TODO: test this. A_ : List[str] = encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(_lowerCamelCase , _lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase )) ) return common_inputs def a_ ( self : Optional[int] , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : int = -1 , _lowerCamelCase : int = -1 , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[TensorType] = None , ): """simple docstring""" A_ : Union[str, Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch A_ , A_ : Optional[Any] = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values A_ : Union[str, Any] = seqlen + 2 A_ , A_ : Tuple = self.num_layers A_ , A_ : Optional[int] = self.num_attention_heads A_ : List[str] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) A_ : str = common_inputs['''attention_mask'''].dtype A_ : Optional[Any] = torch.cat( [common_inputs['''attention_mask'''], torch.ones(_lowerCamelCase , _lowerCamelCase , dtype=_lowerCamelCase )] , dim=1 ) A_ : int = [ (torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase )) for _ in range(_lowerCamelCase ) ] return common_inputs def a_ ( self : str , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : int = -1 , _lowerCamelCase : int = -1 , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[TensorType] = None , ): """simple docstring""" A_ : List[Any] = compute_effective_axis_dimension( _lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX A_ : Dict = tokenizer.num_special_tokens_to_add(_lowerCamelCase ) A_ : int = compute_effective_axis_dimension( _lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence A_ : List[str] = [''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size A_ : List[str] = dict(tokenizer(_lowerCamelCase , return_tensors=_lowerCamelCase ) ) return common_inputs def a_ ( self : Tuple , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : int = -1 , _lowerCamelCase : int = -1 , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[TensorType] = None , ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: A_ : Tuple = self._generate_dummy_inputs_for_default_and_seqaseq_lm( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) elif self.task == "causal-lm": A_ : Optional[Any] = self._generate_dummy_inputs_for_causal_lm( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) else: A_ : Union[str, Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) return common_inputs def a_ ( self : str , _lowerCamelCase : Dict , _lowerCamelCase : Tuple , _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any] ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: A_ : Tuple = super()._flatten_past_key_values_(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) else: A_ : List[Any] = super(_lowerCamelCase , self )._flatten_past_key_values_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase )
167
0
import math from collections.abc import Callable def lowercase_ ( A__ , A__ , A__ ) -> Optional[Any]: """simple docstring""" snake_case = xa snake_case = xa while True: if x_n == x_na or function(a__ ) == function(a__ ): raise ZeroDivisionError("float division by zero, could not find root" ) snake_case = x_na - ( function(a__ ) / ((function(a__ ) - function(a__ )) / (x_na - x_n)) ) if abs(x_na - x_na ) < 10**-5: return x_na snake_case = x_na snake_case = x_na def lowercase_ ( A__ ) -> Optional[int]: """simple docstring""" return math.pow(a__ , 3 ) - (2 * x) - 5 if __name__ == "__main__": print(intersection(f, 3, 3.5))
371
from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json" ), "google/realm-orqa-nq-openqa": "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json", "google/realm-orqa-nq-reader": "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json", "google/realm-orqa-wq-openqa": "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json", "google/realm-orqa-wq-reader": "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json", # See all REALM models at https://huggingface.co/models?filter=realm } class lowerCamelCase ( A_ ): UpperCAmelCase__ : str = "realm" def __init__(self : Optional[int] , _A : Optional[Any]=3_0_5_2_2 , _A : Tuple=7_6_8 , _A : List[str]=1_2_8 , _A : Optional[Any]=1_2 , _A : Dict=1_2 , _A : Tuple=8 , _A : Dict=3_0_7_2 , _A : Union[str, Any]="gelu_new" , _A : Any=0.1 , _A : int=0.1 , _A : Union[str, Any]=5_1_2 , _A : List[str]=2 , _A : Any=0.02 , _A : int=1E-12 , _A : Tuple=2_5_6 , _A : Optional[Any]=1_0 , _A : Any=1E-3 , _A : int=5 , _A : int=3_2_0 , _A : Dict=1_3_3_5_3_7_1_8 , _A : Any=5_0_0_0 , _A : Union[str, Any]=1 , _A : Dict=0 , _A : int=2 , **_A : Union[str, Any] , ) -> Optional[Any]: super().__init__(pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A ) # Common config snake_case = vocab_size snake_case = max_position_embeddings snake_case = hidden_size snake_case = retriever_proj_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = num_candidates snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = initializer_range snake_case = type_vocab_size snake_case = layer_norm_eps # Reader config snake_case = span_hidden_size snake_case = max_span_width snake_case = reader_layer_norm_eps snake_case = reader_beam_size snake_case = reader_seq_len # Retrieval config snake_case = num_block_records snake_case = searcher_beam_size
137
0
"""simple docstring""" from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = cva.getAffineTransform(lowerCAmelCase_ , lowerCAmelCase_ ) return cva.warpAffine(lowerCAmelCase_ , lowerCAmelCase_ , (rows, cols) ) if __name__ == "__main__": # read original image a__ : Any = cva.imread( str(Path(__file__).resolve().parent.parent / '''image_data''' / '''lena.jpg''') ) # turn image in gray scale value a__ : Optional[Any] = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape a__ , a__ : Union[str, Any] = gray_img.shape # set different points to rotate image a__ : Optional[Any] = np.array([[5_0, 5_0], [2_0_0, 5_0], [5_0, 2_0_0]], np.floataa) a__ : List[str] = np.array([[1_0, 1_0_0], [2_0_0, 5_0], [1_0_0, 2_5_0]], np.floataa) a__ : Optional[Any] = np.array([[5_0, 5_0], [1_5_0, 5_0], [1_2_0, 2_0_0]], np.floataa) a__ : Tuple = np.array([[1_0, 1_0_0], [8_0, 5_0], [1_8_0, 2_5_0]], np.floataa) # add all rotated images in a list a__ : Union[str, Any] = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations a__ : Any = plt.figure(1) a__ : Optional[int] = ['''Original''', '''Rotation 1''', '''Rotation 2''', '''Rotation 3'''] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, '''gray''') plt.title(titles[i]) plt.axis('''off''') plt.subplots_adjust(left=0.0, bottom=0.05, right=1.0, top=0.95) plt.show()
54
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """studio-ousia/luke-base""": """https://huggingface.co/studio-ousia/luke-base/resolve/main/config.json""", """studio-ousia/luke-large""": """https://huggingface.co/studio-ousia/luke-large/resolve/main/config.json""", } class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = '''luke''' def __init__( self : Any , _A : int=5_0267 , _A : str=50_0000 , _A : Dict=768 , _A : int=256 , _A : Tuple=12 , _A : Optional[Any]=12 , _A : Any=3072 , _A : Tuple="gelu" , _A : Any=0.1 , _A : Dict=0.1 , _A : Any=512 , _A : Tuple=2 , _A : int=0.02 , _A : Any=1e-12 , _A : Dict=True , _A : Optional[Any]=None , _A : List[str]=1 , _A : List[str]=0 , _A : Dict=2 , **_A : List[str] , ): """simple docstring""" super().__init__(pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A ) __SCREAMING_SNAKE_CASE : Optional[Any] = vocab_size __SCREAMING_SNAKE_CASE : Any = entity_vocab_size __SCREAMING_SNAKE_CASE : int = hidden_size __SCREAMING_SNAKE_CASE : List[Any] = entity_emb_size __SCREAMING_SNAKE_CASE : Union[str, Any] = num_hidden_layers __SCREAMING_SNAKE_CASE : Tuple = num_attention_heads __SCREAMING_SNAKE_CASE : Dict = hidden_act __SCREAMING_SNAKE_CASE : Dict = intermediate_size __SCREAMING_SNAKE_CASE : Optional[int] = hidden_dropout_prob __SCREAMING_SNAKE_CASE : Any = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE : List[Any] = max_position_embeddings __SCREAMING_SNAKE_CASE : List[str] = type_vocab_size __SCREAMING_SNAKE_CASE : Dict = initializer_range __SCREAMING_SNAKE_CASE : Tuple = layer_norm_eps __SCREAMING_SNAKE_CASE : int = use_entity_aware_attention __SCREAMING_SNAKE_CASE : Any = classifier_dropout
303
0
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { 'transfo-xl-wt103': 'https://huggingface.co/transfo-xl-wt103/resolve/main/config.json', } class __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" snake_case__ = "transfo-xl" snake_case__ = ["mems"] snake_case__ = { "n_token": "vocab_size", "hidden_size": "d_model", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Tuple , SCREAMING_SNAKE_CASE__ : Any=267_735 , SCREAMING_SNAKE_CASE__ : Optional[Any]=[20_000, 40_000, 200_000] , SCREAMING_SNAKE_CASE__ : Tuple=1_024 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=1_024 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=16 , SCREAMING_SNAKE_CASE__ : List[str]=64 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=4_096 , SCREAMING_SNAKE_CASE__ : Dict=4 , SCREAMING_SNAKE_CASE__ : List[str]=False , SCREAMING_SNAKE_CASE__ : Tuple=18 , SCREAMING_SNAKE_CASE__ : List[str]=1_600 , SCREAMING_SNAKE_CASE__ : Any=1_000 , SCREAMING_SNAKE_CASE__ : Optional[int]=True , SCREAMING_SNAKE_CASE__ : Optional[int]=True , SCREAMING_SNAKE_CASE__ : Tuple=0 , SCREAMING_SNAKE_CASE__ : Optional[int]=-1 , SCREAMING_SNAKE_CASE__ : Dict=True , SCREAMING_SNAKE_CASE__ : Any=0.1 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0.0 , SCREAMING_SNAKE_CASE__ : Dict=True , SCREAMING_SNAKE_CASE__ : Union[str, Any]="normal" , SCREAMING_SNAKE_CASE__ : str=0.01 , SCREAMING_SNAKE_CASE__ : Optional[int]=0.01 , SCREAMING_SNAKE_CASE__ : Optional[int]=0.02 , SCREAMING_SNAKE_CASE__ : List[Any]=1e-5 , SCREAMING_SNAKE_CASE__ : str=0 , **SCREAMING_SNAKE_CASE__ : Tuple , ) -> List[str]: lowerCAmelCase__ = vocab_size lowerCAmelCase__ = [] self.cutoffs.extend(SCREAMING_SNAKE_CASE__ ) if proj_share_all_but_first: lowerCAmelCase__ = [False] + [True] * len(self.cutoffs ) else: lowerCAmelCase__ = [False] + [False] * len(self.cutoffs ) lowerCAmelCase__ = d_model lowerCAmelCase__ = d_embed lowerCAmelCase__ = d_head lowerCAmelCase__ = d_inner lowerCAmelCase__ = div_val lowerCAmelCase__ = pre_lnorm lowerCAmelCase__ = n_layer lowerCAmelCase__ = n_head lowerCAmelCase__ = mem_len lowerCAmelCase__ = same_length lowerCAmelCase__ = attn_type lowerCAmelCase__ = clamp_len lowerCAmelCase__ = sample_softmax lowerCAmelCase__ = adaptive lowerCAmelCase__ = dropout lowerCAmelCase__ = dropatt lowerCAmelCase__ = untie_r lowerCAmelCase__ = init lowerCAmelCase__ = init_range lowerCAmelCase__ = proj_init_std lowerCAmelCase__ = init_std lowerCAmelCase__ = layer_norm_epsilon super().__init__(eos_token_id=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @property def a ( self : Optional[Any] ) -> Union[str, Any]: # Message copied from Transformer-XL documentation logger.info(f'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def a ( self : Any , SCREAMING_SNAKE_CASE__ : List[str] ) -> List[str]: # Message copied from Transformer-XL documentation raise NotImplementedError( f'The model {self.model_type} is one of the few models that has no sequence length limit.' )
221
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class __lowerCamelCase : """simple docstring""" def __init__( self : str , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[Any]=13 , SCREAMING_SNAKE_CASE__ : Optional[Any]=7 , SCREAMING_SNAKE_CASE__ : str=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=True , SCREAMING_SNAKE_CASE__ : Dict=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=True , SCREAMING_SNAKE_CASE__ : int=99 , SCREAMING_SNAKE_CASE__ : List[str]=32 , SCREAMING_SNAKE_CASE__ : Tuple=2 , SCREAMING_SNAKE_CASE__ : List[str]=4 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=37 , SCREAMING_SNAKE_CASE__ : List[Any]="gelu" , SCREAMING_SNAKE_CASE__ : int=0.1 , SCREAMING_SNAKE_CASE__ : Dict=0.1 , SCREAMING_SNAKE_CASE__ : List[Any]=512 , SCREAMING_SNAKE_CASE__ : List[Any]=16 , SCREAMING_SNAKE_CASE__ : List[str]=2 , SCREAMING_SNAKE_CASE__ : Optional[int]=0.02 , SCREAMING_SNAKE_CASE__ : Tuple=3 , SCREAMING_SNAKE_CASE__ : int=4 , SCREAMING_SNAKE_CASE__ : Optional[Any]=None , SCREAMING_SNAKE_CASE__ : str=0 , ) -> List[Any]: lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = seq_length lowerCAmelCase__ = is_training lowerCAmelCase__ = use_input_mask lowerCAmelCase__ = use_token_type_ids lowerCAmelCase__ = use_labels lowerCAmelCase__ = vocab_size lowerCAmelCase__ = hidden_size lowerCAmelCase__ = num_hidden_layers lowerCAmelCase__ = num_attention_heads lowerCAmelCase__ = intermediate_size lowerCAmelCase__ = hidden_act lowerCAmelCase__ = hidden_dropout_prob lowerCAmelCase__ = attention_probs_dropout_prob lowerCAmelCase__ = max_position_embeddings lowerCAmelCase__ = type_vocab_size lowerCAmelCase__ = type_sequence_label_size lowerCAmelCase__ = initializer_range lowerCAmelCase__ = num_labels lowerCAmelCase__ = num_choices lowerCAmelCase__ = scope lowerCAmelCase__ = projection_dim def a ( self : Any ) -> Any: lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ = None if self.use_input_mask: # follow test_modeling_tf_ctrl.py lowerCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ = None if self.use_token_type_ids: lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None if self.use_labels: lowerCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=SCREAMING_SNAKE_CASE__ , initializer_range=self.initializer_range , ) lowerCAmelCase__ = DPRConfig(projection_dim=self.projection_dim , **config.to_dict() ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a ( self : Dict , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[Any] ) -> List[str]: lowerCAmelCase__ = TFDPRContextEncoder(config=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , attention_mask=SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def a ( self : Dict , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : str ) -> List[str]: lowerCAmelCase__ = TFDPRQuestionEncoder(config=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , attention_mask=SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def a ( self : Dict , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple ) -> List[Any]: lowerCAmelCase__ = TFDPRReader(config=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ , attention_mask=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.relevance_logits.shape , (self.batch_size,) ) def a ( self : Optional[int] ) -> Union[str, Any]: lowerCAmelCase__ = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) = config_and_inputs lowerCAmelCase__ = {"input_ids": input_ids} return config, inputs_dict @require_tf class __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): """simple docstring""" snake_case__ = ( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) snake_case__ = {"feature-extraction": TFDPRQuestionEncoder} if is_tf_available() else {} snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False def a ( self : Dict ) -> int: lowerCAmelCase__ = TFDPRModelTester(self ) lowerCAmelCase__ = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , hidden_size=37 ) def a ( self : List[Any] ) -> List[Any]: self.config_tester.run_common_tests() def a ( self : Tuple ) -> Dict: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*SCREAMING_SNAKE_CASE__ ) def a ( self : Union[str, Any] ) -> Optional[int]: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*SCREAMING_SNAKE_CASE__ ) def a ( self : List[Any] ) -> Dict: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*SCREAMING_SNAKE_CASE__ ) @slow def a ( self : List[str] ) -> Union[str, Any]: for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ = TFDPRContextEncoder.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ = TFDPRContextEncoder.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ = TFDPRQuestionEncoder.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ = TFDPRReader.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) @require_tf class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" @slow def a ( self : Optional[Any] ) -> List[str]: lowerCAmelCase__ = TFDPRQuestionEncoder.from_pretrained("facebook/dpr-question_encoder-single-nq-base" ) lowerCAmelCase__ = tf.constant( [[101, 7_592, 1_010, 2_003, 2_026, 3_899, 10_140, 1_029, 102]] ) # [CLS] hello, is my dog cute? [SEP] lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ )[0] # embedding shape = (1, 768) # compare the actual values for a slice. lowerCAmelCase__ = tf.constant( [ [ 0.03_236_253, 0.12_753_335, 0.16_818_509, 0.00_279_786, 0.3_896_933, 0.24_264_945, 0.2_178_971, -0.02_335_227, -0.08_481_959, -0.14_324_117, ] ] ) self.assertTrue(numpy.allclose(output[:, :10].numpy() , expected_slice.numpy() , atol=1e-4 ) )
221
1
from __future__ import annotations import requests __lowerCamelCase : int = set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def _snake_case ( lowerCAmelCase : str , lowerCAmelCase : int = 1 , lowerCAmelCase : str = "new" , lowerCAmelCase : list | None = None ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(lowerCAmelCase ) - valid_terms ) ): SCREAMING_SNAKE_CASE_ : List[Any] = f'Invalid search term: {invalid_search_terms}' raise ValueError(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] = requests.get( f'https://reddit.com/r/{subreddit}/{age}.json?limit={limit}' , headers={"User-agent": "A random string"} , ) if response.status_code == 4_2_9: raise requests.HTTPError SCREAMING_SNAKE_CASE_ : Union[str, Any] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(lowerCAmelCase )} SCREAMING_SNAKE_CASE_ : List[str] = {} for id_ in range(lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : int = { item: data["data"]["children"][id_]["data"][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
18
"""simple docstring""" # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys lowercase__ = """3""" print("""Python version:""", sys.version) print("""OS platform:""", platform.platform()) print("""OS architecture:""", platform.machine()) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) print("""Cuda version:""", torch.version.cuda) print("""CuDNN version:""", torch.backends.cudnn.version()) print("""Number of GPUs available:""", torch.cuda.device_count()) except ImportError: print("""Torch version:""", None) try: import transformers print("""transformers version:""", transformers.__version__) except ImportError: print("""transformers version:""", None)
241
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCamelCase : Any = { '''configuration_poolformer''': [ '''POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PoolFormerConfig''', '''PoolFormerOnnxConfig''', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Dict = ['''PoolFormerFeatureExtractor'''] _lowerCamelCase : Tuple = ['''PoolFormerImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[Any] = [ '''POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''PoolFormerForImageClassification''', '''PoolFormerModel''', '''PoolFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys _lowerCamelCase : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
206
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class lowercase ( a ): lowercase__ : Optional[Any] = ( """This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image.""" """It takes two arguments named `image` which should be the original image, and `label` which should be a text """ """describing the elements what should be identified in the segmentation mask. The tool returns the mask.""" ) lowercase__ : Optional[int] = """CIDAS/clipseg-rd64-refined""" lowercase__ : Tuple = """image_segmenter""" lowercase__ : Optional[Any] = CLIPSegForImageSegmentation lowercase__ : int = ["""image""", """text"""] lowercase__ : List[str] = ["""image"""] def __init__( self : str , *_UpperCamelCase : str , **_UpperCamelCase : Union[str, Any] ) -> Optional[int]: '''simple docstring''' requires_backends(self , ["vision"] ) super().__init__(*_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : int , _UpperCamelCase : "Image" , _UpperCamelCase : str ) -> Optional[int]: '''simple docstring''' return self.pre_processor(text=[label] , images=[image] , padding=_UpperCamelCase , return_tensors="pt" ) def __snake_case( self : Union[str, Any] , _UpperCamelCase : str ) -> Union[str, Any]: '''simple docstring''' with torch.no_grad(): SCREAMING_SNAKE_CASE = self.model(**_UpperCamelCase ).logits return logits def __snake_case( self : Any , _UpperCamelCase : Optional[Any] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = outputs.cpu().detach().numpy() SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 1 return Image.fromarray((array * 255).astype(np.uinta ) )
206
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a__ : int = { '''configuration_blenderbot_small''': [ '''BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlenderbotSmallConfig''', '''BlenderbotSmallOnnxConfig''', ], '''tokenization_blenderbot_small''': ['''BlenderbotSmallTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = ['''BlenderbotSmallTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : str = [ '''BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlenderbotSmallForCausalLM''', '''BlenderbotSmallForConditionalGeneration''', '''BlenderbotSmallModel''', '''BlenderbotSmallPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Tuple = [ '''TFBlenderbotSmallForConditionalGeneration''', '''TFBlenderbotSmallModel''', '''TFBlenderbotSmallPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] = [ '''FlaxBlenderbotSmallForConditionalGeneration''', '''FlaxBlenderbotSmallModel''', '''FlaxBlenderbotSmallPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotSmallConfig, BlenderbotSmallOnnxConfig, ) from .tokenization_blenderbot_small import BlenderbotSmallTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_small_fast import BlenderbotSmallTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotSmallForCausalLM, BlenderbotSmallForConditionalGeneration, BlenderbotSmallModel, BlenderbotSmallPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot_small import ( TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel, TFBlenderbotSmallPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, FlaxBlenderbotSmallPreTrainedModel, ) else: import sys a__ : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
54
"""simple docstring""" import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor _lowercase : Any = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' def __init__( self : Any, *lowerCamelCase : str, **lowerCamelCase : Optional[Any] )-> None: warnings.warn( '''The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use OwlViTImageProcessor instead.''', lowerCamelCase, ) super().__init__(*lowerCamelCase, **lowerCamelCase )
238
0
'''simple docstring''' from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=a ) class lowerCAmelCase__ ( a ): """simple docstring""" lowerCAmelCase__ = field(default="summarization" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCAmelCase__ = Features({"text": Value("string" )} ) lowerCAmelCase__ = Features({"summary": Value("string" )} ) lowerCAmelCase__ = "text" lowerCAmelCase__ = "summary" @property def UpperCAmelCase__ ( self : Dict ) -> Dict[str, str]: """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
355
'''simple docstring''' import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def a__ ( a__ ): """simple docstring""" return x + 2 class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self : Any ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE = """x = 3""" __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) assert result == 3 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3} ) __SCREAMING_SNAKE_CASE = """x = y""" __SCREAMING_SNAKE_CASE = {"""y""": 5} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 5, """y""": 5} ) def UpperCAmelCase__ ( self : Optional[int] ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = """y = add_two(x)""" __SCREAMING_SNAKE_CASE = {"""x""": 3} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {"""add_two""": add_two} , state=__SCREAMING_SNAKE_CASE ) assert result == 5 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """y""": 5} ) # Won't work without the tool with CaptureStdout() as out: __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) assert result is None assert "tried to execute add_two" in out.out def UpperCAmelCase__ ( self : str ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = """x = 3""" __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) assert result == 3 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3} ) def UpperCAmelCase__ ( self : List[str] ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = """test_dict = {'x': x, 'y': add_two(x)}""" __SCREAMING_SNAKE_CASE = {"""x""": 3} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {"""add_two""": add_two} , state=__SCREAMING_SNAKE_CASE ) self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """y""": 5} ) self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """test_dict""": {"""x""": 3, """y""": 5}} ) def UpperCAmelCase__ ( self : Optional[Any] ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE = """x = 3\ny = 5""" __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """y""": 5} ) def UpperCAmelCase__ ( self : Union[str, Any] ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE = """text = f'This is x: {x}.'""" __SCREAMING_SNAKE_CASE = {"""x""": 3} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) # evaluate returns the value of the last assignment. assert result == "This is x: 3." self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """text""": """This is x: 3."""} ) def UpperCAmelCase__ ( self : Dict ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = """if x <= 3:\n y = 2\nelse:\n y = 5""" __SCREAMING_SNAKE_CASE = {"""x""": 3} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) # evaluate returns the value of the last assignment. assert result == 2 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """y""": 2} ) __SCREAMING_SNAKE_CASE = {"""x""": 8} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 8, """y""": 5} ) def UpperCAmelCase__ ( self : Dict ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = """test_list = [x, add_two(x)]""" __SCREAMING_SNAKE_CASE = {"""x""": 3} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {"""add_two""": add_two} , state=__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , [3, 5] ) self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """test_list""": [3, 5]} ) def UpperCAmelCase__ ( self : Tuple ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = """y = x""" __SCREAMING_SNAKE_CASE = {"""x""": 3} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {} , state=__SCREAMING_SNAKE_CASE ) assert result == 3 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """y""": 3} ) def UpperCAmelCase__ ( self : Any ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = """test_list = [x, add_two(x)]\ntest_list[1]""" __SCREAMING_SNAKE_CASE = {"""x""": 3} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {"""add_two""": add_two} , state=__SCREAMING_SNAKE_CASE ) assert result == 5 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """test_list""": [3, 5]} ) __SCREAMING_SNAKE_CASE = """test_dict = {'x': x, 'y': add_two(x)}\ntest_dict['y']""" __SCREAMING_SNAKE_CASE = {"""x""": 3} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {"""add_two""": add_two} , state=__SCREAMING_SNAKE_CASE ) assert result == 5 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 3, """test_dict""": {"""x""": 3, """y""": 5}} ) def UpperCAmelCase__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE = """x = 0\nfor i in range(3):\n x = i""" __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = evaluate(__SCREAMING_SNAKE_CASE , {"""range""": range} , state=__SCREAMING_SNAKE_CASE ) assert result == 2 self.assertDictEqual(__SCREAMING_SNAKE_CASE , {"""x""": 2, """i""": 2} )
331
0
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , ): snake_case_ = {} if train_file is not None: snake_case_ = [train_file] if eval_file is not None: snake_case_ = [eval_file] if test_file is not None: snake_case_ = [test_file] snake_case_ = datasets.load_dataset('''csv''' , data_files=SCREAMING_SNAKE_CASE__ ) snake_case_ = list(ds[list(files.keys() )[0]].features.keys() ) snake_case_ = features_name.pop(SCREAMING_SNAKE_CASE__ ) snake_case_ = list(set(ds[list(files.keys() )[0]][label_name] ) ) snake_case_ = {label: i for i, label in enumerate(SCREAMING_SNAKE_CASE__ )} snake_case_ = tokenizer.model_input_names snake_case_ = {} if len(SCREAMING_SNAKE_CASE__ ) == 1: for k in files.keys(): snake_case_ = ds[k].map( lambda SCREAMING_SNAKE_CASE__ : tokenizer.batch_encode_plus( example[features_name[0]] , truncation=SCREAMING_SNAKE_CASE__ , max_length=SCREAMING_SNAKE_CASE__ , padding='''max_length''' ) , batched=SCREAMING_SNAKE_CASE__ , ) elif len(SCREAMING_SNAKE_CASE__ ) == 2: for k in files.keys(): snake_case_ = ds[k].map( lambda SCREAMING_SNAKE_CASE__ : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=SCREAMING_SNAKE_CASE__ , max_length=SCREAMING_SNAKE_CASE__ , padding='''max_length''' , ) , batched=SCREAMING_SNAKE_CASE__ , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: snake_case_ = {k: v for k, v in ex.items() if k in input_names} snake_case_ = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: snake_case_ = {k: v for k, v in ex.items() if k in input_names} snake_case_ = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: snake_case_ = {k: v for k, v in ex.items() if k in input_names} snake_case_ = labelaid[ex[label_name]] yield (d, label) snake_case_ = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: snake_case_ = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) snake_case_ = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: snake_case_ = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) snake_case_ = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE__ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: snake_case_ = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid lowerCAmelCase_ = logging.getLogger(__name__) @dataclass class snake_case_ : '''simple docstring''' SCREAMING_SNAKE_CASE : int = field(metadata={"help": "Which column contains the label"} ) SCREAMING_SNAKE_CASE : str = field(default=__A , metadata={"help": "The path of the training file"} ) SCREAMING_SNAKE_CASE : Optional[str] = field(default=__A , metadata={"help": "The path of the development file"} ) SCREAMING_SNAKE_CASE : Optional[str] = field(default=__A , metadata={"help": "The path of the test file"} ) SCREAMING_SNAKE_CASE : int = field( default=128 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) SCREAMING_SNAKE_CASE : bool = field( default=__A , metadata={"help": "Overwrite the cached training and evaluation sets"} ) @dataclass class snake_case_ : '''simple docstring''' SCREAMING_SNAKE_CASE : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) SCREAMING_SNAKE_CASE : Optional[str] = field( default=__A , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) SCREAMING_SNAKE_CASE : Optional[str] = field( default=__A , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) SCREAMING_SNAKE_CASE : bool = field(default=__A , metadata={"help": "Set this flag to use fast tokenization."} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. SCREAMING_SNAKE_CASE : Optional[str] = field( default=__A , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) def __SCREAMING_SNAKE_CASE (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) snake_case_, snake_case_, snake_case_ = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO , ) logger.info( F'''n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, ''' F'''16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case_, snake_case_, snake_case_, snake_case_ = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=SCREAMING_SNAKE_CASE__ , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) snake_case_ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(SCREAMING_SNAKE_CASE__ ) , labelaid=SCREAMING_SNAKE_CASE__ , idalabel={id: label for label, id in labelaid.items()} , finetuning_task='''text-classification''' , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): snake_case_ = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool('''.bin''' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE__ , cache_dir=model_args.cache_dir , ) def compute_metrics(SCREAMING_SNAKE_CASE__ ) -> Dict: snake_case_ = np.argmax(p.predictions , axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer snake_case_ = TFTrainer( model=SCREAMING_SNAKE_CASE__ , args=SCREAMING_SNAKE_CASE__ , train_dataset=SCREAMING_SNAKE_CASE__ , eval_dataset=SCREAMING_SNAKE_CASE__ , compute_metrics=SCREAMING_SNAKE_CASE__ , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case_ = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case_ = trainer.evaluate() snake_case_ = os.path.join(training_args.output_dir , '''eval_results.txt''' ) with open(SCREAMING_SNAKE_CASE__ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(F''' {key} = {value}''' ) writer.write(F'''{key} = {value}\n''' ) results.update(SCREAMING_SNAKE_CASE__ ) return results if __name__ == "__main__": main()
8
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def __SCREAMING_SNAKE_CASE (*SCREAMING_SNAKE_CASE__ ): if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): snake_case_ = list(SCREAMING_SNAKE_CASE__ ) for i in range(len(SCREAMING_SNAKE_CASE__ ) ): snake_case_ = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): snake_case_ = [ '''CUDA out of memory.''', # CUDA OOM '''cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.''', # CUDNN SNAFU '''DefaultCPUAllocator: can\'t allocate memory''', # CPU OOM ] if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = 128 ): if function is None: return functools.partial(SCREAMING_SNAKE_CASE__ , starting_batch_size=SCREAMING_SNAKE_CASE__ ) snake_case_ = starting_batch_size def decorator(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() snake_case_ = list(inspect.signature(SCREAMING_SNAKE_CASE__ ).parameters.keys() ) # Guard against user error if len(SCREAMING_SNAKE_CASE__ ) < (len(SCREAMING_SNAKE_CASE__ ) + 1): snake_case_ = ''', '''.join([F'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'''Batch size was passed into `{function.__name__}` as the first argument when called.''' F'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError('''No executable batch size found, reached zero.''' ) try: return function(SCREAMING_SNAKE_CASE__ , *SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) except Exception as e: if should_reduce_batch_size(SCREAMING_SNAKE_CASE__ ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
8
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __snake_case = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
369
'''simple docstring''' import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class lowercase ( A__ ): """simple docstring""" _a = (DDIMParallelScheduler,) _a = (('eta', 0.0), ('num_inference_steps', 50)) def lowerCAmelCase__ ( self , **UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[Any] = { '''num_train_timesteps''': 1000, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''clip_sample''': True, } config.update(**UpperCamelCase_ ) return config def lowerCAmelCase__ ( self , **UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :str = self.scheduler_classes[0] UpperCamelCase__ :Optional[Any] = self.get_scheduler_config(**UpperCamelCase_ ) UpperCamelCase__ :Optional[Any] = scheduler_class(**UpperCamelCase_ ) UpperCamelCase__ , UpperCamelCase__ :Optional[int] = 10, 0.0 UpperCamelCase__ :List[str] = self.dummy_model() UpperCamelCase__ :List[str] = self.dummy_sample_deter scheduler.set_timesteps(UpperCamelCase_ ) for t in scheduler.timesteps: UpperCamelCase__ :Optional[int] = model(UpperCamelCase_ , UpperCamelCase_ ) UpperCamelCase__ :List[Any] = scheduler.step(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ).prev_sample return sample def lowerCAmelCase__ ( self ): '''simple docstring''' for timesteps in [100, 500, 1000]: self.check_over_configs(num_train_timesteps=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for steps_offset in [0, 1]: self.check_over_configs(steps_offset=UpperCamelCase_ ) UpperCamelCase__ :List[str] = self.scheduler_classes[0] UpperCamelCase__ :List[str] = self.get_scheduler_config(steps_offset=1 ) UpperCamelCase__ :str = scheduler_class(**UpperCamelCase_ ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps , torch.LongTensor([801, 601, 401, 201, 1] ) ) def lowerCAmelCase__ ( self ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=UpperCamelCase_ , beta_end=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' self.check_over_configs(thresholding=UpperCamelCase_ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=UpperCamelCase_ , prediction_type=UpperCamelCase_ , sample_max_value=UpperCamelCase_ , ) def lowerCAmelCase__ ( self ): '''simple docstring''' for t in [1, 10, 49]: self.check_over_forward(time_step=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for t, num_inference_steps in zip([1, 10, 50] , [10, 50, 500] ): self.check_over_forward(time_step=UpperCamelCase_ , num_inference_steps=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' for t, eta in zip([1, 10, 49] , [0.0, 0.5, 1.0] ): self.check_over_forward(time_step=UpperCamelCase_ , eta=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = self.scheduler_classes[0] UpperCamelCase__ :List[str] = self.get_scheduler_config() UpperCamelCase__ :int = scheduler_class(**UpperCamelCase_ ) assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(420 , 400 ) - 0.14771 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(980 , 960 ) - 0.32460 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 , 486 ) - 0.00979 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 , 998 ) - 0.02 ) ) < 1e-5 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self.scheduler_classes[0] UpperCamelCase__ :Dict = self.get_scheduler_config() UpperCamelCase__ :Any = scheduler_class(**UpperCamelCase_ ) UpperCamelCase__ , UpperCamelCase__ :List[str] = 10, 0.0 scheduler.set_timesteps(UpperCamelCase_ ) UpperCamelCase__ :Optional[int] = self.dummy_model() UpperCamelCase__ :List[Any] = self.dummy_sample_deter UpperCamelCase__ :int = self.dummy_sample_deter + 0.1 UpperCamelCase__ :List[str] = self.dummy_sample_deter - 0.1 UpperCamelCase__ :str = samplea.shape[0] UpperCamelCase__ :Optional[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) UpperCamelCase__ :List[Any] = torch.arange(UpperCamelCase_ )[0:3, None].repeat(1 , UpperCamelCase_ ) UpperCamelCase__ :Dict = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) UpperCamelCase__ :Any = scheduler.batch_step_no_noise(UpperCamelCase_ , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) , UpperCamelCase_ ) UpperCamelCase__ :Dict = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :Any = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 1147.7904 ) < 1e-2 assert abs(result_mean.item() - 0.4982 ) < 1e-3 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = self.full_loop() UpperCamelCase__ :List[str] = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :Any = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 172.0067 ) < 1e-2 assert abs(result_mean.item() - 0.223967 ) < 1e-3 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Any = self.full_loop(prediction_type='''v_prediction''' ) UpperCamelCase__ :Dict = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :str = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 52.5302 ) < 1e-2 assert abs(result_mean.item() - 0.0684 ) < 1e-3 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = self.full_loop(set_alpha_to_one=UpperCamelCase_ , beta_start=0.01 ) UpperCamelCase__ :str = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :int = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 149.8295 ) < 1e-2 assert abs(result_mean.item() - 0.1951 ) < 1e-3 def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[str] = self.full_loop(set_alpha_to_one=UpperCamelCase_ , beta_start=0.01 ) UpperCamelCase__ :Dict = torch.sum(torch.abs(UpperCamelCase_ ) ) UpperCamelCase__ :List[str] = torch.mean(torch.abs(UpperCamelCase_ ) ) assert abs(result_sum.item() - 149.0784 ) < 1e-2 assert abs(result_mean.item() - 0.1941 ) < 1e-3
219
0
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger __snake_case :Union[str, Any] = get_logger(__name__) class _A : def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : Optional[str] = None): '''simple docstring''' __a = ( os.path.join(__SCREAMING_SNAKE_CASE , config.EXTRACTED_DATASETS_DIR) if cache_dir else config.EXTRACTED_DATASETS_PATH ) __a = Extractor def _lowerCamelCase ( self : List[Any] , __SCREAMING_SNAKE_CASE : str): '''simple docstring''' from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" __a = os.path.abspath(__SCREAMING_SNAKE_CASE) return os.path.join(self.extract_dir , hash_url_to_filename(__SCREAMING_SNAKE_CASE)) def _lowerCamelCase ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : bool): '''simple docstring''' return force_extract or ( not os.path.isfile(__SCREAMING_SNAKE_CASE) and not (os.path.isdir(__SCREAMING_SNAKE_CASE) and os.listdir(__SCREAMING_SNAKE_CASE)) ) def _lowerCamelCase ( self : str , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : bool = False): '''simple docstring''' __a = self.extractor.infer_extractor_format(__SCREAMING_SNAKE_CASE) if not extractor_format: return input_path __a = self._get_output_path(__SCREAMING_SNAKE_CASE) if self._do_extract(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): self.extractor.extract(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) return output_path class _A ( __UpperCAmelCase ): @classmethod @abstractmethod def _lowerCamelCase ( cls : Dict , __SCREAMING_SNAKE_CASE : Union[Path, str] , **__SCREAMING_SNAKE_CASE : Tuple): '''simple docstring''' ... @staticmethod @abstractmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' ... class _A ( __UpperCAmelCase ,__UpperCAmelCase ): UpperCamelCase__ : List[bytes] = [] @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : int): '''simple docstring''' with open(__SCREAMING_SNAKE_CASE , '''rb''') as f: return f.read(__SCREAMING_SNAKE_CASE) @classmethod def _lowerCamelCase ( cls : Any , __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : bytes = b""): '''simple docstring''' if not magic_number: __a = max(len(__SCREAMING_SNAKE_CASE) for cls_magic_number in cls.magic_numbers) try: __a = cls.read_magic_number(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) except OSError: return False return any(magic_number.startswith(__SCREAMING_SNAKE_CASE) for cls_magic_number in cls.magic_numbers) class _A ( __UpperCAmelCase ): @classmethod def _lowerCamelCase ( cls : Optional[int] , __SCREAMING_SNAKE_CASE : Union[Path, str] , **__SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' return tarfile.is_tarfile(__SCREAMING_SNAKE_CASE) @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Tuple): '''simple docstring''' def resolved(__SCREAMING_SNAKE_CASE : str) -> str: return os.path.realpath(os.path.abspath(__SCREAMING_SNAKE_CASE)) def badpath(__SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : str) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE)).startswith(__SCREAMING_SNAKE_CASE) def badlink(__SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : str) -> bool: # Links are interpreted relative to the directory containing the link __a = resolved(os.path.join(__SCREAMING_SNAKE_CASE , os.path.dirname(info.name))) return badpath(info.linkname , base=__SCREAMING_SNAKE_CASE) __a = resolved(__SCREAMING_SNAKE_CASE) for finfo in members: if badpath(finfo.name , __SCREAMING_SNAKE_CASE): logger.error(F'Extraction of {finfo.name} is blocked (illegal path)') elif finfo.issym() and badlink(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): logger.error(F'Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}') elif finfo.islnk() and badlink(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): logger.error(F'Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}') else: yield finfo @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE) __a = tarfile.open(__SCREAMING_SNAKE_CASE) tar_file.extractall(__SCREAMING_SNAKE_CASE , members=TarExtractor.safemembers(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE)) tar_file.close() class _A ( __UpperCAmelCase ): UpperCamelCase__ : Dict = [B'''\x1F\x8B'''] @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' with gzip.open(__SCREAMING_SNAKE_CASE , '''rb''') as gzip_file: with open(__SCREAMING_SNAKE_CASE , '''wb''') as extracted_file: shutil.copyfileobj(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) class _A ( __UpperCAmelCase ): UpperCamelCase__ : Dict = [ B'''PK\x03\x04''', B'''PK\x05\x06''', # empty archive B'''PK\x07\x08''', # spanned archive ] @classmethod def _lowerCamelCase ( cls : Tuple , __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : bytes = b""): '''simple docstring''' if super().is_extractable(__SCREAMING_SNAKE_CASE , magic_number=__SCREAMING_SNAKE_CASE): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__SCREAMING_SNAKE_CASE , '''rb''') as fp: __a = _EndRecData(__SCREAMING_SNAKE_CASE) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET]) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: __a = fp.read(__SCREAMING_SNAKE_CASE) # CD is where we expect it to be if len(__SCREAMING_SNAKE_CASE) == sizeCentralDir: __a = struct.unpack(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE) with zipfile.ZipFile(__SCREAMING_SNAKE_CASE , '''r''') as zip_file: zip_file.extractall(__SCREAMING_SNAKE_CASE) zip_file.close() class _A ( __UpperCAmelCase ): UpperCamelCase__ : Optional[Any] = [B'''\xFD\x37\x7A\x58\x5A\x00'''] @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' with lzma.open(__SCREAMING_SNAKE_CASE) as compressed_file: with open(__SCREAMING_SNAKE_CASE , '''wb''') as extracted_file: shutil.copyfileobj(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) class _A ( __UpperCAmelCase ): UpperCamelCase__ : int = [B'''Rar!\x1a\x07\x00''', B'''Rar!\x1a\x07\x01\x00'''] # RAR_ID # RAR5_ID @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' if not config.RARFILE_AVAILABLE: raise ImportError('''Please pip install rarfile''') import rarfile os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE) __a = rarfile.RarFile(__SCREAMING_SNAKE_CASE) rf.extractall(__SCREAMING_SNAKE_CASE) rf.close() class _A ( __UpperCAmelCase ): UpperCamelCase__ : List[Any] = [B'''\x28\xb5\x2F\xFD'''] @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' if not config.ZSTANDARD_AVAILABLE: raise ImportError('''Please pip install zstandard''') import zstandard as zstd __a = zstd.ZstdDecompressor() with open(__SCREAMING_SNAKE_CASE , '''rb''') as ifh, open(__SCREAMING_SNAKE_CASE , '''wb''') as ofh: dctx.copy_stream(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) class _A ( __UpperCAmelCase ): UpperCamelCase__ : str = [B'''\x42\x5A\x68'''] @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' with bza.open(__SCREAMING_SNAKE_CASE , '''rb''') as compressed_file: with open(__SCREAMING_SNAKE_CASE , '''wb''') as extracted_file: shutil.copyfileobj(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) class _A ( __UpperCAmelCase ): UpperCamelCase__ : Union[str, Any] = [B'''\x37\x7A\xBC\xAF\x27\x1C'''] @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' if not config.PY7ZR_AVAILABLE: raise ImportError('''Please pip install py7zr''') import pyazr os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE) with pyazr.SevenZipFile(__SCREAMING_SNAKE_CASE , '''r''') as archive: archive.extractall(__SCREAMING_SNAKE_CASE) class _A ( __UpperCAmelCase ): UpperCamelCase__ : int = [B'''\x04\x22\x4D\x18'''] @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str]): '''simple docstring''' if not config.LZ4_AVAILABLE: raise ImportError('''Please pip install lz4''') import lza.frame with lza.frame.open(__SCREAMING_SNAKE_CASE , '''rb''') as compressed_file: with open(__SCREAMING_SNAKE_CASE , '''wb''') as extracted_file: shutil.copyfileobj(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) class _A : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) UpperCamelCase__ : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def _lowerCamelCase ( cls : int): '''simple docstring''' return max( len(__SCREAMING_SNAKE_CASE) for extractor in cls.extractors.values() if issubclass(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) for extractor_magic_number in extractor.magic_numbers) @staticmethod def _lowerCamelCase ( __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : int): '''simple docstring''' try: return MagicNumberBaseExtractor.read_magic_number(__SCREAMING_SNAKE_CASE , magic_number_length=__SCREAMING_SNAKE_CASE) except OSError: return b"" @classmethod def _lowerCamelCase ( cls : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : bool = False): '''simple docstring''' warnings.warn( '''Method \'is_extractable\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ''' '''Use \'infer_extractor_format\' instead.''' , category=__SCREAMING_SNAKE_CASE , ) __a = cls.infer_extractor_format(__SCREAMING_SNAKE_CASE) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def _lowerCamelCase ( cls : Any , __SCREAMING_SNAKE_CASE : Union[Path, str]): # <Added version="2.4.0"/> '''simple docstring''' __a = cls._get_magic_number_max_length() __a = cls._read_magic_number(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__SCREAMING_SNAKE_CASE , magic_number=__SCREAMING_SNAKE_CASE): return extractor_format @classmethod def _lowerCamelCase ( cls : List[str] , __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Union[Path, str] , __SCREAMING_SNAKE_CASE : Optional[str] = None , __SCREAMING_SNAKE_CASE : Optional[BaseExtractor] = "deprecated" , ): '''simple docstring''' os.makedirs(os.path.dirname(__SCREAMING_SNAKE_CASE) , exist_ok=__SCREAMING_SNAKE_CASE) # Prevent parallel extractions __a = str(Path(__SCREAMING_SNAKE_CASE).with_suffix('''.lock''')) with FileLock(__SCREAMING_SNAKE_CASE): shutil.rmtree(__SCREAMING_SNAKE_CASE , ignore_errors=__SCREAMING_SNAKE_CASE) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): # passed as positional arg warnings.warn( '''Parameter \'extractor\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ''' '''Use \'extractor_format\' instead.''' , category=__SCREAMING_SNAKE_CASE , ) __a = extractor if extractor != '''deprecated''' else extractor_format else: __a = cls.extractors[extractor_format] return extractor.extract(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) else: warnings.warn( '''Parameter \'extractor_format\' was made required in version 2.4.0 and not passing it will raise an ''' '''exception in 3.0.0.''' , category=__SCREAMING_SNAKE_CASE , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__SCREAMING_SNAKE_CASE): return extractor.extract(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE)
49
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __snake_case :Optional[int] = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __snake_case :List[str] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __snake_case :List[Any] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = len([g for position, g in enumerate(_UpperCAmelCase ) if g == main_target[position]] ) return (item, float(_UpperCAmelCase )) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = random.randint(0 , len(_UpperCAmelCase ) - 1 ) __a = parent_a[:random_slice] + parent_a[random_slice:] __a = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = list(_UpperCAmelCase ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __a = random.choice(_UpperCAmelCase ) return "".join(_UpperCAmelCase ) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): __a = [] # Generate more children proportionally to the fitness score. __a = int(parent_a[1] * 100 ) + 1 __a = 10 if child_n >= 10 else child_n for _ in range(_UpperCAmelCase ): __a = population_score[random.randint(0 , _UpperCAmelCase )][0] __a , __a = crossover(parent_a[0] , _UpperCAmelCase ) # Append new string to the population list. pop.append(mutate(_UpperCAmelCase , _UpperCAmelCase ) ) pop.append(mutate(_UpperCAmelCase , _UpperCAmelCase ) ) return pop def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: __a = f'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(_UpperCAmelCase ) # Verify that the target contains no genes besides the ones inside genes variable. __a = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __a = f'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(_UpperCAmelCase ) # Generate random starting population. __a = [] for _ in range(_UpperCAmelCase ): population.append(''''''.join([random.choice(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) )] ) ) # Just some logs to know what the algorithms is doing. __a , __a = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(_UpperCAmelCase ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __a = [evaluate(_UpperCAmelCase , _UpperCAmelCase ) for item in population] # Check if there is a matching evolution. __a = sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : x[1] , reverse=_UpperCAmelCase ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f'\nGeneration: {generation}' f'\nTotal Population:{total_population}' f'\nBest score: {population_score[0][1]}' f'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __a = population[: int(N_POPULATION / 3 )] population.clear() population.extend(_UpperCAmelCase ) # Normalize population score to be between 0 and 1. __a = [ (item, score / len(_UpperCAmelCase )) for item, score in population_score ] # This is selection for i in range(_UpperCAmelCase ): population.extend(select(population_score[int(_UpperCAmelCase )] , _UpperCAmelCase , _UpperCAmelCase ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(_UpperCAmelCase ) > N_POPULATION: break if __name__ == "__main__": __snake_case :Optional[int] = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __snake_case :List[Any] = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __snake_case ,__snake_case ,__snake_case :Dict = basic(target_str, genes_list) print( f'\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}' )
49
1
lowercase_ = 8.314_4598 def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example lowercase_ = 3_0_0 lowercase_ = 2_8 lowercase_ = rms_speed_of_molecule(temperature, molar_mass) print(F"""Vrms of Nitrogen gas at 300 K is {vrms} m/s""")
367
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase_ = {'configuration_mbart': ['MBART_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MBartConfig', 'MBartOnnxConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ['MBartTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ['MBartTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ 'MBART_PRETRAINED_MODEL_ARCHIVE_LIST', 'MBartForCausalLM', 'MBartForConditionalGeneration', 'MBartForQuestionAnswering', 'MBartForSequenceClassification', 'MBartModel', 'MBartPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ 'TFMBartForConditionalGeneration', 'TFMBartModel', 'TFMBartPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ 'FlaxMBartForConditionalGeneration', 'FlaxMBartForQuestionAnswering', 'FlaxMBartForSequenceClassification', 'FlaxMBartModel', 'FlaxMBartPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mbart import MBART_PRETRAINED_CONFIG_ARCHIVE_MAP, MBartConfig, MBartOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart import MBartTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart_fast import MBartTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mbart import ( MBART_PRETRAINED_MODEL_ARCHIVE_LIST, MBartForCausalLM, MBartForConditionalGeneration, MBartForQuestionAnswering, MBartForSequenceClassification, MBartModel, MBartPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mbart import TFMBartForConditionalGeneration, TFMBartModel, TFMBartPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mbart import ( FlaxMBartForConditionalGeneration, FlaxMBartForQuestionAnswering, FlaxMBartForSequenceClassification, FlaxMBartModel, FlaxMBartPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
194
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType lowercase_ = logging.get_logger(__name__) lowercase_ = { "microsoft/deberta-v2-xlarge": "https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json", "microsoft/deberta-v2-xxlarge": "https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json", "microsoft/deberta-v2-xlarge-mnli": ( "https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json" ), "microsoft/deberta-v2-xxlarge-mnli": ( "https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json" ), } class A ( _UpperCAmelCase ): """simple docstring""" lowerCamelCase = 'deberta-v2' def __init__( self : int,lowercase_ : List[str]=1_2_8_1_0_0,lowercase_ : Union[str, Any]=1_5_3_6,lowercase_ : Any=2_4,lowercase_ : Optional[int]=2_4,lowercase_ : Tuple=6_1_4_4,lowercase_ : Dict="gelu",lowercase_ : str=0.1,lowercase_ : List[Any]=0.1,lowercase_ : int=5_1_2,lowercase_ : Any=0,lowercase_ : Optional[int]=0.02,lowercase_ : List[str]=1E-7,lowercase_ : int=False,lowercase_ : int=-1,lowercase_ : str=0,lowercase_ : Tuple=True,lowercase_ : Dict=None,lowercase_ : int=0,lowercase_ : Tuple="gelu",**lowercase_ : List[Any],)-> Union[str, Any]: '''simple docstring''' super().__init__(**lowercase_ ) A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = initializer_range A__ = relative_attention A__ = max_relative_positions A__ = pad_token_id A__ = position_biased_input # Backwards compatibility if type(lowercase_ ) == str: A__ = [x.strip() for x in pos_att_type.lower().split('|' )] A__ = pos_att_type A__ = vocab_size A__ = layer_norm_eps A__ = kwargs.get('pooler_hidden_size',lowercase_ ) A__ = pooler_dropout A__ = pooler_hidden_act class A ( _UpperCAmelCase ): """simple docstring""" @property def snake_case__ ( self : int )-> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": A__ = {0: 'batch', 1: 'choice', 2: 'sequence'} else: A__ = {0: 'batch', 1: 'sequence'} if self._config.type_vocab_size > 0: return OrderedDict( [('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis)] ) else: return OrderedDict([('input_ids', dynamic_axis), ('attention_mask', dynamic_axis)] ) @property def snake_case__ ( self : List[Any] )-> int: '''simple docstring''' return 1_2 def snake_case__ ( self : Dict,lowercase_ : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"],lowercase_ : int = -1,lowercase_ : int = -1,lowercase_ : int = -1,lowercase_ : bool = False,lowercase_ : Optional["TensorType"] = None,lowercase_ : int = 3,lowercase_ : int = 4_0,lowercase_ : int = 4_0,lowercase_ : "PreTrainedTokenizerBase" = None,)-> Mapping[str, Any]: '''simple docstring''' A__ = super().generate_dummy_inputs(preprocessor=lowercase_,framework=lowercase_ ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
7
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> float: '''simple docstring''' return math.sqrt(sum(pow(a - b, 2 ) for a, b in zip(_UpperCAmelCase, _UpperCAmelCase ) ) ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> list[list[list[float] | float]]: '''simple docstring''' if dataset.ndim != value_array.ndim: lowerCAmelCase : List[Any] = ( 'Wrong input data\'s dimensions... ' f"dataset : {dataset.ndim}, value_array : {value_array.ndim}" ) raise ValueError(_UpperCAmelCase ) try: if dataset.shape[1] != value_array.shape[1]: lowerCAmelCase : Dict = ( 'Wrong input data\'s shape... ' f"dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}" ) raise ValueError(_UpperCAmelCase ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError('Wrong shape' ) if dataset.dtype != value_array.dtype: lowerCAmelCase : Any = ( 'Input data have different datatype... ' f"dataset : {dataset.dtype}, value_array : {value_array.dtype}" ) raise TypeError(_UpperCAmelCase ) lowerCAmelCase : int = [] for value in value_array: lowerCAmelCase : Tuple = euclidean(_UpperCAmelCase, dataset[0] ) lowerCAmelCase : Tuple = dataset[0].tolist() for dataset_value in dataset[1:]: lowerCAmelCase : Dict = euclidean(_UpperCAmelCase, _UpperCAmelCase ) if dist > temp_dist: lowerCAmelCase : Tuple = temp_dist lowerCAmelCase : Tuple = dataset_value.tolist() answer.append([vector, dist] ) return answer def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> float: '''simple docstring''' return np.dot(_UpperCAmelCase, _UpperCAmelCase ) / (norm(_UpperCAmelCase ) * norm(_UpperCAmelCase )) if __name__ == "__main__": import doctest doctest.testmod()
138
0
'''simple docstring''' import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase_ ( A , unittest.TestCase ): """simple docstring""" lowerCamelCase_ = CodeGenTokenizer lowerCamelCase_ = CodeGenTokenizerFast lowerCamelCase_ = True lowerCamelCase_ = {'''add_prefix_space''': True} lowerCamelCase_ = False def lowerCAmelCase_ ( self : Tuple ): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _SCREAMING_SNAKE_CASE = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", "<|endoftext|>", ] _SCREAMING_SNAKE_CASE = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) _SCREAMING_SNAKE_CASE = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] _SCREAMING_SNAKE_CASE = {"unk_token": "<unk>"} _SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(__lowerCamelCase ) ) def lowerCAmelCase_ ( self : Any , **__lowerCamelCase : Dict ): """simple docstring""" kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def lowerCAmelCase_ ( self : Any , **__lowerCamelCase : Union[str, Any] ): """simple docstring""" kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def lowerCAmelCase_ ( self : Tuple , __lowerCamelCase : List[str] ): """simple docstring""" _SCREAMING_SNAKE_CASE = "lower newer" _SCREAMING_SNAKE_CASE = "lower newer" return input_text, output_text def lowerCAmelCase_ ( self : Tuple ): """simple docstring""" _SCREAMING_SNAKE_CASE = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _SCREAMING_SNAKE_CASE = "lower newer" _SCREAMING_SNAKE_CASE = ["\u0120low", "er", "\u0120", "n", "e", "w", "er"] _SCREAMING_SNAKE_CASE = tokenizer.tokenize(__lowerCamelCase , add_prefix_space=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) _SCREAMING_SNAKE_CASE = tokens + [tokenizer.unk_token] _SCREAMING_SNAKE_CASE = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , __lowerCamelCase ) def lowerCAmelCase_ ( self : List[str] ): """simple docstring""" if not self.test_rust_tokenizer: return _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = self.get_rust_tokenizer(add_prefix_space=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = "lower newer" # Testing tokenization _SCREAMING_SNAKE_CASE = tokenizer.tokenize(__lowerCamelCase , add_prefix_space=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = rust_tokenizer.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) # Testing conversion to ids without special tokens _SCREAMING_SNAKE_CASE = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = rust_tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) # Testing conversion to ids with special tokens _SCREAMING_SNAKE_CASE = self.get_rust_tokenizer(add_prefix_space=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = tokenizer.encode(__lowerCamelCase , add_prefix_space=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = rust_tokenizer.encode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) # Testing the unknown token _SCREAMING_SNAKE_CASE = tokens + [rust_tokenizer.unk_token] _SCREAMING_SNAKE_CASE = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , __lowerCamelCase ) def lowerCAmelCase_ ( self : List[str] , *__lowerCamelCase : Any , **__lowerCamelCase : Dict ): """simple docstring""" # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def lowerCAmelCase_ ( self : str , __lowerCamelCase : Dict=1_5 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _SCREAMING_SNAKE_CASE = self.rust_tokenizer_class.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) # Simple input _SCREAMING_SNAKE_CASE = "This is a simple input" _SCREAMING_SNAKE_CASE = ["This is a simple input 1", "This is a simple input 2"] _SCREAMING_SNAKE_CASE = ("This is a simple input", "This is a pair") _SCREAMING_SNAKE_CASE = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(__lowerCamelCase , tokenizer_r.encode , __lowerCamelCase , max_length=__lowerCamelCase , padding="max_length" ) # Simple input self.assertRaises(__lowerCamelCase , tokenizer_r.encode_plus , __lowerCamelCase , max_length=__lowerCamelCase , padding="max_length" ) # Simple input self.assertRaises( __lowerCamelCase , tokenizer_r.batch_encode_plus , __lowerCamelCase , max_length=__lowerCamelCase , padding="max_length" , ) # Pair input self.assertRaises(__lowerCamelCase , tokenizer_r.encode , __lowerCamelCase , max_length=__lowerCamelCase , padding="max_length" ) # Pair input self.assertRaises(__lowerCamelCase , tokenizer_r.encode_plus , __lowerCamelCase , max_length=__lowerCamelCase , padding="max_length" ) # Pair input self.assertRaises( __lowerCamelCase , tokenizer_r.batch_encode_plus , __lowerCamelCase , max_length=__lowerCamelCase , padding="max_length" , ) def lowerCAmelCase_ ( self : Union[str, Any] ): """simple docstring""" _SCREAMING_SNAKE_CASE = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token="<pad>" ) # Simple input _SCREAMING_SNAKE_CASE = "This is a simple input" _SCREAMING_SNAKE_CASE = ["This is a simple input looooooooong", "This is a simple input"] _SCREAMING_SNAKE_CASE = ("This is a simple input", "This is a pair") _SCREAMING_SNAKE_CASE = [ ("This is a simple input loooooong", "This is a simple input"), ("This is a simple pair loooooong", "This is a simple pair"), ] _SCREAMING_SNAKE_CASE = tokenizer.pad_token_id _SCREAMING_SNAKE_CASE = tokenizer(__lowerCamelCase , padding="max_length" , max_length=3_0 , return_tensors="np" ) _SCREAMING_SNAKE_CASE = tokenizer(__lowerCamelCase , padding=__lowerCamelCase , truncate=__lowerCamelCase , return_tensors="np" ) _SCREAMING_SNAKE_CASE = tokenizer(*__lowerCamelCase , padding="max_length" , max_length=6_0 , return_tensors="np" ) _SCREAMING_SNAKE_CASE = tokenizer(__lowerCamelCase , padding=__lowerCamelCase , truncate=__lowerCamelCase , return_tensors="np" ) # s # test single string max_length padding self.assertEqual(out_s["input_ids"].shape[-1] , 3_0 ) self.assertTrue(pad_token_id in out_s["input_ids"] ) self.assertTrue(0 in out_s["attention_mask"] ) # s2 # test automatic padding self.assertEqual(out_sa["input_ids"].shape[-1] , 3_3 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["input_ids"][0] ) self.assertFalse(0 in out_sa["attention_mask"][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["input_ids"][1] ) self.assertTrue(0 in out_sa["attention_mask"][1] ) # p # test single pair max_length padding self.assertEqual(out_p["input_ids"].shape[-1] , 6_0 ) self.assertTrue(pad_token_id in out_p["input_ids"] ) self.assertTrue(0 in out_p["attention_mask"] ) # p2 # test automatic padding pair self.assertEqual(out_pa["input_ids"].shape[-1] , 5_2 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["input_ids"][0] ) self.assertFalse(0 in out_pa["attention_mask"][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["input_ids"][1] ) self.assertTrue(0 in out_pa["attention_mask"][1] ) def lowerCAmelCase_ ( self : int ): """simple docstring""" _SCREAMING_SNAKE_CASE = "$$$" _SCREAMING_SNAKE_CASE = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=__lowerCamelCase , add_bos_token=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = "This is a simple input" _SCREAMING_SNAKE_CASE = ["This is a simple input 1", "This is a simple input 2"] _SCREAMING_SNAKE_CASE = tokenizer.bos_token_id _SCREAMING_SNAKE_CASE = tokenizer(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = tokenizer(__lowerCamelCase ) self.assertEqual(out_s.input_ids[0] , __lowerCamelCase ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _SCREAMING_SNAKE_CASE = tokenizer.decode(out_s.input_ids ) _SCREAMING_SNAKE_CASE = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , __lowerCamelCase ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def lowerCAmelCase_ ( self : Union[str, Any] ): """simple docstring""" _SCREAMING_SNAKE_CASE = CodeGenTokenizer.from_pretrained("Salesforce/codegen-350M-mono" ) _SCREAMING_SNAKE_CASE = "\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#" _SCREAMING_SNAKE_CASE = "\nif len_a > len_b: result = a\nelse: result = b" _SCREAMING_SNAKE_CASE = tokenizer.encode(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = ["^#", re.escape("<|endoftext|>" ), "^'''", "^\"\"\"", "\n\n\n"] _SCREAMING_SNAKE_CASE = tokenizer.decode(__lowerCamelCase , truncate_before_pattern=__lowerCamelCase ) self.assertEqual(__lowerCamelCase , __lowerCamelCase ) def lowerCAmelCase_ ( self : List[str] ): """simple docstring""" pass
111
'''simple docstring''' import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel lowerCamelCase_ = { 'gwf-440k': { 'url': 'https://model-server.zqevans2.workers.dev/gwf-440k.ckpt', 'sample_rate': 4_80_00, 'sample_size': 6_55_36, }, 'jmann-small-190k': { 'url': 'https://model-server.zqevans2.workers.dev/jmann-small-190k.ckpt', 'sample_rate': 4_80_00, 'sample_size': 6_55_36, }, 'jmann-large-580k': { 'url': 'https://model-server.zqevans2.workers.dev/jmann-large-580k.ckpt', 'sample_rate': 4_80_00, 'sample_size': 13_10_72, }, 'maestro-uncond-150k': { 'url': 'https://model-server.zqevans2.workers.dev/maestro-uncond-150k.ckpt', 'sample_rate': 1_60_00, 'sample_size': 6_55_36, }, 'unlocked-uncond-250k': { 'url': 'https://model-server.zqevans2.workers.dev/unlocked-uncond-250k.ckpt', 'sample_rate': 1_60_00, 'sample_size': 6_55_36, }, 'honk-140k': { 'url': 'https://model-server.zqevans2.workers.dev/honk-140k.ckpt', 'sample_rate': 1_60_00, 'sample_size': 6_55_36, }, } def SCREAMING_SNAKE_CASE_ ( __A : List[Any] , __A : List[Any] ) -> Tuple: return torch.atana(__A , __A ) / math.pi * 2 def SCREAMING_SNAKE_CASE_ ( __A : str ) -> Tuple: _SCREAMING_SNAKE_CASE = torch.sin(t * math.pi / 2 ) ** 2 _SCREAMING_SNAKE_CASE = (1 - sigma**2) ** 0.5 return alpha_sigma_to_t(__A , __A ) class lowercase_ ( A ): """simple docstring""" pass class lowercase_ ( nn.Module ): """simple docstring""" def __init__( self : int , __lowerCamelCase : Dict ): """simple docstring""" super().__init__() _SCREAMING_SNAKE_CASE = DiffusionAttnUnetaD(__lowerCamelCase , n_attn_layers=4 ) _SCREAMING_SNAKE_CASE = deepcopy(self.diffusion ) _SCREAMING_SNAKE_CASE = torch.quasirandom.SobolEngine(1 , scramble=__lowerCamelCase ) def SCREAMING_SNAKE_CASE_ ( __A : Any ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]["url"] os.system(f"""wget {url} ./""" ) return f"""./{model_name}.ckpt""" lowerCamelCase_ = { '1': 'resnets.0', '2': 'attentions.0', '3': 'resnets.1', '4': 'attentions.1', '5': 'resnets.2', '6': 'attentions.2', } lowerCamelCase_ = { '8': 'resnets.0', '9': 'attentions.0', '10': 'resnets.1', '11': 'attentions.1', '12': 'resnets.2', '13': 'attentions.2', } lowerCamelCase_ = { '1': 'resnets.0', '2': 'attentions.0', '3': 'resnets.1', '4': 'attentions.1', '5': 'resnets.2', '6': 'attentions.2', '8': 'resnets.3', '9': 'attentions.3', '10': 'resnets.4', '11': 'attentions.4', '12': 'resnets.5', '13': 'attentions.5', } lowerCamelCase_ = { '0': 'resnets.0', '1': 'resnets.1', '2': 'resnets.2', '4': 'resnets.0', '5': 'resnets.1', '6': 'resnets.2', } lowerCamelCase_ = { 'skip': 'conv_skip', 'main.0': 'conv_1', 'main.1': 'group_norm_1', 'main.3': 'conv_2', 'main.4': 'group_norm_2', } lowerCamelCase_ = { 'norm': 'group_norm', 'qkv_proj': ['query', 'key', 'value'], 'out_proj': ['proj_attn'], } def SCREAMING_SNAKE_CASE_ ( __A : Optional[int] ) -> Dict: if name.startswith("skip" ): return name.replace("skip" , RES_CONV_MAP["skip"] ) # name has to be of format main.{digit} if not name.startswith("main." ): raise ValueError(f"""ResConvBlock error with {name}""" ) return name.replace(name[:6] , RES_CONV_MAP[name[:6]] ) def SCREAMING_SNAKE_CASE_ ( __A : Union[str, Any] ) -> Union[str, Any]: for key, value in ATTN_MAP.items(): if name.startswith(__A ) and not isinstance(__A , __A ): return name.replace(__A , __A ) elif name.startswith(__A ): return [name.replace(__A , __A ) for v in value] raise ValueError(f"""Attn error with {name}""" ) def SCREAMING_SNAKE_CASE_ ( __A : Dict , __A : List[Any]=13 ) -> List[Any]: _SCREAMING_SNAKE_CASE = input_string if string.split("." )[0] == "timestep_embed": return string.replace("timestep_embed" , "time_proj" ) _SCREAMING_SNAKE_CASE = 0 if string.startswith("net.3." ): depth += 1 _SCREAMING_SNAKE_CASE = string[6:] elif string.startswith("net." ): _SCREAMING_SNAKE_CASE = string[4:] while string.startswith("main.7." ): depth += 1 _SCREAMING_SNAKE_CASE = string[7:] if string.startswith("main." ): _SCREAMING_SNAKE_CASE = string[5:] # mid block if string[:2].isdigit(): _SCREAMING_SNAKE_CASE = string[:2] _SCREAMING_SNAKE_CASE = string[2:] else: _SCREAMING_SNAKE_CASE = string[0] _SCREAMING_SNAKE_CASE = string[1:] if depth == max_depth: _SCREAMING_SNAKE_CASE = MID_NUM_TO_LAYER[layer_num] _SCREAMING_SNAKE_CASE = "mid_block" elif depth > 0 and int(__A ) < 7: _SCREAMING_SNAKE_CASE = DOWN_NUM_TO_LAYER[layer_num] _SCREAMING_SNAKE_CASE = f"""down_blocks.{depth}""" elif depth > 0 and int(__A ) > 7: _SCREAMING_SNAKE_CASE = UP_NUM_TO_LAYER[layer_num] _SCREAMING_SNAKE_CASE = f"""up_blocks.{max_depth - depth - 1}""" elif depth == 0: _SCREAMING_SNAKE_CASE = DEPTH_0_TO_LAYER[layer_num] _SCREAMING_SNAKE_CASE = f"""up_blocks.{max_depth - 1}""" if int(__A ) > 3 else "down_blocks.0" if not string_left.startswith("." ): raise ValueError(f"""Naming error with {input_string} and string_left: {string_left}.""" ) _SCREAMING_SNAKE_CASE = string_left[1:] if "resnets" in new_layer: _SCREAMING_SNAKE_CASE = convert_resconv_naming(__A ) elif "attentions" in new_layer: _SCREAMING_SNAKE_CASE = convert_attn_naming(__A ) _SCREAMING_SNAKE_CASE = new_string_left if not isinstance(__A , __A ): _SCREAMING_SNAKE_CASE = prefix + "." + new_layer + "." + string_left else: _SCREAMING_SNAKE_CASE = [prefix + "." + new_layer + "." + s for s in string_left] return new_string def SCREAMING_SNAKE_CASE_ ( __A : List[Any] ) -> int: _SCREAMING_SNAKE_CASE = {} for k, v in state_dict.items(): if k.endswith("kernel" ): # up- and downsample layers, don't have trainable weights continue _SCREAMING_SNAKE_CASE = rename(__A ) # check if we need to transform from Conv => Linear for attention if isinstance(__A , __A ): _SCREAMING_SNAKE_CASE = transform_conv_attns(__A , __A , __A ) else: _SCREAMING_SNAKE_CASE = v return new_state_dict def SCREAMING_SNAKE_CASE_ ( __A : Tuple , __A : Dict , __A : Tuple ) -> Optional[int]: if len(__A ) == 1: if len(v.shape ) == 3: # weight _SCREAMING_SNAKE_CASE = v[:, :, 0] else: # bias _SCREAMING_SNAKE_CASE = v else: # qkv matrices _SCREAMING_SNAKE_CASE = v.shape[0] _SCREAMING_SNAKE_CASE = trippled_shape // 3 for i in range(3 ): if len(v.shape ) == 3: _SCREAMING_SNAKE_CASE = v[i * single_shape : (i + 1) * single_shape, :, 0] else: _SCREAMING_SNAKE_CASE = v[i * single_shape : (i + 1) * single_shape] return new_state_dict def SCREAMING_SNAKE_CASE_ ( __A : str ) -> Optional[int]: _SCREAMING_SNAKE_CASE = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) _SCREAMING_SNAKE_CASE = args.model_path.split("/" )[-1].split("." )[0] if not os.path.isfile(args.model_path ): assert ( model_name == args.model_path ), f"""Make sure to provide one of the official model names {MODELS_MAP.keys()}""" _SCREAMING_SNAKE_CASE = download(__A ) _SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]["sample_rate"] _SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]["sample_size"] _SCREAMING_SNAKE_CASE = Object() _SCREAMING_SNAKE_CASE = sample_size _SCREAMING_SNAKE_CASE = sample_rate _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = UNetaDModel(sample_size=__A , sample_rate=__A ) _SCREAMING_SNAKE_CASE = diffusers_model.state_dict() _SCREAMING_SNAKE_CASE = DiffusionUncond(__A ) orig_model.load_state_dict(torch.load(args.model_path , map_location=__A )["state_dict"] ) _SCREAMING_SNAKE_CASE = orig_model.diffusion_ema.eval() _SCREAMING_SNAKE_CASE = orig_model.state_dict() _SCREAMING_SNAKE_CASE = rename_orig_weights(__A ) _SCREAMING_SNAKE_CASE = set(renamed_state_dict.keys() ) - set(diffusers_state_dict.keys() ) _SCREAMING_SNAKE_CASE = set(diffusers_state_dict.keys() ) - set(renamed_state_dict.keys() ) assert len(__A ) == 0, f"""Problem with {renamed_minus_diffusers}""" assert all(k.endswith("kernel" ) for k in list(__A ) ), f"""Problem with {diffusers_minus_renamed}""" for key, value in renamed_state_dict.items(): assert ( diffusers_state_dict[key].squeeze().shape == value.squeeze().shape ), f"""Shape for {key} doesn't match. Diffusers: {diffusers_state_dict[key].shape} vs. {value.shape}""" if key == "time_proj.weight": _SCREAMING_SNAKE_CASE = value.squeeze() _SCREAMING_SNAKE_CASE = value diffusers_model.load_state_dict(__A ) _SCREAMING_SNAKE_CASE = 1_00 _SCREAMING_SNAKE_CASE = 33 _SCREAMING_SNAKE_CASE = IPNDMScheduler(num_train_timesteps=__A ) _SCREAMING_SNAKE_CASE = torch.manual_seed(__A ) _SCREAMING_SNAKE_CASE = torch.randn([1, 2, config.sample_size] , generator=__A ).to(__A ) _SCREAMING_SNAKE_CASE = torch.linspace(1 , 0 , steps + 1 , device=__A )[:-1] _SCREAMING_SNAKE_CASE = get_crash_schedule(__A ) _SCREAMING_SNAKE_CASE = DanceDiffusionPipeline(unet=__A , scheduler=__A ) _SCREAMING_SNAKE_CASE = torch.manual_seed(33 ) _SCREAMING_SNAKE_CASE = pipe(num_inference_steps=__A , generator=__A ).audios _SCREAMING_SNAKE_CASE = sampling.iplms_sample(__A , __A , __A , {} ) _SCREAMING_SNAKE_CASE = generated.clamp(-1 , 1 ) _SCREAMING_SNAKE_CASE = (generated - audio).abs().sum() _SCREAMING_SNAKE_CASE = (generated - audio).abs().max() if args.save: pipe.save_pretrained(args.checkpoint_path ) print("Diff sum" , __A ) print("Diff max" , __A ) assert diff_max < 1e-3, f"""Diff max: {diff_max} is too much :-/""" print(f"""Conversion for {model_name} successful!""" ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument('--model_path', default=None, type=str, required=True, help='Path to the model to convert.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument('--checkpoint_path', default=None, type=str, required=True, help='Path to the output model.') lowerCamelCase_ = parser.parse_args() main(args)
111
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase : Optional[int] = { "configuration_jukebox": [ "JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP", "JukeboxConfig", "JukeboxPriorConfig", "JukeboxVQVAEConfig", ], "tokenization_jukebox": ["JukeboxTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : int = [ "JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST", "JukeboxModel", "JukeboxPreTrainedModel", "JukeboxVQVAE", "JukeboxPrior", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys _lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
28
import os import unittest from transformers import BatchEncoding from transformers.models.bert.tokenization_bert import ( BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.models.prophetnet.tokenization_prophetnet import VOCAB_FILES_NAMES, ProphetNetTokenizer from transformers.testing_utils import require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin class lowercase_ ( lowercase , unittest.TestCase ): '''simple docstring''' __snake_case = ProphetNetTokenizer __snake_case = False def __lowerCAmelCase ( self : Optional[int] ) ->Optional[Any]: """simple docstring""" super().setUp() a = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __lowerCAmelCase ( self : List[str] , __UpperCAmelCase : str ) ->Dict: """simple docstring""" a = '''UNwant\u00E9d,running''' a = '''unwanted, running''' return input_text, output_text def __lowerCAmelCase ( self : Optional[int] ) ->Optional[Any]: """simple docstring""" a = self.tokenizer_class(self.vocab_file ) a = tokenizer.tokenize('''UNwant\u00E9d,running''' ) self.assertListEqual(__UpperCAmelCase , ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [9, 6, 7, 12, 10, 11] ) def __lowerCAmelCase ( self : int ) ->Any: """simple docstring""" a = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''' ) , ['''ah''', '''\u535A''', '''\u63A8''', '''zz'''] ) def __lowerCAmelCase ( self : Any ) ->int: """simple docstring""" a = BasicTokenizer(do_lower_case=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''hello''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def __lowerCAmelCase ( self : Union[str, Any] ) ->Optional[int]: """simple docstring""" a = BasicTokenizer(do_lower_case=__UpperCAmelCase , strip_accents=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''h\u00E9llo'''] ) def __lowerCAmelCase ( self : Dict ) ->str: """simple docstring""" a = BasicTokenizer(do_lower_case=__UpperCAmelCase , strip_accents=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def __lowerCAmelCase ( self : Any ) ->Dict: """simple docstring""" a = BasicTokenizer(do_lower_case=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def __lowerCAmelCase ( self : Tuple ) ->Optional[Any]: """simple docstring""" a = BasicTokenizer(do_lower_case=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def __lowerCAmelCase ( self : Tuple ) ->Tuple: """simple docstring""" a = BasicTokenizer(do_lower_case=__UpperCAmelCase , strip_accents=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def __lowerCAmelCase ( self : int ) ->Optional[int]: """simple docstring""" a = BasicTokenizer(do_lower_case=__UpperCAmelCase , strip_accents=__UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def __lowerCAmelCase ( self : Any ) ->int: """simple docstring""" a = BasicTokenizer(do_lower_case=__UpperCAmelCase , never_split=['''[UNK]'''] ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]'''] ) def __lowerCAmelCase ( self : Union[str, Any] ) ->int: """simple docstring""" a = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing'''] a = {} for i, token in enumerate(__UpperCAmelCase ): a = i a = WordpieceTokenizer(vocab=__UpperCAmelCase , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''unwanted running''' ) , ['''un''', '''##want''', '''##ed''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.tokenize('''unwantedX running''' ) , ['''[UNK]''', '''runn''', '''##ing'''] ) @require_torch def __lowerCAmelCase ( self : int ) ->int: """simple docstring""" a = self.tokenizer_class.from_pretrained('''microsoft/prophetnet-large-uncased''' ) a = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] a = [1_037, 2_146, 20_423, 2_005, 7_680, 7_849, 3_989, 1_012, 102] a = tokenizer(__UpperCAmelCase , padding=__UpperCAmelCase , return_tensors='''pt''' ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) a = list(batch.input_ids.numpy()[0] ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) def __lowerCAmelCase ( self : Optional[Any] ) ->List[str]: """simple docstring""" self.assertTrue(_is_whitespace(''' ''' ) ) self.assertTrue(_is_whitespace('''\t''' ) ) self.assertTrue(_is_whitespace('''\r''' ) ) self.assertTrue(_is_whitespace('''\n''' ) ) self.assertTrue(_is_whitespace('''\u00A0''' ) ) self.assertFalse(_is_whitespace('''A''' ) ) self.assertFalse(_is_whitespace('''-''' ) ) def __lowerCAmelCase ( self : Any ) ->List[str]: """simple docstring""" self.assertTrue(_is_control('''\u0005''' ) ) self.assertFalse(_is_control('''A''' ) ) self.assertFalse(_is_control(''' ''' ) ) self.assertFalse(_is_control('''\t''' ) ) self.assertFalse(_is_control('''\r''' ) ) def __lowerCAmelCase ( self : List[Any] ) ->List[str]: """simple docstring""" self.assertTrue(_is_punctuation('''-''' ) ) self.assertTrue(_is_punctuation('''$''' ) ) self.assertTrue(_is_punctuation('''`''' ) ) self.assertTrue(_is_punctuation('''.''' ) ) self.assertFalse(_is_punctuation('''A''' ) ) self.assertFalse(_is_punctuation(''' ''' ) ) @slow def __lowerCAmelCase ( self : List[str] ) ->List[str]: """simple docstring""" a = self.tokenizer_class.from_pretrained('''microsoft/prophetnet-large-uncased''' ) a = tokenizer.encode('''sequence builders''' , add_special_tokens=__UpperCAmelCase ) a = tokenizer.encode('''multi-sequence build''' , add_special_tokens=__UpperCAmelCase ) a = tokenizer.build_inputs_with_special_tokens(__UpperCAmelCase ) a = tokenizer.build_inputs_with_special_tokens(__UpperCAmelCase , __UpperCAmelCase ) assert encoded_sentence == text + [102] assert encoded_pair == text + [102] + text_a + [102]
0
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase : List[Any] = { '''configuration_mobilebert''': [ '''MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileBertConfig''', '''MobileBertOnnxConfig''', ], '''tokenization_mobilebert''': ['''MobileBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Tuple = ['''MobileBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Tuple = [ '''MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileBertForMaskedLM''', '''MobileBertForMultipleChoice''', '''MobileBertForNextSentencePrediction''', '''MobileBertForPreTraining''', '''MobileBertForQuestionAnswering''', '''MobileBertForSequenceClassification''', '''MobileBertForTokenClassification''', '''MobileBertLayer''', '''MobileBertModel''', '''MobileBertPreTrainedModel''', '''load_tf_weights_in_mobilebert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : int = [ '''TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFMobileBertForMaskedLM''', '''TFMobileBertForMultipleChoice''', '''TFMobileBertForNextSentencePrediction''', '''TFMobileBertForPreTraining''', '''TFMobileBertForQuestionAnswering''', '''TFMobileBertForSequenceClassification''', '''TFMobileBertForTokenClassification''', '''TFMobileBertMainLayer''', '''TFMobileBertModel''', '''TFMobileBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys _lowerCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
340
"""simple docstring""" import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _lowerCAmelCase : int = logging.get_logger(__name__) def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase=False ) -> List[str]: '''simple docstring''' _lowerCamelCase : Tuple = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith("head" ): _lowerCamelCase : Tuple = "segformer.encoder." + key if key.startswith("backbone" ): _lowerCamelCase : Any = key.replace("backbone" , "segformer.encoder" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 _lowerCamelCase : int = key[key.find("patch_embed" ) + len("patch_embed" )] _lowerCamelCase : int = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(_lowerCamelCase )-1}""" ) if "norm" in key: _lowerCamelCase : Optional[Any] = key.replace("norm" , "layer_norm" ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 _lowerCamelCase : Dict = key[key.find("segformer.encoder.layer_norm" ) + len("segformer.encoder.layer_norm" )] _lowerCamelCase : Tuple = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(_lowerCamelCase )-1}""" ) if "layer_norm1" in key: _lowerCamelCase : Union[str, Any] = key.replace("layer_norm1" , "layer_norm_1" ) if "layer_norm2" in key: _lowerCamelCase : int = key.replace("layer_norm2" , "layer_norm_2" ) if "block" in key: # replace for example block1 by block.0 _lowerCamelCase : Union[str, Any] = key[key.find("block" ) + len("block" )] _lowerCamelCase : Optional[Any] = key.replace(F"""block{idx}""" , F"""block.{int(_lowerCamelCase )-1}""" ) if "attn.q" in key: _lowerCamelCase : Optional[int] = key.replace("attn.q" , "attention.self.query" ) if "attn.proj" in key: _lowerCamelCase : List[str] = key.replace("attn.proj" , "attention.output.dense" ) if "attn" in key: _lowerCamelCase : Tuple = key.replace("attn" , "attention.self" ) if "fc1" in key: _lowerCamelCase : Optional[Any] = key.replace("fc1" , "dense1" ) if "fc2" in key: _lowerCamelCase : Dict = key.replace("fc2" , "dense2" ) if "linear_pred" in key: _lowerCamelCase : int = key.replace("linear_pred" , "classifier" ) if "linear_fuse" in key: _lowerCamelCase : str = key.replace("linear_fuse.conv" , "linear_fuse" ) _lowerCamelCase : Optional[Any] = key.replace("linear_fuse.bn" , "batch_norm" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 _lowerCamelCase : Union[str, Any] = key[key.find("linear_c" ) + len("linear_c" )] _lowerCamelCase : Optional[int] = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(_lowerCamelCase )-1}""" ) if key.startswith("head" ): _lowerCamelCase : List[str] = key.replace("head" , "classifier" ) _lowerCamelCase : Union[str, Any] = value return new_state_dict def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase ) -> Optional[Any]: '''simple docstring''' for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) _lowerCamelCase : Optional[Any] = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.weight""" ) _lowerCamelCase : Optional[Any] = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict _lowerCamelCase : int = kv_weight[ : config.hidden_sizes[i], : ] _lowerCamelCase : int = kv_bias[: config.hidden_sizes[i]] _lowerCamelCase : Optional[int] = kv_weight[ config.hidden_sizes[i] :, : ] _lowerCamelCase : Optional[Any] = kv_bias[ config.hidden_sizes[i] : ] def lowerCamelCase_( ) -> Dict: '''simple docstring''' _lowerCamelCase : int = "http://images.cocodataset.org/val2017/000000039769.jpg" _lowerCamelCase : Union[str, Any] = Image.open(requests.get(_lowerCamelCase , stream=_lowerCamelCase ).raw ) return image @torch.no_grad() def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Dict: '''simple docstring''' _lowerCamelCase : Any = SegformerConfig() _lowerCamelCase : int = False # set attributes based on model_name _lowerCamelCase : Any = "huggingface/label-files" if "segformer" in model_name: _lowerCamelCase : str = model_name[len("segformer." ) : len("segformer." ) + 2] if "ade" in model_name: _lowerCamelCase : str = 150 _lowerCamelCase : Dict = "ade20k-id2label.json" _lowerCamelCase : Dict = (1, 150, 128, 128) elif "city" in model_name: _lowerCamelCase : List[str] = 19 _lowerCamelCase : Tuple = "cityscapes-id2label.json" _lowerCamelCase : Tuple = (1, 19, 128, 128) else: raise ValueError(F"""Model {model_name} not supported""" ) elif "mit" in model_name: _lowerCamelCase : List[str] = True _lowerCamelCase : Tuple = model_name[4:6] _lowerCamelCase : Tuple = 1000 _lowerCamelCase : List[Any] = "imagenet-1k-id2label.json" _lowerCamelCase : List[Any] = (1, 1000) else: raise ValueError(F"""Model {model_name} not supported""" ) # set config attributes _lowerCamelCase : Optional[Any] = json.load(open(hf_hub_download(_lowerCamelCase , _lowerCamelCase , repo_type="dataset" ) , "r" ) ) _lowerCamelCase : List[str] = {int(_lowerCamelCase ): v for k, v in idalabel.items()} _lowerCamelCase : Optional[Any] = idalabel _lowerCamelCase : Union[str, Any] = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": _lowerCamelCase : int = [64, 128, 320, 512] _lowerCamelCase : int = 256 elif size == "b2": _lowerCamelCase : Tuple = [64, 128, 320, 512] _lowerCamelCase : List[Any] = 768 _lowerCamelCase : Any = [3, 4, 6, 3] elif size == "b3": _lowerCamelCase : Tuple = [64, 128, 320, 512] _lowerCamelCase : Union[str, Any] = 768 _lowerCamelCase : Optional[Any] = [3, 4, 18, 3] elif size == "b4": _lowerCamelCase : str = [64, 128, 320, 512] _lowerCamelCase : Optional[Any] = 768 _lowerCamelCase : Dict = [3, 8, 27, 3] elif size == "b5": _lowerCamelCase : int = [64, 128, 320, 512] _lowerCamelCase : Tuple = 768 _lowerCamelCase : Tuple = [3, 6, 40, 3] else: raise ValueError(F"""Size {size} not supported""" ) # load image processor (only resize + normalize) _lowerCamelCase : Dict = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_lowerCamelCase , align=_lowerCamelCase , do_random_crop=_lowerCamelCase ) # prepare image _lowerCamelCase : List[str] = prepare_img() _lowerCamelCase : Dict = image_processor(images=_lowerCamelCase , return_tensors="pt" ).pixel_values logger.info(F"""Converting model {model_name}...""" ) # load original state dict if encoder_only: _lowerCamelCase : Tuple = torch.load(_lowerCamelCase , map_location=torch.device("cpu" ) ) else: _lowerCamelCase : int = torch.load(_lowerCamelCase , map_location=torch.device("cpu" ) )["state_dict"] # rename keys _lowerCamelCase : str = rename_keys(_lowerCamelCase , encoder_only=_lowerCamelCase ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(_lowerCamelCase , _lowerCamelCase ) # create HuggingFace model and load state dict if encoder_only: _lowerCamelCase : Tuple = False _lowerCamelCase : Optional[int] = SegformerForImageClassification(_lowerCamelCase ) else: _lowerCamelCase : List[str] = SegformerForSemanticSegmentation(_lowerCamelCase ) model.load_state_dict(_lowerCamelCase ) model.eval() # forward pass _lowerCamelCase : Any = model(_lowerCamelCase ) _lowerCamelCase : Dict = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": _lowerCamelCase : str = torch.tensor( [ [[-4.6_3_1_0, -5.5_2_3_2, -6.2_3_5_6], [-5.1_9_2_1, -6.1_4_4_4, -6.5_9_9_6], [-5.4_4_2_4, -6.2_7_9_0, -6.7_5_7_4]], [[-1_2.1_3_9_1, -1_3.3_1_2_2, -1_3.9_5_5_4], [-1_2.8_7_3_2, -1_3.9_3_5_2, -1_4.3_5_6_3], [-1_2.9_4_3_8, -1_3.8_2_2_6, -1_4.2_5_1_3]], [[-1_2.5_1_3_4, -1_3.4_6_8_6, -1_4.4_9_1_5], [-1_2.8_6_6_9, -1_4.4_3_4_3, -1_4.7_7_5_8], [-1_3.2_5_2_3, -1_4.5_8_1_9, -1_5.0_6_9_4]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": _lowerCamelCase : Any = torch.tensor( [ [[-7.5_8_2_0, -8.7_2_3_1, -8.3_2_1_5], [-8.0_6_0_0, -1_0.3_5_2_9, -1_0.0_3_0_4], [-7.5_2_0_8, -9.4_1_0_3, -9.6_2_3_9]], [[-1_2.6_9_1_8, -1_3.8_9_9_4, -1_3.7_1_3_7], [-1_3.3_1_9_6, -1_5.7_5_2_3, -1_5.4_7_8_9], [-1_2.9_3_4_3, -1_4.8_7_5_7, -1_4.9_6_8_9]], [[-1_1.1_9_1_1, -1_1.9_4_2_1, -1_1.3_2_4_3], [-1_1.3_3_4_2, -1_3.6_8_3_9, -1_3.3_5_8_1], [-1_0.3_9_0_9, -1_2.1_8_3_2, -1_2.4_8_5_8]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": _lowerCamelCase : int = torch.tensor( [ [[-1_1.8_1_7_3, -1_4.3_8_5_0, -1_6.3_1_2_8], [-1_4.5_6_4_8, -1_6.5_8_0_4, -1_8.6_5_6_8], [-1_4.7_2_2_3, -1_5.7_3_8_7, -1_8.4_2_1_8]], [[-1_5.7_2_9_0, -1_7.9_1_7_1, -1_9.4_4_2_3], [-1_8.3_1_0_5, -1_9.9_4_4_8, -2_1.4_6_6_1], [-1_7.9_2_9_6, -1_8.6_4_9_7, -2_0.7_9_1_0]], [[-1_5.0_7_8_3, -1_7.0_3_3_6, -1_8.2_7_8_9], [-1_6.8_7_7_1, -1_8.6_8_7_0, -2_0.1_6_1_2], [-1_6.2_4_5_4, -1_7.1_4_2_6, -1_9.5_0_5_5]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": _lowerCamelCase : Optional[Any] = torch.tensor( [ [[-9.0_8_7_8, -1_0.2_0_8_1, -1_0.1_8_9_1], [-9.3_1_4_4, -1_0.7_9_4_1, -1_0.9_8_4_3], [-9.2_2_9_4, -1_0.3_8_5_5, -1_0.5_7_0_4]], [[-1_2.2_3_1_6, -1_3.9_0_6_8, -1_3.6_1_0_2], [-1_2.9_1_6_1, -1_4.3_7_0_2, -1_4.3_2_3_5], [-1_2.5_2_3_3, -1_3.7_1_7_4, -1_3.7_9_3_2]], [[-1_4.6_2_7_5, -1_5.2_4_9_0, -1_4.9_7_2_7], [-1_4.3_4_0_0, -1_5.9_6_8_7, -1_6.2_8_2_7], [-1_4.1_4_8_4, -1_5.4_0_3_3, -1_5.8_9_3_7]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": _lowerCamelCase : List[str] = torch.tensor( [ [[-1_2.3_1_4_4, -1_3.2_4_4_7, -1_4.0_8_0_2], [-1_3.3_6_1_4, -1_4.5_8_1_6, -1_5.6_1_1_7], [-1_3.3_3_4_0, -1_4.4_4_3_3, -1_6.2_2_1_9]], [[-1_9.2_7_8_1, -2_0.4_1_2_8, -2_0.7_5_0_6], [-2_0.6_1_5_3, -2_1.6_5_6_6, -2_2.0_9_9_8], [-1_9.9_8_0_0, -2_1.0_4_3_0, -2_2.1_4_9_4]], [[-1_8.8_7_3_9, -1_9.7_8_0_4, -2_1.1_8_3_4], [-2_0.1_2_3_3, -2_1.6_7_6_5, -2_3.2_9_4_4], [-2_0.0_3_1_5, -2_1.2_6_4_1, -2_3.6_9_4_4]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": _lowerCamelCase : Any = torch.tensor( [ [[-9.5_5_2_4, -1_2.0_8_3_5, -1_1.7_3_4_8], [-1_0.5_2_2_9, -1_3.6_4_4_6, -1_4.5_6_6_2], [-9.5_8_4_2, -1_2.8_8_5_1, -1_3.9_4_1_4]], [[-1_5.3_4_3_2, -1_7.5_3_2_3, -1_7.0_8_1_8], [-1_6.3_3_3_0, -1_8.9_2_5_5, -1_9.2_1_0_1], [-1_5.1_3_4_0, -1_7.7_8_4_8, -1_8.3_9_7_1]], [[-1_2.6_0_7_2, -1_4.9_4_8_6, -1_4.6_6_3_1], [-1_3.7_6_2_9, -1_7.0_9_0_7, -1_7.7_7_4_5], [-1_2.7_8_9_9, -1_6.1_6_9_5, -1_7.1_6_7_1]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": _lowerCamelCase : Dict = torch.tensor( [ [[-1_1.9_2_9_5, -1_3.4_0_5_7, -1_4.8_1_0_6], [-1_3.3_4_3_1, -1_4.8_1_7_9, -1_5.3_7_8_1], [-1_4.2_8_3_6, -1_5.5_9_4_2, -1_6.1_5_8_8]], [[-1_1.4_9_0_6, -1_2.8_0_6_7, -1_3.6_5_6_4], [-1_3.1_1_8_9, -1_4.0_5_0_0, -1_4.1_5_4_3], [-1_3.8_7_4_8, -1_4.5_1_3_6, -1_4.8_7_8_9]], [[0.5_3_7_4, 0.1_0_6_7, -0.4_7_4_2], [0.1_1_4_1, -0.2_2_5_5, -0.7_0_9_9], [-0.3_0_0_0, -0.5_9_2_4, -1.3_1_0_5]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": _lowerCamelCase : Optional[int] = torch.tensor( [ [[-7.8_2_1_7, -9.8_7_6_7, -1_0.1_7_1_7], [-9.4_4_3_8, -1_0.9_0_5_8, -1_1.4_0_4_7], [-9.7_9_3_9, -1_2.3_4_9_5, -1_2.1_0_7_9]], [[-7.1_5_1_4, -9.5_3_3_6, -1_0.0_8_6_0], [-9.7_7_7_6, -1_1.6_8_2_2, -1_1.8_4_3_9], [-1_0.1_4_1_1, -1_2.7_6_5_5, -1_2.8_9_7_2]], [[0.3_0_2_1, 0.0_8_0_5, -0.2_3_1_0], [-0.0_3_2_8, -0.1_6_0_5, -0.2_7_1_4], [-0.1_4_0_8, -0.5_4_7_7, -0.6_9_7_6]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": _lowerCamelCase : Tuple = torch.tensor( [ [ [-1.13_72e01, -1.27_87e01, -1.34_77e01], [-1.25_36e01, -1.41_94e01, -1.44_09e01], [-1.32_17e01, -1.48_88e01, -1.53_27e01], ], [ [-1.47_91e01, -1.71_22e01, -1.82_77e01], [-1.71_63e01, -1.91_92e01, -1.95_33e01], [-1.78_97e01, -1.99_91e01, -2.03_15e01], ], [ [7.67_23e-01, 4.19_21e-01, -7.78_78e-02], [4.77_72e-01, 9.55_57e-03, -2.80_82e-01], [3.60_32e-01, -2.48_26e-01, -5.11_68e-01], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": _lowerCamelCase : Union[str, Any] = torch.tensor( [ [[-9.4_9_5_9, -1_1.3_0_8_7, -1_1.7_4_7_9], [-1_1.0_0_2_5, -1_2.6_5_4_0, -1_2.3_3_1_9], [-1_1.4_0_6_4, -1_3.0_4_8_7, -1_2.9_9_0_5]], [[-9.8_9_0_5, -1_1.3_0_8_4, -1_2.0_8_5_4], [-1_1.1_7_2_6, -1_2.7_6_9_8, -1_2.9_5_8_3], [-1_1.5_9_8_5, -1_3.3_2_7_8, -1_4.1_7_7_4]], [[0.2_2_1_3, 0.0_1_9_2, -0.2_4_6_6], [-0.1_7_3_1, -0.4_2_1_3, -0.4_8_7_4], [-0.3_1_2_6, -0.6_5_4_1, -1.1_3_8_9]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": _lowerCamelCase : List[Any] = torch.tensor( [ [[-1_3.5_7_4_8, -1_3.9_1_1_1, -1_2.6_5_0_0], [-1_4.3_5_0_0, -1_5.3_6_8_3, -1_4.2_3_2_8], [-1_4.7_5_3_2, -1_6.0_4_2_4, -1_5.6_0_8_7]], [[-1_7.1_6_5_1, -1_5.8_7_2_5, -1_2.9_6_5_3], [-1_7.2_5_8_0, -1_7.3_7_1_8, -1_4.8_2_2_3], [-1_6.6_0_5_8, -1_6.8_7_8_3, -1_6.7_4_5_2]], [[-3.6_4_5_6, -3.0_2_0_9, -1.4_2_0_3], [-3.0_7_9_7, -3.1_9_5_9, -2.0_0_0_0], [-1.8_7_5_7, -1.9_2_1_7, -1.6_9_9_7]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": _lowerCamelCase : Tuple = torch.tensor( [ [[-1_6.0_9_7_6, -1_6.4_8_5_6, -1_7.3_9_6_2], [-1_6.6_2_3_4, -1_9.0_3_4_2, -1_9.7_6_8_5], [-1_6.0_9_0_0, -1_8.0_6_6_1, -1_9.1_1_8_0]], [[-1_8.4_7_5_0, -1_8.8_4_8_8, -1_9.5_0_7_4], [-1_9.4_0_3_0, -2_2.1_5_7_0, -2_2.5_9_7_7], [-1_9.1_1_9_1, -2_0.8_4_8_6, -2_2.3_7_8_3]], [[-4.5_1_7_8, -5.5_0_3_7, -6.5_1_0_9], [-5.0_8_8_4, -7.2_1_7_4, -8.0_3_3_4], [-4.4_1_5_6, -5.8_1_1_7, -7.2_9_7_0]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": _lowerCamelCase : Any = torch.tensor( [ [[-1_4.2_0_8_1, -1_4.4_7_3_2, -1_4.1_9_7_7], [-1_4.5_8_6_7, -1_6.4_4_2_3, -1_6.6_3_5_6], [-1_3.4_4_4_1, -1_4.9_6_8_5, -1_6.8_6_9_6]], [[-1_4.4_5_7_6, -1_4.7_0_7_3, -1_5.0_4_5_1], [-1_5.0_8_1_6, -1_7.6_2_3_7, -1_7.9_8_7_3], [-1_4.4_2_1_3, -1_6.0_1_9_9, -1_8.5_9_9_2]], [[-4.7_3_4_9, -4.9_5_8_8, -5.0_9_6_6], [-4.3_2_1_0, -6.9_3_2_5, -7.2_5_9_1], [-3.4_3_1_2, -4.7_4_8_4, -7.1_9_1_7]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": _lowerCamelCase : List[str] = torch.tensor( [ [[-1_1.7_7_3_7, -1_1.9_5_2_6, -1_1.3_2_7_3], [-1_3.6_6_9_2, -1_4.4_5_7_4, -1_3.8_8_7_8], [-1_3.8_9_3_7, -1_4.6_9_2_4, -1_5.9_3_4_5]], [[-1_4.6_7_0_6, -1_4.5_3_3_0, -1_4.1_3_0_6], [-1_6.1_5_0_2, -1_6.8_1_8_0, -1_6.4_2_6_9], [-1_6.8_3_3_8, -1_7.8_9_3_9, -2_0.1_7_4_6]], [[1.0_4_9_1, 0.8_2_8_9, 1.0_3_1_0], [1.1_0_4_4, 0.5_2_1_9, 0.8_0_5_5], [1.0_8_9_9, 0.6_9_2_6, 0.5_5_9_0]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": _lowerCamelCase : str = torch.tensor( [ [[-1_2.5_6_4_1, -1_3.4_7_7_7, -1_3.0_6_8_4], [-1_3.9_5_8_7, -1_5.8_9_8_3, -1_6.6_5_5_7], [-1_3.3_1_0_9, -1_5.7_3_5_0, -1_6.3_1_4_1]], [[-1_4.7_0_7_4, -1_5.4_3_5_2, -1_4.5_9_4_4], [-1_6.6_3_5_3, -1_8.1_6_6_3, -1_8.6_1_2_0], [-1_5.1_7_0_2, -1_8.0_3_2_9, -1_8.1_5_4_7]], [[-1.7_9_9_0, -2.0_9_5_1, -1.7_7_8_4], [-2.6_3_9_7, -3.8_2_4_5, -3.9_6_8_6], [-1.5_2_6_4, -2.8_1_2_6, -2.9_3_1_6]], ] ) else: _lowerCamelCase : Dict = logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , _lowerCamelCase , atol=1e-2 ) # finally, save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(_lowerCamelCase ).mkdir(exist_ok=_lowerCamelCase ) model.save_pretrained(_lowerCamelCase ) image_processor.save_pretrained(_lowerCamelCase ) if __name__ == "__main__": _lowerCAmelCase : str = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''segformer.b0.512x512.ade.160k''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) _lowerCAmelCase : str = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
340
1