code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } UpperCAmelCase_ = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } UpperCAmelCase_ = {'facebook/blenderbot_small-90M': 512} def lowerCamelCase__ ( A__ : int ): '''simple docstring''' __lowerCamelCase = set() __lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowerCamelCase = char __lowerCamelCase = set(A__ ) return pairs class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : int = VOCAB_FILES_NAMES UpperCAmelCase__ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : str = ['input_ids', 'attention_mask'] def __init__( self: str , UpperCamelCase_: List[str] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: List[str]="__start__" , UpperCamelCase_: Optional[Any]="__end__" , UpperCamelCase_: Optional[Any]="__unk__" , UpperCamelCase_: Union[str, Any]="__null__" , **UpperCamelCase_: int , ): super().__init__(unk_token=UpperCamelCase_ , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , **UpperCamelCase_ ) with open(UpperCamelCase_ , encoding="""utf-8""" ) as vocab_handle: __lowerCamelCase = json.load(UpperCamelCase_ ) __lowerCamelCase = {v: k for k, v in self.encoder.items()} with open(UpperCamelCase_ , encoding="""utf-8""" ) as merges_handle: __lowerCamelCase = merges_handle.read().split("""\n""" )[1:-1] __lowerCamelCase = [tuple(merge.split() ) for merge in merges] __lowerCamelCase = dict(zip(UpperCamelCase_ , range(len(UpperCamelCase_ ) ) ) ) __lowerCamelCase = {} @property def lowerCAmelCase__ ( self: List[str] ): return len(self.encoder ) def lowerCAmelCase__ ( self: List[str] ): return dict(self.encoder , **self.added_tokens_encoder ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: str ): if token in self.cache: return self.cache[token] __lowerCamelCase = re.sub("""([.,!?()])""" , r""" \1""" , UpperCamelCase_ ) __lowerCamelCase = re.sub("""(')""" , r""" \1 """ , UpperCamelCase_ ) __lowerCamelCase = re.sub(r"""\s{2,}""" , """ """ , UpperCamelCase_ ) if "\n" in token: __lowerCamelCase = token.replace("""\n""" , """ __newln__""" ) __lowerCamelCase = token.split(""" """ ) __lowerCamelCase = [] for token in tokens: if not len(UpperCamelCase_ ): continue __lowerCamelCase = token.lower() __lowerCamelCase = tuple(UpperCamelCase_ ) __lowerCamelCase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __lowerCamelCase = get_pairs(UpperCamelCase_ ) if not pairs: words.append(UpperCamelCase_ ) continue while True: __lowerCamelCase = min(UpperCamelCase_ , key=lambda UpperCamelCase_ : self.bpe_ranks.get(UpperCamelCase_ , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __lowerCamelCase, __lowerCamelCase = bigram __lowerCamelCase = [] __lowerCamelCase = 0 while i < len(UpperCamelCase_ ): try: __lowerCamelCase = word.index(UpperCamelCase_ , UpperCamelCase_ ) new_word.extend(word[i:j] ) __lowerCamelCase = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(UpperCamelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowerCamelCase = tuple(UpperCamelCase_ ) __lowerCamelCase = new_word if len(UpperCamelCase_ ) == 1: break else: __lowerCamelCase = get_pairs(UpperCamelCase_ ) __lowerCamelCase = """@@ """.join(UpperCamelCase_ ) __lowerCamelCase = word[:-4] __lowerCamelCase = word words.append(UpperCamelCase_ ) return " ".join(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: str ): __lowerCamelCase = [] __lowerCamelCase = re.findall(r"""\S+\n?""" , UpperCamelCase_ ) for token in words: split_tokens.extend(list(self.bpe(UpperCamelCase_ ).split(""" """ ) ) ) return split_tokens def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: str ): __lowerCamelCase = token.lower() return self.encoder.get(UpperCamelCase_ , self.encoder.get(self.unk_token ) ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: int ): return self.decoder.get(UpperCamelCase_ , self.unk_token ) def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: List[str] ): __lowerCamelCase = """ """.join(UpperCamelCase_ ).replace("""@@ """ , """""" ).strip() return out_string def lowerCAmelCase__ ( self: str , UpperCamelCase_: str , UpperCamelCase_: Optional[str] = None ): if not os.path.isdir(UpperCamelCase_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowerCamelCase = os.path.join( UpperCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowerCamelCase = os.path.join( UpperCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(UpperCamelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=UpperCamelCase_ , ensure_ascii=UpperCamelCase_ ) + """\n""" ) __lowerCamelCase = 0 with open(UpperCamelCase_ , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda UpperCamelCase_ : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) __lowerCamelCase = token_index writer.write(""" """.join(UpperCamelCase_ ) + """\n""" ) index += 1 return vocab_file, merge_file
80
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase_ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') class lowerCamelCase__( __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : int = BartphoTokenizer UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : List[str] = True def lowerCAmelCase__ ( self: Tuple ): super().setUp() __lowerCamelCase = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] __lowerCamelCase = dict(zip(UpperCamelCase_ , range(len(UpperCamelCase_ ) ) ) ) __lowerCamelCase = {"""unk_token""": """<unk>"""} __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F'{token} {vocab_tokens[token]}\n' ) __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase__ ( self: List[str] , **UpperCamelCase_: List[str] ): kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): __lowerCamelCase = """This is a là test""" __lowerCamelCase = """This is a<unk><unk> test""" return input_text, output_text def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) __lowerCamelCase = """This is a là test""" __lowerCamelCase = """▁This ▁is ▁a ▁l à ▁t est""".split() __lowerCamelCase = tokenizer.tokenize(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase_ ) , UpperCamelCase_ )
80
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_electra import ElectraTokenizer UpperCAmelCase_ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} UpperCAmelCase_ = { 'vocab_file': { 'google/electra-small-generator': ( 'https://huggingface.co/google/electra-small-generator/resolve/main/vocab.txt' ), 'google/electra-base-generator': 'https://huggingface.co/google/electra-base-generator/resolve/main/vocab.txt', 'google/electra-large-generator': ( 'https://huggingface.co/google/electra-large-generator/resolve/main/vocab.txt' ), 'google/electra-small-discriminator': ( 'https://huggingface.co/google/electra-small-discriminator/resolve/main/vocab.txt' ), 'google/electra-base-discriminator': ( 'https://huggingface.co/google/electra-base-discriminator/resolve/main/vocab.txt' ), 'google/electra-large-discriminator': ( 'https://huggingface.co/google/electra-large-discriminator/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'google/electra-small-generator': ( 'https://huggingface.co/google/electra-small-generator/resolve/main/tokenizer.json' ), 'google/electra-base-generator': ( 'https://huggingface.co/google/electra-base-generator/resolve/main/tokenizer.json' ), 'google/electra-large-generator': ( 'https://huggingface.co/google/electra-large-generator/resolve/main/tokenizer.json' ), 'google/electra-small-discriminator': ( 'https://huggingface.co/google/electra-small-discriminator/resolve/main/tokenizer.json' ), 'google/electra-base-discriminator': ( 'https://huggingface.co/google/electra-base-discriminator/resolve/main/tokenizer.json' ), 'google/electra-large-discriminator': ( 'https://huggingface.co/google/electra-large-discriminator/resolve/main/tokenizer.json' ), }, } UpperCAmelCase_ = { 'google/electra-small-generator': 512, 'google/electra-base-generator': 512, 'google/electra-large-generator': 512, 'google/electra-small-discriminator': 512, 'google/electra-base-discriminator': 512, 'google/electra-large-discriminator': 512, } UpperCAmelCase_ = { 'google/electra-small-generator': {'do_lower_case': True}, 'google/electra-base-generator': {'do_lower_case': True}, 'google/electra-large-generator': {'do_lower_case': True}, 'google/electra-small-discriminator': {'do_lower_case': True}, 'google/electra-base-discriminator': {'do_lower_case': True}, 'google/electra-large-discriminator': {'do_lower_case': True}, } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Optional[Any] = VOCAB_FILES_NAMES UpperCAmelCase__ : Any = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Tuple = ElectraTokenizer def __init__( self: Any , UpperCamelCase_: str=None , UpperCamelCase_: List[Any]=None , UpperCamelCase_: Tuple=True , UpperCamelCase_: str="[UNK]" , UpperCamelCase_: Dict="[SEP]" , UpperCamelCase_: int="[PAD]" , UpperCamelCase_: List[Any]="[CLS]" , UpperCamelCase_: Optional[Any]="[MASK]" , UpperCamelCase_: Any=True , UpperCamelCase_: Union[str, Any]=None , **UpperCamelCase_: Optional[Any] , ): super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , do_lower_case=UpperCamelCase_ , unk_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , tokenize_chinese_chars=UpperCamelCase_ , strip_accents=UpperCamelCase_ , **UpperCamelCase_ , ) __lowerCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , UpperCamelCase_ ) != do_lower_case or normalizer_state.get("""strip_accents""" , UpperCamelCase_ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , UpperCamelCase_ ) != tokenize_chinese_chars ): __lowerCamelCase = getattr(UpperCamelCase_ , normalizer_state.pop("""type""" ) ) __lowerCamelCase = do_lower_case __lowerCamelCase = strip_accents __lowerCamelCase = tokenize_chinese_chars __lowerCamelCase = normalizer_class(**UpperCamelCase_ ) __lowerCamelCase = do_lower_case def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: str , UpperCamelCase_: Dict=None ): __lowerCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: List[int] , UpperCamelCase_: Optional[List[int]] = None ): __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase__ ( self: str , UpperCamelCase_: str , UpperCamelCase_: Optional[str] = None ): __lowerCamelCase = self._tokenizer.model.save(UpperCamelCase_ , name=UpperCamelCase_ ) return tuple(UpperCamelCase_ )
80
def lowerCamelCase__ ( A__ : dict ): '''simple docstring''' __lowerCamelCase = set() # To detect a back edge, keep track of vertices currently in the recursion stack __lowerCamelCase = set() return any( node not in visited and depth_first_search(A__ , A__ , A__ , A__ ) for node in graph ) def lowerCamelCase__ ( A__ : dict , A__ : int , A__ : set , A__ : set ): '''simple docstring''' visited.add(A__ ) rec_stk.add(A__ ) for node in graph[vertex]: if node not in visited: if depth_first_search(A__ , A__ , A__ , A__ ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(A__ ) return False if __name__ == "__main__": from doctest import testmod testmod()
80
1
UpperCAmelCase_ = 8.31_4462 # Unit - J mol-1 K-1 def lowerCamelCase__ ( A__ : float , A__ : float , A__ : float ): '''simple docstring''' if moles < 0 or kelvin < 0 or volume < 0: raise ValueError("""Invalid inputs. Enter positive value.""" ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / volume def lowerCamelCase__ ( A__ : float , A__ : float , A__ : float ): '''simple docstring''' if moles < 0 or kelvin < 0 or pressure < 0: raise ValueError("""Invalid inputs. Enter positive value.""" ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / pressure if __name__ == "__main__": from doctest import testmod testmod()
80
from __future__ import annotations def lowerCamelCase__ ( A__ : list[float] , A__ : list[float] ): '''simple docstring''' __lowerCamelCase = sorted(numsa + numsa ) __lowerCamelCase, __lowerCamelCase = divmod(len(A__ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ = [float(x) for x in input('Enter the elements of first array: ').split()] UpperCAmelCase_ = [float(x) for x in input('Enter the elements of second array: ').split()] print(f"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
80
1
from manim import * class lowerCamelCase__( __lowerCamelCase): def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = Rectangle(height=0.5 , width=0.5 ) __lowerCamelCase = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) __lowerCamelCase = Rectangle(height=0.25 , width=0.25 ) __lowerCamelCase = [mem.copy() for i in range(6 )] __lowerCamelCase = [mem.copy() for i in range(6 )] __lowerCamelCase = VGroup(*UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0 ) __lowerCamelCase = VGroup(*UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0 ) __lowerCamelCase = VGroup(UpperCamelCase_ , UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0 ) __lowerCamelCase = Text("""CPU""" , font_size=24 ) __lowerCamelCase = Group(UpperCamelCase_ , UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0.5 , aligned_edge=UpperCamelCase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(UpperCamelCase_ ) __lowerCamelCase = [mem.copy() for i in range(4 )] __lowerCamelCase = VGroup(*UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0 ) __lowerCamelCase = Text("""GPU""" , font_size=24 ) __lowerCamelCase = Group(UpperCamelCase_ , UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0.5 , aligned_edge=UpperCamelCase_ ) gpu.move_to([-1, -1, 0] ) self.add(UpperCamelCase_ ) __lowerCamelCase = [mem.copy() for i in range(6 )] __lowerCamelCase = VGroup(*UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0 ) __lowerCamelCase = Text("""Model""" , font_size=24 ) __lowerCamelCase = Group(UpperCamelCase_ , UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0.5 , aligned_edge=UpperCamelCase_ ) model.move_to([3, -1.0, 0] ) self.add(UpperCamelCase_ ) __lowerCamelCase = [] __lowerCamelCase = [] for i, rect in enumerate(UpperCamelCase_ ): __lowerCamelCase = fill.copy().set_fill(UpperCamelCase_ , opacity=0.8 ) target.move_to(UpperCamelCase_ ) model_arr.append(UpperCamelCase_ ) __lowerCamelCase = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(UpperCamelCase_ , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(UpperCamelCase_ ) self.add(*UpperCamelCase_ , *UpperCamelCase_ ) __lowerCamelCase = [meta_mem.copy() for i in range(6 )] __lowerCamelCase = [meta_mem.copy() for i in range(6 )] __lowerCamelCase = VGroup(*UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0 ) __lowerCamelCase = VGroup(*UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0 ) __lowerCamelCase = VGroup(UpperCamelCase_ , UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0 ) __lowerCamelCase = Text("""Disk""" , font_size=24 ) __lowerCamelCase = Group(UpperCamelCase_ , UpperCamelCase_ ).arrange(UpperCamelCase_ , buff=0.5 , aligned_edge=UpperCamelCase_ ) disk.move_to([-4, -1.25, 0] ) self.add(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __lowerCamelCase = MarkupText( F'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = MarkupText( F'<span fgcolor=\'{BLUE}\'>●</span> Checkpoint' , font_size=18 , ) blue_text.next_to(UpperCamelCase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(UpperCamelCase_ ) __lowerCamelCase = MarkupText( F'Now watch as an input is passed through the model\nand how the memory is utilized and handled.' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCamelCase_ ) ) __lowerCamelCase = Square(0.3 ) input.set_fill(UpperCamelCase_ , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , UpperCamelCase_ , buff=0.5 ) self.play(Write(UpperCamelCase_ ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=UpperCamelCase_ , buff=0.02 ) self.play(MoveToTarget(UpperCamelCase_ ) ) self.play(FadeOut(UpperCamelCase_ ) ) __lowerCamelCase = Arrow(start=UpperCamelCase_ , end=UpperCamelCase_ , color=UpperCamelCase_ , buff=0.5 ) a.next_to(model_arr[0].get_left() , UpperCamelCase_ , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) __lowerCamelCase = MarkupText( F'As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCamelCase_ , run_time=3 ) ) __lowerCamelCase = {"""run_time""": 1, """fade_in""": True, """fade_out""": True, """buff""": 0.02} self.play( Write(UpperCamelCase_ ) , Circumscribe(model_arr[0] , color=UpperCamelCase_ , **UpperCamelCase_ ) , Circumscribe(model_cpu_arr[0] , color=UpperCamelCase_ , **UpperCamelCase_ ) , Circumscribe(gpu_rect[0] , color=UpperCamelCase_ , **UpperCamelCase_ ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) __lowerCamelCase = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , UpperCamelCase_ , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) __lowerCamelCase = AnimationGroup( FadeOut(UpperCamelCase_ , run_time=0.5 ) , MoveToTarget(UpperCamelCase_ , run_time=0.5 ) , FadeIn(UpperCamelCase_ , run_time=0.5 ) , lag_ratio=0.2 ) self.play(UpperCamelCase_ ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: __lowerCamelCase = 0.7 self.play( Circumscribe(model_arr[i] , **UpperCamelCase_ ) , Circumscribe(cpu_left_col_base[i] , **UpperCamelCase_ ) , Circumscribe(cpu_left_col_base[i + 1] , color=UpperCamelCase_ , **UpperCamelCase_ ) , Circumscribe(gpu_rect[0] , color=UpperCamelCase_ , **UpperCamelCase_ ) , Circumscribe(model_arr[i + 1] , color=UpperCamelCase_ , **UpperCamelCase_ ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=UpperCamelCase_ , **UpperCamelCase_ ) , Circumscribe(cpu_left_col_base[-1] , color=UpperCamelCase_ , **UpperCamelCase_ ) , Circumscribe(gpu_rect[0] , color=UpperCamelCase_ , **UpperCamelCase_ ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) __lowerCamelCase = a_c __lowerCamelCase = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(UpperCamelCase_ ) , FadeOut(UpperCamelCase_ , run_time=0.5 ) , ) __lowerCamelCase = MarkupText(F'Inference on a model too large for GPU memory\nis successfully completed.' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCamelCase_ , run_time=3 ) , MoveToTarget(UpperCamelCase_ ) ) self.wait()
80
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) self.assertTrue(torch.allclose(gelu_python(UpperCamelCase_ ) , torch_builtin(UpperCamelCase_ ) ) ) self.assertFalse(torch.allclose(gelu_python(UpperCamelCase_ ) , gelu_new(UpperCamelCase_ ) ) ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = get_activation("""gelu_10""" ) __lowerCamelCase = torch_builtin(UpperCamelCase_ ) __lowerCamelCase = geluaa(UpperCamelCase_ ) __lowerCamelCase = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(UpperCamelCase_ ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def lowerCAmelCase__ ( self: str ): get_activation("""gelu""" ) get_activation("""gelu_10""" ) get_activation("""gelu_fast""" ) get_activation("""gelu_new""" ) get_activation("""gelu_python""" ) get_activation("""gelu_pytorch_tanh""" ) get_activation("""linear""" ) get_activation("""mish""" ) get_activation("""quick_gelu""" ) get_activation("""relu""" ) get_activation("""sigmoid""" ) get_activation("""silu""" ) get_activation("""swish""" ) get_activation("""tanh""" ) with self.assertRaises(UpperCamelCase_ ): get_activation("""bogus""" ) with self.assertRaises(UpperCamelCase_ ): get_activation(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = 1 __lowerCamelCase = get_activation("""gelu""" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(UpperCamelCase_ ): __lowerCamelCase = acta.a
80
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'facebook/xlm-roberta-xl': 'https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json', 'facebook/xlm-roberta-xxl': 'https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : int = 'xlm-roberta-xl' def __init__( self: List[Any] , UpperCamelCase_: List[Any]=25_08_80 , UpperCamelCase_: Union[str, Any]=25_60 , UpperCamelCase_: Tuple=36 , UpperCamelCase_: int=32 , UpperCamelCase_: Tuple=1_02_40 , UpperCamelCase_: Any="gelu" , UpperCamelCase_: int=0.1 , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: Union[str, Any]=5_14 , UpperCamelCase_: Optional[Any]=1 , UpperCamelCase_: Any=0.02 , UpperCamelCase_: str=1E-05 , UpperCamelCase_: Optional[int]=1 , UpperCamelCase_: int=0 , UpperCamelCase_: Any=2 , UpperCamelCase_: Any="absolute" , UpperCamelCase_: Tuple=True , UpperCamelCase_: List[Any]=None , **UpperCamelCase_: int , ): super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = position_embedding_type __lowerCamelCase = use_cache __lowerCamelCase = classifier_dropout class lowerCamelCase__( __lowerCamelCase): @property def lowerCAmelCase__ ( self: Optional[Any] ): if self.task == "multiple-choice": __lowerCamelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowerCamelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
80
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class lowerCamelCase__( __lowerCamelCase): @slow @require_torch def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = EncoderDecoderModel.from_encoder_decoder_pretrained("""prajjwal1/bert-tiny""" , """prajjwal1/bert-tiny""" ) __lowerCamelCase = BertTokenizer.from_pretrained("""bert-base-uncased""" ) __lowerCamelCase = bertabert.config.encoder.vocab_size __lowerCamelCase = tokenizer.sep_token_id __lowerCamelCase = tokenizer.cls_token_id __lowerCamelCase = 1_28 __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""train[:1%]""" ) __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""validation[:1%]""" ) __lowerCamelCase = train_dataset.select(range(32 ) ) __lowerCamelCase = val_dataset.select(range(16 ) ) __lowerCamelCase = 4 def _map_to_encoder_decoder_inputs(UpperCamelCase_: List[Any] ): # Tokenizer will automatically set [BOS] <text> [EOS] __lowerCamelCase = tokenizer(batch["""article"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=5_12 ) __lowerCamelCase = tokenizer(batch["""highlights"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=1_28 ) __lowerCamelCase = inputs.input_ids __lowerCamelCase = inputs.attention_mask __lowerCamelCase = outputs.input_ids __lowerCamelCase = outputs.input_ids.copy() __lowerCamelCase = [ [-1_00 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["""labels"""] ] __lowerCamelCase = outputs.attention_mask assert all(len(UpperCamelCase_ ) == 5_12 for x in inputs.input_ids ) assert all(len(UpperCamelCase_ ) == 1_28 for x in outputs.input_ids ) return batch def _compute_metrics(UpperCamelCase_: int ): __lowerCamelCase = pred.label_ids __lowerCamelCase = pred.predictions # all unnecessary tokens are removed __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = sum([int(pred_str[i] == label_str[i] ) for i in range(len(UpperCamelCase_ ) )] ) / len(UpperCamelCase_ ) return {"accuracy": accuracy} # map train dataset __lowerCamelCase = train_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) train_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) # same for validation dataset __lowerCamelCase = val_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) val_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) __lowerCamelCase = self.get_auto_remove_tmp_dir() __lowerCamelCase = SeqaSeqTrainingArguments( output_dir=UpperCamelCase_ , per_device_train_batch_size=UpperCamelCase_ , per_device_eval_batch_size=UpperCamelCase_ , predict_with_generate=UpperCamelCase_ , evaluation_strategy="""steps""" , do_train=UpperCamelCase_ , do_eval=UpperCamelCase_ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer __lowerCamelCase = SeqaSeqTrainer( model=UpperCamelCase_ , args=UpperCamelCase_ , compute_metrics=_compute_metrics , train_dataset=UpperCamelCase_ , eval_dataset=UpperCamelCase_ , tokenizer=UpperCamelCase_ , ) # start training trainer.train()
80
1
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__: def __init__( self: Union[str, Any] , UpperCamelCase_: str = None , UpperCamelCase_: uuid.UUID = None , UpperCamelCase_: Dict=None , UpperCamelCase_: Any=None ): if not conversation_id: __lowerCamelCase = uuid.uuida() if past_user_inputs is None: __lowerCamelCase = [] if generated_responses is None: __lowerCamelCase = [] __lowerCamelCase = conversation_id __lowerCamelCase = past_user_inputs __lowerCamelCase = generated_responses __lowerCamelCase = text def __eq__( self: Optional[Any] , UpperCamelCase_: Union[str, Any] ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: str , UpperCamelCase_: bool = False ): if self.new_user_input: if overwrite: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' F'with: "{text}".' ) __lowerCamelCase = text else: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' F'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input' ) else: __lowerCamelCase = text def lowerCAmelCase__ ( self: List[str] ): if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __lowerCamelCase = None def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): self.generated_responses.append(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple ): for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self: Union[str, Any] ): __lowerCamelCase = F'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): __lowerCamelCase = """user""" if is_user else """bot""" output += F'{name} >> {text} \n' return output @add_end_docstrings( __lowerCamelCase , r'\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n ' , ) class lowerCamelCase__( __lowerCamelCase): def __init__( self: List[str] , *UpperCamelCase_: List[Any] , **UpperCamelCase_: str ): super().__init__(*UpperCamelCase_ , **UpperCamelCase_ ) if self.tokenizer.pad_token_id is None: __lowerCamelCase = self.tokenizer.eos_token def lowerCAmelCase__ ( self: str , UpperCamelCase_: int=None , UpperCamelCase_: Any=None , UpperCamelCase_: Union[str, Any]=None , **UpperCamelCase_: int ): __lowerCamelCase = {} __lowerCamelCase = {} __lowerCamelCase = {} if min_length_for_response is not None: __lowerCamelCase = min_length_for_response if minimum_tokens is not None: __lowerCamelCase = minimum_tokens if "max_length" in generate_kwargs: __lowerCamelCase = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __lowerCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCamelCase_ ) return preprocess_params, forward_params, postprocess_params def __call__( self: Any , UpperCamelCase_: Union[Conversation, List[Conversation]] , UpperCamelCase_: Optional[int]=0 , **UpperCamelCase_: Optional[int] ): __lowerCamelCase = super().__call__(UpperCamelCase_ , num_workers=UpperCamelCase_ , **UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and len(UpperCamelCase_ ) == 1: return outputs[0] return outputs def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Conversation , UpperCamelCase_: Optional[Any]=32 ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( F'Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ' """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): __lowerCamelCase = self.tokenizer._build_conversation_input_ids(UpperCamelCase_ ) else: # If the tokenizer cannot handle conversations, we default to only the old version __lowerCamelCase = self._legacy_parse_and_tokenize(UpperCamelCase_ ) if self.framework == "pt": __lowerCamelCase = torch.LongTensor([input_ids] ) elif self.framework == "tf": __lowerCamelCase = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str=10 , **UpperCamelCase_: List[str] ): __lowerCamelCase = generate_kwargs.get("""max_length""" , self.model.config.max_length ) __lowerCamelCase = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(F'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})' ) __lowerCamelCase = max_length - minimum_tokens __lowerCamelCase = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: __lowerCamelCase = model_inputs["""attention_mask"""][:, -trim:] __lowerCamelCase = model_inputs.pop("""conversation""" ) __lowerCamelCase = max_length __lowerCamelCase = self.model.generate(**UpperCamelCase_ , **UpperCamelCase_ ) if self.model.config.is_encoder_decoder: __lowerCamelCase = 1 else: __lowerCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Optional[Any] , UpperCamelCase_: int=True ): __lowerCamelCase = model_outputs["""output_ids"""] __lowerCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_ , ) __lowerCamelCase = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCamelCase_ ) return conversation def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Conversation ): __lowerCamelCase = self.tokenizer.eos_token_id __lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) ) if len(UpperCamelCase_ ) > self.tokenizer.model_max_length: __lowerCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
80
class lowerCamelCase__: # Public class to implement a graph def __init__( self: Dict , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): __lowerCamelCase = row __lowerCamelCase = col __lowerCamelCase = graph def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): # Checking all 8 elements surrounding nth element __lowerCamelCase = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __lowerCamelCase = [-1, 0, 1, -1, 1, -1, 0, 1] __lowerCamelCase = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): # And finally, count all islands. __lowerCamelCase = [[False for j in range(self.COL )] for i in range(self.ROW )] __lowerCamelCase = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) count += 1 return count
80
1
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class lowerCamelCase__: def __init__( self: List[str] , UpperCamelCase_: Optional[int] , UpperCamelCase_: List[Any]=13 , UpperCamelCase_: Dict=32 , UpperCamelCase_: str=2 , UpperCamelCase_: int=3 , UpperCamelCase_: List[str]=16 , UpperCamelCase_: List[Any]=[1, 2, 1] , UpperCamelCase_: Optional[Any]=[2, 2, 4] , UpperCamelCase_: Any=2 , UpperCamelCase_: Optional[int]=2.0 , UpperCamelCase_: Tuple=True , UpperCamelCase_: Dict=0.0 , UpperCamelCase_: Optional[Any]=0.0 , UpperCamelCase_: Union[str, Any]=0.1 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: int=False , UpperCamelCase_: Dict=True , UpperCamelCase_: List[str]=0.02 , UpperCamelCase_: List[str]=1E-5 , UpperCamelCase_: Optional[int]=True , UpperCamelCase_: int=None , UpperCamelCase_: Tuple=True , UpperCamelCase_: str=10 , UpperCamelCase_: List[Any]=8 , UpperCamelCase_: Optional[int]=["stage1", "stage2", "stage3"] , UpperCamelCase_: int=[1, 2, 3] , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = embed_dim __lowerCamelCase = depths __lowerCamelCase = num_heads __lowerCamelCase = window_size __lowerCamelCase = mlp_ratio __lowerCamelCase = qkv_bias __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_absolute_embeddings __lowerCamelCase = patch_norm __lowerCamelCase = layer_norm_eps __lowerCamelCase = initializer_range __lowerCamelCase = is_training __lowerCamelCase = scope __lowerCamelCase = use_labels __lowerCamelCase = type_sequence_label_size __lowerCamelCase = encoder_stride __lowerCamelCase = out_features __lowerCamelCase = out_indices def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = self.get_config() return config, pixel_values, labels def lowerCAmelCase__ ( self: Optional[Any] ): return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Dict ): __lowerCamelCase = MaskFormerSwinModel(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = model(UpperCamelCase_ ) __lowerCamelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowerCamelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: Any , UpperCamelCase_: List[str] , UpperCamelCase_: List[str] ): __lowerCamelCase = MaskFormerSwinBackbone(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = model(UpperCamelCase_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(UpperCamelCase_ ): __lowerCamelCase = ["""stem"""] __lowerCamelCase = MaskFormerSwinBackbone(config=UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = config_and_inputs __lowerCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : int = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) UpperCAmelCase__ : List[str] = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : Union[str, Any] = False UpperCAmelCase__ : str = False UpperCAmelCase__ : str = False def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = MaskFormerSwinModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=UpperCamelCase_ , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def lowerCAmelCase__ ( self: Optional[int] ): pass def lowerCAmelCase__ ( self: Tuple ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCAmelCase__ ( self: Any ): return def lowerCAmelCase__ ( self: str ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase_ ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*UpperCamelCase_ ) @unittest.skip("""Swin does not use inputs_embeds""" ) def lowerCAmelCase__ ( self: Any ): pass @unittest.skip("""Swin does not support feedforward chunking""" ) def lowerCAmelCase__ ( self: Optional[Any] ): pass def lowerCAmelCase__ ( self: Any ): __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(UpperCamelCase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase_ , nn.Linear ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(UpperCamelCase_ ) __lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase_ ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def lowerCAmelCase__ ( self: Tuple ): pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def lowerCAmelCase__ ( self: List[Any] ): pass def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Tuple , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Tuple , UpperCamelCase_: Optional[int] ): __lowerCamelCase = model_class(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() with torch.no_grad(): __lowerCamelCase = model(**self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) ) __lowerCamelCase = outputs.hidden_states __lowerCamelCase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(UpperCamelCase_ ) , UpperCamelCase_ ) # Swin has a different seq_length __lowerCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowerCamelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowerCamelCase = True self.check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCamelCase = True self.check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = 3 __lowerCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowerCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowerCamelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowerCamelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowerCamelCase = True self.check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCamelCase = True self.check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def lowerCAmelCase__ ( self: Dict ): pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def lowerCAmelCase__ ( self: str ): pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def lowerCAmelCase__ ( self: str ): pass def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(UpperCamelCase_: Optional[int] ): __lowerCamelCase = 0 return t def check_equivalence(UpperCamelCase_: List[Any] , UpperCamelCase_: List[Any] , UpperCamelCase_: Tuple , UpperCamelCase_: Any={} ): with torch.no_grad(): __lowerCamelCase = model(**UpperCamelCase_ , return_dict=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = model(**UpperCamelCase_ , return_dict=UpperCamelCase_ , **UpperCamelCase_ ).to_tuple() def recursive_check(UpperCamelCase_: str , UpperCamelCase_: List[str] ): if isinstance(UpperCamelCase_ , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(UpperCamelCase_ , UpperCamelCase_ ): recursive_check(UpperCamelCase_ , UpperCamelCase_ ) elif isinstance(UpperCamelCase_ , UpperCamelCase_ ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(UpperCamelCase_ , UpperCamelCase_ ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(UpperCamelCase_ ) , set_nan_tensor_to_zero(UpperCamelCase_ ) , atol=1E-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:' F' {torch.isnan(UpperCamelCase_ ).any()} and `inf`: {torch.isinf(UpperCamelCase_ )}. Dict has' F' `nan`: {torch.isnan(UpperCamelCase_ ).any()} and `inf`: {torch.isinf(UpperCamelCase_ )}.' ) , ) recursive_check(UpperCamelCase_ , UpperCamelCase_ ) for model_class in self.all_model_classes: __lowerCamelCase = model_class(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) check_equivalence(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ , return_labels=UpperCamelCase_ ) __lowerCamelCase = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ , return_labels=UpperCamelCase_ ) check_equivalence(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) check_equivalence(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , {"""output_hidden_states""": True} ) __lowerCamelCase = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ , return_labels=UpperCamelCase_ ) __lowerCamelCase = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ , return_labels=UpperCamelCase_ ) check_equivalence(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , {"""output_hidden_states""": True} ) @require_torch class lowerCamelCase__( unittest.TestCase , __lowerCamelCase): UpperCAmelCase__ : Dict = (MaskFormerSwinBackbone,) if is_torch_available() else () UpperCAmelCase__ : Union[str, Any] = MaskFormerSwinConfig def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = MaskFormerSwinModelTester(self ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowerCamelCase = backbone_class(UpperCamelCase_ ) backbone.to(UpperCamelCase_ ) backbone.eval() __lowerCamelCase = backbone(**UpperCamelCase_ ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , UpperCamelCase_ ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowerCamelCase = backbone(**UpperCamelCase_ , output_hidden_states=UpperCamelCase_ ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowerCamelCase = backbone(**UpperCamelCase_ , output_attentions=UpperCamelCase_ ) self.assertIsNotNone(outputs.attentions )
80
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = DPTConfig() if "large" in checkpoint_url: __lowerCamelCase = 1024 __lowerCamelCase = 4096 __lowerCamelCase = 24 __lowerCamelCase = 16 __lowerCamelCase = [5, 11, 17, 23] __lowerCamelCase = [256, 512, 1024, 1024] __lowerCamelCase = (1, 384, 384) if "ade" in checkpoint_url: __lowerCamelCase = True __lowerCamelCase = 150 __lowerCamelCase = """huggingface/label-files""" __lowerCamelCase = """ade20k-id2label.json""" __lowerCamelCase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type="""dataset""" ) ) , """r""" ) ) __lowerCamelCase = {int(A__ ): v for k, v in idalabel.items()} __lowerCamelCase = idalabel __lowerCamelCase = {v: k for k, v in idalabel.items()} __lowerCamelCase = [1, 150, 480, 480] return config, expected_shape def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' __lowerCamelCase = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.encoder""" ) if "pretrained.model" in name: __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.embeddings""" ) if "patch_embed" in name: __lowerCamelCase = name.replace("""patch_embed""" , """patch_embeddings""" ) if "pos_embed" in name: __lowerCamelCase = name.replace("""pos_embed""" , """position_embeddings""" ) if "attn.proj" in name: __lowerCamelCase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "proj" in name and "project" not in name: __lowerCamelCase = name.replace("""proj""" , """projection""" ) if "blocks" in name: __lowerCamelCase = name.replace("""blocks""" , """layer""" ) if "mlp.fc1" in name: __lowerCamelCase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: __lowerCamelCase = name.replace("""mlp.fc2""" , """output.dense""" ) if "norm1" in name: __lowerCamelCase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: __lowerCamelCase = name.replace("""norm2""" , """layernorm_after""" ) if "scratch.output_conv" in name: __lowerCamelCase = name.replace("""scratch.output_conv""" , """head""" ) if "scratch" in name: __lowerCamelCase = name.replace("""scratch""" , """neck""" ) if "layer1_rn" in name: __lowerCamelCase = name.replace("""layer1_rn""" , """convs.0""" ) if "layer2_rn" in name: __lowerCamelCase = name.replace("""layer2_rn""" , """convs.1""" ) if "layer3_rn" in name: __lowerCamelCase = name.replace("""layer3_rn""" , """convs.2""" ) if "layer4_rn" in name: __lowerCamelCase = name.replace("""layer4_rn""" , """convs.3""" ) if "refinenet" in name: __lowerCamelCase = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __lowerCamelCase = name.replace(f'refinenet{layer_idx}' , f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: __lowerCamelCase = name.replace("""out_conv""" , """projection""" ) if "resConfUnit1" in name: __lowerCamelCase = name.replace("""resConfUnit1""" , """residual_layer1""" ) if "resConfUnit2" in name: __lowerCamelCase = name.replace("""resConfUnit2""" , """residual_layer2""" ) if "conv1" in name: __lowerCamelCase = name.replace("""conv1""" , """convolution1""" ) if "conv2" in name: __lowerCamelCase = name.replace("""conv2""" , """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.0.project.0""" , """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.0.project.0""" , """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.0.project.0""" , """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.0.project.0""" , """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.3""" , """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.4""" , """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.3""" , """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.4""" , """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.3""" , """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.3""" , """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.4""" , """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: __lowerCamelCase = name.replace("""pretrained""" , """dpt""" ) if "bn" in name: __lowerCamelCase = name.replace("""bn""" , """batch_norm""" ) if "head" in name: __lowerCamelCase = name.replace("""head""" , """head.head""" ) if "encoder.norm" in name: __lowerCamelCase = name.replace("""encoder.norm""" , """layernorm""" ) if "auxlayer" in name: __lowerCamelCase = name.replace("""auxlayer""" , """auxiliary_head.head""" ) return name def lowerCamelCase__ ( A__ : Tuple , A__ : Any ): '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __lowerCamelCase = in_proj_weight[: config.hidden_size, :] __lowerCamelCase = in_proj_bias[: config.hidden_size] __lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] __lowerCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" __lowerCamelCase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( A__ : Optional[int] , A__ : Union[str, Any] , A__ : List[str] , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = get_dpt_config(A__ ) # load original state_dict from URL __lowerCamelCase = torch.hub.load_state_dict_from_url(A__ , map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(A__ ) # rename keys for key in state_dict.copy().keys(): __lowerCamelCase = state_dict.pop(A__ ) __lowerCamelCase = val # read in qkv matrices read_in_q_k_v(A__ , A__ ) # load HuggingFace model __lowerCamelCase = DPTForSemanticSegmentation(A__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(A__ ) model.load_state_dict(A__ ) model.eval() # Check outputs on an image __lowerCamelCase = 480 if """ade""" in checkpoint_url else 384 __lowerCamelCase = DPTImageProcessor(size=A__ ) __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(A__ , return_tensors="""pt""" ) # forward pass __lowerCamelCase = model(**A__ ).logits if """ade""" in checkpoint_url else model(**A__ ).predicted_depth # Assert logits __lowerCamelCase = torch.tensor([[6.3_199, 6.3_629, 6.4_148], [6.3_850, 6.3_615, 6.4_166], [6.3_519, 6.3_176, 6.3_575]] ) if "ade" in checkpoint_url: __lowerCamelCase = torch.tensor([[4.0_480, 4.2_420, 4.4_360], [4.3_124, 4.5_693, 4.8_261], [4.5_768, 4.8_965, 5.2_163]] ) assert outputs.shape == torch.Size(A__ ) assert ( torch.allclose(outputs[0, 0, :3, :3] , A__ , atol=1E-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3] , A__ ) ) Path(A__ ).mkdir(exist_ok=A__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(A__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=A__ , ) image_processor.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=A__ , ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) UpperCAmelCase_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
80
1
def lowerCamelCase__ ( A__ : float , A__ : float ): '''simple docstring''' if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
80
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
80
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'post_extract_proj': 'feature_projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.upsample.0': 'encoder.upsample.projection', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'layer_norm', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } def lowerCamelCase__ ( A__ : str , A__ : str , A__ : str , A__ : int , A__ : Union[str, Any] ): '''simple docstring''' for attribute in key.split(""".""" ): __lowerCamelCase = getattr(A__ , A__ ) if weight_type is not None: __lowerCamelCase = getattr(A__ , A__ ).shape else: __lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": __lowerCamelCase = value elif weight_type == "weight_g": __lowerCamelCase = value elif weight_type == "weight_v": __lowerCamelCase = value elif weight_type == "bias": __lowerCamelCase = value else: __lowerCamelCase = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def lowerCamelCase__ ( A__ : Tuple , A__ : Optional[int] , A__ : List[str] ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = fairseq_model.state_dict() __lowerCamelCase = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): __lowerCamelCase = False if "conv_layers" in name: load_conv_layer( A__ , A__ , A__ , A__ , hf_model.config.feat_extract_norm == """group""" , ) __lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): __lowerCamelCase = """sew.""" + mapped_key if (is_finetuned and mapped_key != """lm_head""") else mapped_key if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: __lowerCamelCase = True if "*" in mapped_key: __lowerCamelCase = name.split(A__ )[0].split(""".""" )[-2] __lowerCamelCase = mapped_key.replace("""*""" , A__ ) if "weight_g" in name: __lowerCamelCase = """weight_g""" elif "weight_v" in name: __lowerCamelCase = """weight_v""" elif "weight" in name: __lowerCamelCase = """weight""" elif "bias" in name: __lowerCamelCase = """bias""" else: __lowerCamelCase = None set_recursively(A__ , A__ , A__ , A__ , A__ ) continue if not is_used: unused_weights.append(A__ ) logger.warning(f'Unused weights: {unused_weights}' ) def lowerCamelCase__ ( A__ : Dict , A__ : int , A__ : Dict , A__ : Optional[Any] , A__ : Optional[Any] ): '''simple docstring''' __lowerCamelCase = full_name.split("""conv_layers.""" )[-1] __lowerCamelCase = name.split(""".""" ) __lowerCamelCase = int(items[0] ) __lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) __lowerCamelCase = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(A__ ) def lowerCamelCase__ ( A__ : List[str] , A__ : Optional[int] ): '''simple docstring''' __lowerCamelCase = SEWConfig() if is_finetuned: __lowerCamelCase = model.wav_encoder.wav_model.cfg else: __lowerCamelCase = model.cfg __lowerCamelCase = fs_config.conv_bias __lowerCamelCase = eval(fs_config.conv_feature_layers ) __lowerCamelCase = [x[0] for x in conv_layers] __lowerCamelCase = [x[1] for x in conv_layers] __lowerCamelCase = [x[2] for x in conv_layers] __lowerCamelCase = """gelu""" __lowerCamelCase = """layer""" if fs_config.extractor_mode == """layer_norm""" else """group""" __lowerCamelCase = 0.0 __lowerCamelCase = fs_config.activation_fn.name __lowerCamelCase = fs_config.encoder_embed_dim __lowerCamelCase = 0.02 __lowerCamelCase = fs_config.encoder_ffn_embed_dim __lowerCamelCase = 1E-5 __lowerCamelCase = fs_config.encoder_layerdrop __lowerCamelCase = fs_config.encoder_attention_heads __lowerCamelCase = fs_config.conv_pos_groups __lowerCamelCase = fs_config.conv_pos __lowerCamelCase = len(A__ ) __lowerCamelCase = fs_config.encoder_layers __lowerCamelCase = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: __lowerCamelCase = model.cfg __lowerCamelCase = fs_config.final_dropout __lowerCamelCase = fs_config.layerdrop __lowerCamelCase = fs_config.activation_dropout __lowerCamelCase = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 __lowerCamelCase = fs_config.attention_dropout __lowerCamelCase = fs_config.dropout_input __lowerCamelCase = fs_config.dropout __lowerCamelCase = fs_config.mask_channel_length __lowerCamelCase = fs_config.mask_channel_prob __lowerCamelCase = fs_config.mask_length __lowerCamelCase = fs_config.mask_prob __lowerCamelCase = """Wav2Vec2FeatureExtractor""" __lowerCamelCase = """Wav2Vec2CTCTokenizer""" return config @torch.no_grad() def lowerCamelCase__ ( A__ : List[str] , A__ : int , A__ : Any=None , A__ : str=None , A__ : Optional[int]=True ): '''simple docstring''' if is_finetuned: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) else: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: __lowerCamelCase = SEWConfig.from_pretrained(A__ ) else: __lowerCamelCase = convert_config(model[0] , A__ ) __lowerCamelCase = model[0].eval() __lowerCamelCase = True if config.feat_extract_norm == """layer""" else False __lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=A__ , return_attention_mask=A__ , ) if is_finetuned: if dict_path: __lowerCamelCase = Dictionary.load(A__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowerCamelCase = target_dict.pad_index __lowerCamelCase = target_dict.bos_index __lowerCamelCase = target_dict.pad_index __lowerCamelCase = target_dict.bos_index __lowerCamelCase = target_dict.eos_index __lowerCamelCase = len(target_dict.symbols ) __lowerCamelCase = os.path.join(A__ , """vocab.json""" ) if not os.path.isdir(A__ ): logger.error("""--pytorch_dump_folder_path ({}) should be a directory""".format(A__ ) ) return os.makedirs(A__ , exist_ok=A__ ) with open(A__ , """w""" , encoding="""utf-8""" ) as vocab_handle: json.dump(target_dict.indices , A__ ) __lowerCamelCase = WavaVecaCTCTokenizer( A__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="""|""" , do_lower_case=A__ , ) __lowerCamelCase = WavaVecaProcessor(feature_extractor=A__ , tokenizer=A__ ) processor.save_pretrained(A__ ) __lowerCamelCase = SEWForCTC(A__ ) else: __lowerCamelCase = SEWModel(A__ ) feature_extractor.save_pretrained(A__ ) recursively_load_weights(A__ , A__ , A__ ) hf_model.save_pretrained(A__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--is_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) UpperCAmelCase_ = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
80
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/config.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/config.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/config.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/config.json', 'bert-base-multilingual-uncased': 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json', 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/config.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/config.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-base-cased-finetuned-mrpc': 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json', 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json', 'bert-base-german-dbmdz-uncased': 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json', 'cl-tohoku/bert-base-japanese': 'https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json', 'cl-tohoku/bert-base-japanese-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json' ), 'wietsedv/bert-base-dutch-cased': 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json', # See all BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = 'bert' def __init__( self: List[str] , UpperCamelCase_: str=3_05_22 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Tuple=12 , UpperCamelCase_: int=12 , UpperCamelCase_: int=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[Any]=0.1 , UpperCamelCase_: Optional[int]=5_12 , UpperCamelCase_: List[Any]=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: List[str]=1E-12 , UpperCamelCase_: Dict=0 , UpperCamelCase_: List[Any]="absolute" , UpperCamelCase_: Tuple=True , UpperCamelCase_: Tuple=None , **UpperCamelCase_: Optional[Any] , ): super().__init__(pad_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = position_embedding_type __lowerCamelCase = use_cache __lowerCamelCase = classifier_dropout class lowerCamelCase__( __lowerCamelCase): @property def lowerCAmelCase__ ( self: Any ): if self.task == "multiple-choice": __lowerCamelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowerCamelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
80
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to properly calculate the metrics on the # validation dataset when in a distributed system, and builds off the # `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## UpperCAmelCase_ = 16 UpperCAmelCase_ = 32 def lowerCamelCase__ ( A__ : Accelerator , A__ : int = 16 ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained("""bert-base-cased""" ) __lowerCamelCase = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(A__ : Any ): # max_length=None => use the model max length (it's actually the default) __lowerCamelCase = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=A__ , max_length=A__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): __lowerCamelCase = datasets.map( A__ , batched=A__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __lowerCamelCase = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(A__ : Any ): # On TPU it's best to pad everything to the same length or training will be very slow. __lowerCamelCase = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": __lowerCamelCase = 16 elif accelerator.mixed_precision != "no": __lowerCamelCase = 8 else: __lowerCamelCase = None return tokenizer.pad( A__ , padding="""longest""" , max_length=A__ , pad_to_multiple_of=A__ , return_tensors="""pt""" , ) # Instantiate dataloaders. __lowerCamelCase = DataLoader( tokenized_datasets["""train"""] , shuffle=A__ , collate_fn=A__ , batch_size=A__ ) __lowerCamelCase = DataLoader( tokenized_datasets["""validation"""] , shuffle=A__ , collate_fn=A__ , batch_size=A__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders UpperCAmelCase_ = mocked_dataloaders # noqa: F811 def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Any ): '''simple docstring''' if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , A__ ) == "1": __lowerCamelCase = 2 # Initialize accelerator __lowerCamelCase = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __lowerCamelCase = config["""lr"""] __lowerCamelCase = int(config["""num_epochs"""] ) __lowerCamelCase = int(config["""seed"""] ) __lowerCamelCase = int(config["""batch_size"""] ) __lowerCamelCase = evaluate.load("""glue""" , """mrpc""" ) # If the batch size is too big we use gradient accumulation __lowerCamelCase = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: __lowerCamelCase = batch_size // MAX_GPU_BATCH_SIZE __lowerCamelCase = MAX_GPU_BATCH_SIZE set_seed(A__ ) __lowerCamelCase, __lowerCamelCase = get_dataloaders(A__ , A__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __lowerCamelCase = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=A__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __lowerCamelCase = model.to(accelerator.device ) # Instantiate optimizer __lowerCamelCase = AdamW(params=model.parameters() , lr=A__ ) # Instantiate scheduler __lowerCamelCase = get_linear_schedule_with_warmup( optimizer=A__ , num_warmup_steps=100 , num_training_steps=(len(A__ ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = accelerator.prepare( A__ , A__ , A__ , A__ , A__ ) # Now we train the model for epoch in range(A__ ): model.train() for step, batch in enumerate(A__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) __lowerCamelCase = model(**A__ ) __lowerCamelCase = outputs.loss __lowerCamelCase = loss / gradient_accumulation_steps accelerator.backward(A__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() __lowerCamelCase = 0 for step, batch in enumerate(A__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __lowerCamelCase = model(**A__ ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) __lowerCamelCase, __lowerCamelCase = accelerator.gather((predictions, batch["""labels"""]) ) # New Code # # First we check if it's a distributed system if accelerator.use_distributed: # Then see if we're on the last batch of our eval dataloader if step == len(A__ ) - 1: # Last batch needs to be truncated on distributed systems as it contains additional samples __lowerCamelCase = predictions[: len(eval_dataloader.dataset ) - samples_seen] __lowerCamelCase = references[: len(eval_dataloader.dataset ) - samples_seen] else: # Otherwise we add the number of samples seen samples_seen += references.shape[0] # All of this can be avoided if you use `Accelerator.gather_for_metrics` instead of `Accelerator.gather`: # accelerator.gather_for_metrics((predictions, batch["labels"])) metric.add_batch( predictions=A__ , references=A__ , ) __lowerCamelCase = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'epoch {epoch}:' , A__ ) def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=A__ , default=A__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) __lowerCamelCase = parser.parse_args() __lowerCamelCase = {"""lr""": 2E-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(A__ , A__ ) if __name__ == "__main__": main()
80
from __future__ import annotations from math import ceil, floor, sqrt def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0] __lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target __lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target __lowerCamelCase = 0 # an estimate of b, using the quadratic formula __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the triangle number corresponding to b_floor __lowerCamelCase = 42 # the triangle number corresponding to b_ceil __lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): __lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 __lowerCamelCase = floor(A__ ) __lowerCamelCase = ceil(A__ ) __lowerCamelCase = triangle_numbers[b_floor] __lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_first_guess * triangle_a __lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_second_guess * triangle_a __lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(f"""{solution() = }""")
80
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
80
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[str] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet, attn in zip(self.resnets , self.attentions ): __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: str , UpperCamelCase_: Any , UpperCamelCase_: Optional[int] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet in self.resnets: __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: Tuple , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: List[Any]=True ): for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Optional[Any]=True ): for resnet in self.resnets: # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: int ): # there is always at least one resnet __lowerCamelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] __lowerCamelCase = [] for _ in range(self.num_layers ): __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions def __call__( self: int , UpperCamelCase_: Any , UpperCamelCase_: int , UpperCamelCase_: Dict , UpperCamelCase_: Optional[int]=True ): __lowerCamelCase = self.resnets[0](UpperCamelCase_ , UpperCamelCase_ ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) return hidden_states
80
1
def lowerCamelCase__ ( A__ : Tuple , A__ : Tuple , A__ : List[Any] , A__ : str ): '''simple docstring''' if height >= 1: move_tower(height - 1 , A__ , A__ , A__ ) move_disk(A__ , A__ ) move_tower(height - 1 , A__ , A__ , A__ ) def lowerCamelCase__ ( A__ : Tuple , A__ : Any ): '''simple docstring''' print("""moving disk from""" , A__ , """to""" , A__ ) def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = int(input("""Height of hanoi: """ ).strip() ) move_tower(A__ , """A""" , """B""" , """C""" ) if __name__ == "__main__": main()
80
import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging UpperCAmelCase_ = ['bart.large', 'bart.large.mnli', 'bart.large.cnn', 'bart_xsum/model.pt'] UpperCAmelCase_ = {'bart.large': BartModel, 'bart.large.mnli': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('0.9.0'): raise Exception('requires fairseq >= 0.9.0') logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = ' Hello world! cécé herlolip' UpperCAmelCase_ = [ ('model.classification_heads.mnli.dense.weight', 'classification_head.dense.weight'), ('model.classification_heads.mnli.dense.bias', 'classification_head.dense.bias'), ('model.classification_heads.mnli.out_proj.weight', 'classification_head.out_proj.weight'), ('model.classification_heads.mnli.out_proj.bias', 'classification_head.out_proj.bias'), ] def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """_float_tensor""", ] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Tuple , A__ : Any , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase = dct.pop(A__ ) __lowerCamelCase = val def lowerCamelCase__ ( A__ : Tuple ): '''simple docstring''' __lowerCamelCase = torch.load(A__ , map_location="""cpu""" ) __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , """bart.large.cnn""" ).eval() hub_interface.model.load_state_dict(sd["""model"""] ) return hub_interface def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = emb.weight.shape __lowerCamelCase = nn.Linear(A__ , A__ , bias=A__ ) __lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Optional[int] , A__ : Dict=None ): '''simple docstring''' if not os.path.exists(A__ ): __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , A__ ).eval() else: __lowerCamelCase = load_xsum_checkpoint(A__ ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: __lowerCamelCase = checkpoint_path.replace(""".""" , """-""" ) __lowerCamelCase = BartConfig.from_pretrained(A__ ) __lowerCamelCase = bart.encode(A__ ).unsqueeze(0 ) __lowerCamelCase = BartTokenizer.from_pretrained(A__ ).encode(A__ , return_tensors="""pt""" ).unsqueeze(0 ) if not torch.eq(A__ , A__ ).all(): raise ValueError( f'converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}' ) if checkpoint_path == "bart.large.mnli": __lowerCamelCase = bart.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""model.decoder.embed_tokens.weight"""] for src, dest in mnli_rename_keys: rename_key(A__ , A__ , A__ ) __lowerCamelCase = BartForSequenceClassification(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = bart.predict("""mnli""" , A__ , return_logits=A__ ) __lowerCamelCase = model(A__ )[0] # logits else: # no classification heads to worry about __lowerCamelCase = bart.model.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""decoder.embed_tokens.weight"""] __lowerCamelCase = bart.extract_features(A__ ) if hf_checkpoint_name == "facebook/bart-large": __lowerCamelCase = BartModel(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = model(A__ ).model[0] else: __lowerCamelCase = BartForConditionalGeneration(A__ ).eval() # an existing summarization ckpt model.model.load_state_dict(A__ ) if hasattr(A__ , """lm_head""" ): __lowerCamelCase = make_linear_from_emb(model.model.shared ) __lowerCamelCase = model.model(A__ )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f'`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}' ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError("""Some values in `fairseq_output` are different from `new_model_outputs`""" ) Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default=None, type=str, help='Which huggingface architecture to use: bart-large-xsum' ) UpperCAmelCase_ = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
80
1
from __future__ import annotations def lowerCamelCase__ ( A__ : int , A__ : int ): '''simple docstring''' if b == 0: return (1, 0) ((__lowerCamelCase), (__lowerCamelCase)) = extended_euclid(A__ , a % b ) __lowerCamelCase = a // b return (y, x - k * y) def lowerCamelCase__ ( A__ : int , A__ : int , A__ : int , A__ : int ): '''simple docstring''' ((__lowerCamelCase), (__lowerCamelCase)) = extended_euclid(A__ , A__ ) __lowerCamelCase = na * na __lowerCamelCase = ra * x * na + ra * y * na return (n % m + m) % m def lowerCamelCase__ ( A__ : int , A__ : int ): '''simple docstring''' ((__lowerCamelCase), (__lowerCamelCase)) = extended_euclid(A__ , A__ ) if b < 0: __lowerCamelCase = (b % n + n) % n return b def lowerCamelCase__ ( A__ : int , A__ : int , A__ : int , A__ : int ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = invert_modulo(A__ , A__ ), invert_modulo(A__ , A__ ) __lowerCamelCase = na * na __lowerCamelCase = ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name='chinese_remainder_theorem', verbose=True) testmod(name='chinese_remainder_theorem2', verbose=True) testmod(name='invert_modulo', verbose=True) testmod(name='extended_euclid', verbose=True)
80
import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class lowerCamelCase__: def __init__( self: Tuple , UpperCamelCase_: Any , UpperCamelCase_: List[Any]=14 , UpperCamelCase_: int=7 , UpperCamelCase_: Union[str, Any]=True , UpperCamelCase_: Dict=True , UpperCamelCase_: Optional[Any]=True , UpperCamelCase_: Tuple=True , UpperCamelCase_: List[str]=True , UpperCamelCase_: int=99 , UpperCamelCase_: str=32 , UpperCamelCase_: List[Any]=5 , UpperCamelCase_: Optional[int]=4 , UpperCamelCase_: List[Any]=37 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[str]=5_12 , UpperCamelCase_: Dict=16 , UpperCamelCase_: List[str]=2 , UpperCamelCase_: Optional[Any]=0.02 , UpperCamelCase_: List[str]=3 , UpperCamelCase_: Tuple=4 , UpperCamelCase_: Tuple=None , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_input_mask __lowerCamelCase = use_labels __lowerCamelCase = use_mc_token_ids __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = self.vocab_size - 1 def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None if self.use_mc_token_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = self.get_config() __lowerCamelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def lowerCAmelCase__ ( self: Dict ): return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: str , UpperCamelCase_: Dict , UpperCamelCase_: Tuple , UpperCamelCase_: Any , UpperCamelCase_: List[str] , *UpperCamelCase_: Optional[Any] ): __lowerCamelCase = CTRLModel(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , head_mask=UpperCamelCase_ ) model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ ) __lowerCamelCase = model(UpperCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Dict , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: List[Any] , *UpperCamelCase_: Tuple ): __lowerCamelCase = CTRLLMHeadModel(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ) = config_and_inputs __lowerCamelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask} return config, inputs_dict def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , *UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = self.num_labels __lowerCamelCase = CTRLForSequenceClassification(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : Any = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () UpperCAmelCase__ : Optional[Any] = (CTRLLMHeadModel,) if is_torch_available() else () UpperCAmelCase__ : int = ( { 'feature-extraction': CTRLModel, 'text-classification': CTRLForSequenceClassification, 'text-generation': CTRLLMHeadModel, 'zero-shot': CTRLForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ : List[str] = True UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : Optional[Any] = False def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Any , UpperCamelCase_: List[str] , UpperCamelCase_: Tuple , UpperCamelCase_: Tuple , UpperCamelCase_: List[str] ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = CTRLModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=UpperCamelCase_ , n_embd=37 ) def lowerCAmelCase__ ( self: Optional[int] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__ ( self: Optional[Any] ): self.config_tester.run_common_tests() def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*UpperCamelCase_ ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*UpperCamelCase_ ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCAmelCase__ ( self: List[Any] ): pass @slow def lowerCAmelCase__ ( self: Optional[Any] ): for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = CTRLModel.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) @unittest.skip("""The model doesn't support left padding""" ) # and it's not used enough to be worth fixing :) def lowerCAmelCase__ ( self: Optional[Any] ): pass @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: List[str] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = CTRLLMHeadModel.from_pretrained("""ctrl""" ) model.to(UpperCamelCase_ ) __lowerCamelCase = torch.tensor( [[1_18_59, 0, 16_11, 8]] , dtype=torch.long , device=UpperCamelCase_ ) # Legal the president is __lowerCamelCase = [ 1_18_59, 0, 16_11, 8, 5, 1_50, 2_64_49, 2, 19, 3_48, 4_69, 3, 25_95, 48, 2_07_40, 24_65_33, 24_65_33, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a __lowerCamelCase = model.generate(UpperCamelCase_ , do_sample=UpperCamelCase_ ) self.assertListEqual(output_ids[0].tolist() , UpperCamelCase_ )
80
1
def lowerCamelCase__ ( A__ : int = 1000 ): '''simple docstring''' __lowerCamelCase = -1 __lowerCamelCase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c __lowerCamelCase = (n * n - 2 * a * n) // (2 * n - 2 * a) __lowerCamelCase = n - a - b if c * c == (a * a + b * b): __lowerCamelCase = a * b * c if candidate >= product: __lowerCamelCase = candidate return product if __name__ == "__main__": print(f"""{solution() = }""")
80
def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0 for i in range(n + 1 )] __lowerCamelCase = 1 __lowerCamelCase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , A__ ): __lowerCamelCase = 1 __lowerCamelCase = 0 for i in range(A__ ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(f"""{solution() = }""")
80
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCAmelCase_ = False class lowerCamelCase__( unittest.TestCase): pass @nightly @require_torch_gpu class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Optional[int] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt="""first prompt""" , image=UpperCamelCase_ , text_to_image_strength=0.75 , generator=UpperCamelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCamelCase_ ) __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained(UpperCamelCase_ , torch_dtype=torch.floataa ) pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) __lowerCamelCase = generator.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt="""first prompt""" , image=UpperCamelCase_ , text_to_image_strength=0.75 , generator=UpperCamelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCAmelCase__ ( self: int ): __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) __lowerCamelCase = """cyberpunk 2077""" __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt=UpperCamelCase_ , image=UpperCamelCase_ , text_to_image_strength=0.75 , generator=UpperCamelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images __lowerCamelCase = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase = np.array([0.1448, 0.1619, 0.1741, 0.1086, 0.1147, 0.1128, 0.1199, 0.1165, 0.1001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = """A painting of a squirrel eating a burger """ __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.text_to_image( prompt=UpperCamelCase_ , generator=UpperCamelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images __lowerCamelCase = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = pipe.image_variation(UpperCamelCase_ , generator=UpperCamelCase_ , output_type="""numpy""" ).images __lowerCamelCase = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase = np.array([0.3076, 0.3123, 0.3284, 0.3782, 0.3770, 0.3894, 0.4297, 0.4331, 0.4456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
80
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase): UpperCAmelCase__ : Dict = 1 @register_to_config def __init__( self: List[str] , UpperCamelCase_: int = 10_00 , UpperCamelCase_: Optional[Union[np.ndarray, List[float]]] = None ): # set `betas`, `alphas`, `timesteps` self.set_timesteps(UpperCamelCase_ ) # standard deviation of the initial noise distribution __lowerCamelCase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowerCamelCase = 4 # running values __lowerCamelCase = [] def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: int , UpperCamelCase_: Union[str, torch.device] = None ): __lowerCamelCase = num_inference_steps __lowerCamelCase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowerCamelCase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowerCamelCase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowerCamelCase = torch.sin(steps * math.pi / 2 ) ** 2 __lowerCamelCase = (1.0 - self.betas**2) ** 0.5 __lowerCamelCase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowerCamelCase = timesteps.to(UpperCamelCase_ ) __lowerCamelCase = [] def lowerCAmelCase__ ( self: int , UpperCamelCase_: torch.FloatTensor , UpperCamelCase_: int , UpperCamelCase_: torch.FloatTensor , UpperCamelCase_: bool = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowerCamelCase = (self.timesteps == timestep).nonzero().item() __lowerCamelCase = timestep_index + 1 __lowerCamelCase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(UpperCamelCase_ ) if len(self.ets ) == 1: __lowerCamelCase = self.ets[-1] elif len(self.ets ) == 2: __lowerCamelCase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowerCamelCase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowerCamelCase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowerCamelCase = self._get_prev_sample(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: torch.FloatTensor , *UpperCamelCase_: Dict , **UpperCamelCase_: Union[str, Any] ): return sample def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Any ): __lowerCamelCase = self.alphas[timestep_index] __lowerCamelCase = self.betas[timestep_index] __lowerCamelCase = self.alphas[prev_timestep_index] __lowerCamelCase = self.betas[prev_timestep_index] __lowerCamelCase = (sample - sigma * ets) / max(UpperCamelCase_ , 1E-8 ) __lowerCamelCase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self: List[Any] ): return self.config.num_train_timesteps
80
1
import math def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = input("""Enter message: """ ) __lowerCamelCase = int(input(f'Enter key [2-{len(A__ ) - 1}]: ' ) ) __lowerCamelCase = input("""Encryption/Decryption [e/d]: """ ) if mode.lower().startswith("""e""" ): __lowerCamelCase = encrypt_message(A__ , A__ ) elif mode.lower().startswith("""d""" ): __lowerCamelCase = decrypt_message(A__ , A__ ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f'Output:\n{text + "|"}' ) def lowerCamelCase__ ( A__ : int , A__ : str ): '''simple docstring''' __lowerCamelCase = [""""""] * key for col in range(A__ ): __lowerCamelCase = col while pointer < len(A__ ): cipher_text[col] += message[pointer] pointer += key return "".join(A__ ) def lowerCamelCase__ ( A__ : int , A__ : str ): '''simple docstring''' __lowerCamelCase = math.ceil(len(A__ ) / key ) __lowerCamelCase = key __lowerCamelCase = (num_cols * num_rows) - len(A__ ) __lowerCamelCase = [""""""] * num_cols __lowerCamelCase = 0 __lowerCamelCase = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): __lowerCamelCase = 0 row += 1 return "".join(A__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
80
import os from collections.abc import Iterator def lowerCamelCase__ ( A__ : str = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(A__ ): __lowerCamelCase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(A__ )[1] in (".py", ".ipynb"): yield os.path.join(A__ , A__ ).lstrip("""./""" ) def lowerCamelCase__ ( A__ : Optional[int] ): '''simple docstring''' return f'{i * " "}*' if i else "\n##" def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(A__ ) or old_parts[i] != new_part) and new_part: print(f'{md_prefix(A__ )} {new_part.replace("_" , " " ).title()}' ) return new_path def lowerCamelCase__ ( A__ : str = "." ): '''simple docstring''' __lowerCamelCase = """""" for filepath in sorted(good_file_paths(A__ ) ): __lowerCamelCase, __lowerCamelCase = os.path.split(A__ ) if filepath != old_path: __lowerCamelCase = print_path(A__ , A__ ) __lowerCamelCase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowerCamelCase = f'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowerCamelCase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(f'{md_prefix(A__ )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md('.')
80
1
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = {'vocab_file': 'vocab.txt', 'emoji_file': 'emoji.json'} UpperCAmelCase_ = { 'vocab_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt', }, 'emoji_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json', }, } UpperCAmelCase_ = { 'abeja/gpt-neox-japanese-2.7b': 2_048, } def lowerCamelCase__ ( A__ : List[str] , A__ : Optional[int] ): '''simple docstring''' with open(A__ , """r""" , encoding="""utf-8""" ) as f: __lowerCamelCase = json.loads(f.read() ) __lowerCamelCase = collections.OrderedDict() __lowerCamelCase = collections.OrderedDict() __lowerCamelCase = collections.OrderedDict() with open(A__ , """r""" , encoding="""utf-8""" ) as f: __lowerCamelCase = f.readlines() __lowerCamelCase = [[t.rstrip("""\n""" )] if (t == """,""" or """,""" not in t) else t.rstrip("""\n""" ).split(""",""" ) for t in token] for idx, b in enumerate(A__ ): __lowerCamelCase = b __lowerCamelCase = idx for wd in b: __lowerCamelCase = idx return vocab, raw_vocab, ids_to_tokens, emoji class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Optional[int] = VOCAB_FILES_NAMES UpperCAmelCase__ : int = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Dict = ['input_ids', 'attention_mask'] def __init__( self: Optional[Any] , UpperCamelCase_: Tuple , UpperCamelCase_: List[str] , UpperCamelCase_: str="<|endoftext|>" , UpperCamelCase_: Dict="<|endoftext|>" , UpperCamelCase_: List[str]="<|startoftext|>" , UpperCamelCase_: List[Any]="<|endoftext|>" , UpperCamelCase_: Optional[Any]=False , **UpperCamelCase_: Tuple , ): super().__init__( unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , do_clean_text=UpperCamelCase_ , **UpperCamelCase_ , ) if not os.path.isfile(UpperCamelCase_ ): raise ValueError( F'Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained' """ model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) if not os.path.isfile(UpperCamelCase_ ): raise ValueError( F'Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google' """ pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) __lowerCamelCase = do_clean_text __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = load_vocab_and_emoji(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def lowerCAmelCase__ ( self: Tuple ): # self.vocab contains support for character fluctuation unique to Japanese, and has a large number of vocab return len(self.raw_vocab ) def lowerCAmelCase__ ( self: List[Any] ): return dict(self.raw_vocab , **self.added_tokens_encoder ) def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: Dict ): return self.subword_tokenizer.tokenize(UpperCamelCase_ , clean=self.do_clean_text ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: Optional[Any] ): return self.vocab.get(UpperCamelCase_ , self.vocab.get(self.unk_token ) ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Any ): return self.subword_tokenizer.convert_id_to_token(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = """""".join(UpperCamelCase_ ).strip() return out_string def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: "Conversation" ): __lowerCamelCase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) + [self.eos_token_id] ) if len(UpperCamelCase_ ) > self.model_max_length: __lowerCamelCase = input_ids[-self.model_max_length :] return input_ids def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str , UpperCamelCase_: Optional[str] = None ): __lowerCamelCase = 0 if os.path.isdir(UpperCamelCase_ ): __lowerCamelCase = os.path.join( UpperCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowerCamelCase = os.path.join( UpperCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""emoji_file"""] ) else: __lowerCamelCase = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowerCamelCase = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""emoji_file"""] ) with open(UpperCamelCase_ , """w""" , encoding="""utf-8""" ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' """ Please check that the vocabulary is not corrupted!""" ) __lowerCamelCase = token_index writer.write(""",""".join(UpperCamelCase_ ) + """\n""" ) index += 1 with open(UpperCamelCase_ , """w""" , encoding="""utf-8""" ) as writer: json.dump(self.emoji , UpperCamelCase_ ) return vocab_file, emoji_file class lowerCamelCase__( __lowerCamelCase): def __init__( self: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Dict , UpperCamelCase_: int ): __lowerCamelCase = vocab # same as swe __lowerCamelCase = ids_to_tokens # same as bpe __lowerCamelCase = emoji __lowerCamelCase = np.max([len(UpperCamelCase_ ) for w in self.vocab.keys()] ) __lowerCamelCase = re.compile(r"""(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)""" ) __lowerCamelCase = re.compile(r"""[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*""" ) __lowerCamelCase = re.compile(r"""[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}""" ) __lowerCamelCase = re.compile( r"""([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __lowerCamelCase = re.compile( r"""(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __lowerCamelCase = re.compile( r"""((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*""" ) __lowerCamelCase = """─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿""" __lowerCamelCase = """▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟""" __lowerCamelCase = str.maketrans({k: """<BLOCK>""" for k in keisen + blocks} ) def __len__( self: int ): return len(self.ids_to_tokens ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: List[Any] ): __lowerCamelCase = self.content_repattera.sub("""<URL>""" , UpperCamelCase_ ) __lowerCamelCase = self.content_repattera.sub("""<EMAIL>""" , UpperCamelCase_ ) __lowerCamelCase = self.content_repattera.sub("""<TEL>""" , UpperCamelCase_ ) __lowerCamelCase = self.content_repattera.sub("""<DATE>""" , UpperCamelCase_ ) __lowerCamelCase = self.content_repattera.sub("""<DATE>""" , UpperCamelCase_ ) __lowerCamelCase = self.content_repattera.sub("""<PRICE>""" , UpperCamelCase_ ) __lowerCamelCase = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: __lowerCamelCase = content.replace("""<BLOCK><BLOCK>""" , """<BLOCK>""" ) return content def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: str=False ): __lowerCamelCase = text.replace(""" """ , """<SP>""" ) __lowerCamelCase = text.replace(""" """ , """<SP>""" ) __lowerCamelCase = text.replace("""\r\n""" , """<BR>""" ) __lowerCamelCase = text.replace("""\n""" , """<BR>""" ) __lowerCamelCase = text.replace("""\r""" , """<BR>""" ) __lowerCamelCase = text.replace("""\t""" , """<TAB>""" ) __lowerCamelCase = text.replace("""—""" , """ー""" ) __lowerCamelCase = text.replace("""−""" , """ー""" ) for k, v in self.emoji["emoji"].items(): if k in text: __lowerCamelCase = text.replace(UpperCamelCase_ , UpperCamelCase_ ) if clean: __lowerCamelCase = self.clean_text(UpperCamelCase_ ) def check_simbol(UpperCamelCase_: int ): __lowerCamelCase = x.encode() if len(UpperCamelCase_ ) == 1 and len(UpperCamelCase_ ) == 2: __lowerCamelCase = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0xc2a1 and c <= 0xc2bf) or (c >= 0xc780 and c <= 0xc783) or (c >= 0xcab9 and c <= 0xcbbf) or (c >= 0xcc80 and c <= 0xcda2) ): return True return False def checkuae(UpperCamelCase_: Optional[Any] ): __lowerCamelCase = x.encode() if len(UpperCamelCase_ ) == 1 and len(UpperCamelCase_ ) == 3: __lowerCamelCase = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0xe28080 and c <= 0xe2b07f: return True return False __lowerCamelCase = 0 __lowerCamelCase = [] while pos < len(UpperCamelCase_ ): __lowerCamelCase = min(len(UpperCamelCase_ ) , pos + self.maxlen + 1 ) if text[pos] == """<""" else pos + 3 __lowerCamelCase = [] # (token_id, token, pos) for e in range(UpperCamelCase_ , UpperCamelCase_ , -1 ): __lowerCamelCase = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(UpperCamelCase_ ) > 2: __lowerCamelCase = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(UpperCamelCase_ ) > 0: # the smallest token_id is adopted __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = sorted(UpperCamelCase_ , key=lambda UpperCamelCase_ : x[0] )[0] result.append(UpperCamelCase_ ) __lowerCamelCase = e else: __lowerCamelCase = pos + 1 __lowerCamelCase = text[pos:end] if check_simbol(UpperCamelCase_ ): result.append("""<KIGOU>""" ) elif checkuae(UpperCamelCase_ ): result.append("""<U2000U2BFF>""" ) else: for i in wd.encode("""utf-8""" ): result.append("""<|byte%d|>""" % i ) __lowerCamelCase = end return result def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any]="\n" ): __lowerCamelCase = [] __lowerCamelCase = [] __lowerCamelCase = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(UpperCamelCase_ ) > 0: words.append(bytearray(UpperCamelCase_ ).decode("""utf-8""" , errors="""replace""" ) ) __lowerCamelCase = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji["""emoji_inv"""][word] ) elif word == "<SP>": words.append(""" """ ) elif word == "<BR>": words.append(UpperCamelCase_ ) elif word == "<TAB>": words.append("""\t""" ) elif word == "<BLOCK>": words.append("""▀""" ) elif word == "<KIGOU>": words.append("""ǀ""" ) elif word == "<U2000U2BFF>": words.append("""‖""" ) else: words.append(UpperCamelCase_ ) if len(UpperCamelCase_ ) > 0: words.append(bytearray(UpperCamelCase_ ).decode("""utf-8""" , errors="""replace""" ) ) __lowerCamelCase = """""".join(UpperCamelCase_ ) return text
80
from __future__ import annotations def lowerCamelCase__ ( A__ : list ): '''simple docstring''' if not nums: raise ValueError("""List is empty""" ) return sum(A__ ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
80
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'adapter_layer': 'encoder.layers.*.adapter_layer', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', 'pooling_layer.linear': 'projector', 'pooling_layer.projection': 'classifier', } UpperCAmelCase_ = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', 'projector', 'classifier', ] def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase = {} with open(A__ , """r""" ) as file: for line_number, line in enumerate(A__ ): __lowerCamelCase = line.strip() if line: __lowerCamelCase = line.split() __lowerCamelCase = line_number __lowerCamelCase = words[0] __lowerCamelCase = value return result def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Tuple , A__ : str , A__ : Union[str, Any] , A__ : Optional[int] ): '''simple docstring''' for attribute in key.split(""".""" ): __lowerCamelCase = getattr(A__ , A__ ) __lowerCamelCase = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(A__ ): __lowerCamelCase = PARAM_MAPPING[full_name.split(""".""" )[-1]] __lowerCamelCase = """param""" if weight_type is not None and weight_type != "param": __lowerCamelCase = getattr(A__ , A__ ).shape elif weight_type is not None and weight_type == "param": __lowerCamelCase = hf_pointer for attribute in hf_param_name.split(""".""" ): __lowerCamelCase = getattr(A__ , A__ ) __lowerCamelCase = shape_pointer.shape # let's reduce dimension __lowerCamelCase = value[0] else: __lowerCamelCase = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": __lowerCamelCase = value elif weight_type == "weight_g": __lowerCamelCase = value elif weight_type == "weight_v": __lowerCamelCase = value elif weight_type == "bias": __lowerCamelCase = value elif weight_type == "param": for attribute in hf_param_name.split(""".""" ): __lowerCamelCase = getattr(A__ , A__ ) __lowerCamelCase = value else: __lowerCamelCase = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def lowerCamelCase__ ( A__ : List[str] , A__ : Union[str, Any] , A__ : str , A__ : List[Any] , A__ : int ): '''simple docstring''' __lowerCamelCase = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(A__ ): __lowerCamelCase = PARAM_MAPPING[full_name.split(""".""" )[-1]] __lowerCamelCase = """param""" if weight_type is not None and weight_type != "param": __lowerCamelCase = """.""".join([key, weight_type] ) elif weight_type is not None and weight_type == "param": __lowerCamelCase = """.""".join([key, hf_param_name] ) else: __lowerCamelCase = key __lowerCamelCase = value if """lm_head""" in full_key else value[0] UpperCAmelCase_ = { 'W_a': 'linear_1.weight', 'W_b': 'linear_2.weight', 'b_a': 'linear_1.bias', 'b_b': 'linear_2.bias', 'ln_W': 'norm.weight', 'ln_b': 'norm.bias', } def lowerCamelCase__ ( A__ : Any , A__ : List[Any] , A__ : Any=None , A__ : Union[str, Any]=None ): '''simple docstring''' __lowerCamelCase = False for key, mapped_key in MAPPING.items(): __lowerCamelCase = """wav2vec2.""" + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: __lowerCamelCase = True if "*" in mapped_key: __lowerCamelCase = name.split(A__ )[0].split(""".""" )[-2] __lowerCamelCase = mapped_key.replace("""*""" , A__ ) if "weight_g" in name: __lowerCamelCase = """weight_g""" elif "weight_v" in name: __lowerCamelCase = """weight_v""" elif "bias" in name: __lowerCamelCase = """bias""" elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCamelCase = """weight""" else: __lowerCamelCase = None if hf_dict is not None: rename_dict(A__ , A__ , A__ , A__ , A__ ) else: set_recursively(A__ , A__ , A__ , A__ , A__ ) return is_used return is_used def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Tuple , A__ : int ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = fairseq_model.state_dict() __lowerCamelCase = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): __lowerCamelCase = False if "conv_layers" in name: load_conv_layer( A__ , A__ , A__ , A__ , hf_model.config.feat_extract_norm == """group""" , ) __lowerCamelCase = True else: __lowerCamelCase = load_wavaveca_layer(A__ , A__ , A__ ) if not is_used: unused_weights.append(A__ ) logger.warning(f'Unused weights: {unused_weights}' ) def lowerCamelCase__ ( A__ : int , A__ : int , A__ : str , A__ : int , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase = full_name.split("""conv_layers.""" )[-1] __lowerCamelCase = name.split(""".""" ) __lowerCamelCase = int(items[0] ) __lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(A__ ) @torch.no_grad() def lowerCamelCase__ ( A__ : List[Any] , A__ : Optional[int] , A__ : Dict=None , A__ : Tuple=None , A__ : List[str]=True , A__ : Optional[int]=False ): '''simple docstring''' if config_path is not None: __lowerCamelCase = WavaVecaConfig.from_pretrained(A__ ) else: __lowerCamelCase = WavaVecaConfig() if is_seq_class: __lowerCamelCase = read_txt_into_dict(A__ ) __lowerCamelCase = idalabel __lowerCamelCase = WavaVecaForSequenceClassification(A__ ) __lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=A__ , return_attention_mask=A__ , ) feature_extractor.save_pretrained(A__ ) elif is_finetuned: if dict_path: __lowerCamelCase = Dictionary.load(A__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowerCamelCase = target_dict.pad_index __lowerCamelCase = target_dict.bos_index __lowerCamelCase = target_dict.eos_index __lowerCamelCase = len(target_dict.symbols ) __lowerCamelCase = os.path.join(A__ , """vocab.json""" ) if not os.path.isdir(A__ ): logger.error("""--pytorch_dump_folder_path ({}) should be a directory""".format(A__ ) ) return os.makedirs(A__ , exist_ok=A__ ) __lowerCamelCase = target_dict.indices # fairseq has the <pad> and <s> switched __lowerCamelCase = 0 __lowerCamelCase = 1 with open(A__ , """w""" , encoding="""utf-8""" ) as vocab_handle: json.dump(A__ , A__ ) __lowerCamelCase = WavaVecaCTCTokenizer( A__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="""|""" , do_lower_case=A__ , ) __lowerCamelCase = True if config.feat_extract_norm == """layer""" else False __lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=A__ , return_attention_mask=A__ , ) __lowerCamelCase = WavaVecaProcessor(feature_extractor=A__ , tokenizer=A__ ) processor.save_pretrained(A__ ) __lowerCamelCase = WavaVecaForCTC(A__ ) else: __lowerCamelCase = WavaVecaForPreTraining(A__ ) if is_finetuned or is_seq_class: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) else: __lowerCamelCase = argparse.Namespace(task="""audio_pretraining""" ) __lowerCamelCase = fairseq.tasks.setup_task(A__ ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=A__ ) __lowerCamelCase = model[0].eval() recursively_load_weights(A__ , A__ , not is_finetuned ) hf_wavavec.save_pretrained(A__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) parser.add_argument( '--is_seq_class', action='store_true', help='Whether the model to convert is a fine-tuned sequence classification model or not', ) UpperCAmelCase_ = parser.parse_args() UpperCAmelCase_ = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
80
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase): UpperCAmelCase__ : Any = 'maskformer-swin' UpperCAmelCase__ : List[Any] = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Any , UpperCamelCase_: Any=2_24 , UpperCamelCase_: List[str]=4 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: Optional[int]=96 , UpperCamelCase_: List[str]=[2, 2, 6, 2] , UpperCamelCase_: Optional[Any]=[3, 6, 12, 24] , UpperCamelCase_: str=7 , UpperCamelCase_: int=4.0 , UpperCamelCase_: Optional[int]=True , UpperCamelCase_: Union[str, Any]=0.0 , UpperCamelCase_: Optional[int]=0.0 , UpperCamelCase_: Any=0.1 , UpperCamelCase_: Union[str, Any]="gelu" , UpperCamelCase_: int=False , UpperCamelCase_: Optional[int]=0.02 , UpperCamelCase_: Optional[Any]=1E-5 , UpperCamelCase_: Optional[int]=None , UpperCamelCase_: List[Any]=None , **UpperCamelCase_: Union[str, Any] , ): super().__init__(**UpperCamelCase_ ) __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = embed_dim __lowerCamelCase = depths __lowerCamelCase = len(UpperCamelCase_ ) __lowerCamelCase = num_heads __lowerCamelCase = window_size __lowerCamelCase = mlp_ratio __lowerCamelCase = qkv_bias __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_absolute_embeddings __lowerCamelCase = layer_norm_eps __lowerCamelCase = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCamelCase = int(embed_dim * 2 ** (len(UpperCamelCase_ ) - 1) ) __lowerCamelCase = ["""stem"""] + [F'stage{idx}' for idx in range(1 , len(UpperCamelCase_ ) + 1 )] __lowerCamelCase, __lowerCamelCase = get_aligned_output_features_output_indices( out_features=UpperCamelCase_ , out_indices=UpperCamelCase_ , stage_names=self.stage_names )
80
1
UpperCAmelCase_ = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' UpperCAmelCase_ = [{'type': 'code', 'content': INSTALL_CONTENT}] UpperCAmelCase_ = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
80
from __future__ import annotations def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): __lowerCamelCase, __lowerCamelCase = array[indexa], array[indexa] def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if length > 1: __lowerCamelCase = int(length / 2 ) for i in range(A__ , low + middle ): comp_and_swap(A__ , A__ , i + middle , A__ ) bitonic_merge(A__ , A__ , A__ , A__ ) bitonic_merge(A__ , low + middle , A__ , A__ ) def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if length > 1: __lowerCamelCase = int(length / 2 ) bitonic_sort(A__ , A__ , A__ , 1 ) bitonic_sort(A__ , low + middle , A__ , 0 ) bitonic_merge(A__ , A__ , A__ , A__ ) if __name__ == "__main__": UpperCAmelCase_ = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase_ = [int(item.strip()) for item in user_input.split(',')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('\nSorted array in ascending order is: ', end='') print(*unsorted, sep=', ') bitonic_merge(unsorted, 0, len(unsorted), 0) print('Sorted array in descending order is: ', end='') print(*unsorted, sep=', ')
80
1
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class lowerCamelCase__( _lowerCamelCase): def __init__( self: List[Any] , UpperCamelCase_: List[Any] , UpperCamelCase_: str=13 , UpperCamelCase_: Tuple=7 , UpperCamelCase_: Dict=True , UpperCamelCase_: List[str]=True , UpperCamelCase_: Optional[int]=False , UpperCamelCase_: Optional[int]=True , UpperCamelCase_: str=99 , UpperCamelCase_: int=32 , UpperCamelCase_: Tuple=5 , UpperCamelCase_: Any=4 , UpperCamelCase_: List[Any]=64 , UpperCamelCase_: int="gelu" , UpperCamelCase_: Union[str, Any]=0.1 , UpperCamelCase_: List[str]=0.1 , UpperCamelCase_: Union[str, Any]=5_12 , UpperCamelCase_: Any=16 , UpperCamelCase_: Optional[int]=2 , UpperCamelCase_: List[str]=0.02 , UpperCamelCase_: str=3 , UpperCamelCase_: Optional[Any]=4 , UpperCamelCase_: Union[str, Any]=None , UpperCamelCase_: Dict=2 , UpperCamelCase_: List[Any]=2 , UpperCamelCase_: Optional[int]=2 , UpperCamelCase_: Tuple=2 , UpperCamelCase_: Optional[int]=4 , UpperCamelCase_: Union[str, Any]=1 , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = q_groups __lowerCamelCase = k_groups __lowerCamelCase = v_groups __lowerCamelCase = post_attention_groups __lowerCamelCase = intermediate_groups __lowerCamelCase = output_groups def lowerCAmelCase__ ( self: str ): __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase__ ( self: Tuple ): return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: str , UpperCamelCase_: List[str] , UpperCamelCase_: Optional[int] , UpperCamelCase_: Tuple , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: List[Any] ): __lowerCamelCase = SqueezeBertModel(config=A__ ) model.to(A__ ) model.eval() __lowerCamelCase = model(A__ , A__ ) __lowerCamelCase = model(A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Optional[int] , UpperCamelCase_: Optional[int] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Tuple , UpperCamelCase_: str ): __lowerCamelCase = SqueezeBertForMaskedLM(config=A__ ) model.to(A__ ) model.eval() __lowerCamelCase = model(A__ , attention_mask=A__ , labels=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: Dict , UpperCamelCase_: str , UpperCamelCase_: Optional[int] , UpperCamelCase_: int , UpperCamelCase_: List[Any] , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = SqueezeBertForQuestionAnswering(config=A__ ) model.to(A__ ) model.eval() __lowerCamelCase = model( A__ , attention_mask=A__ , start_positions=A__ , end_positions=A__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: List[Any] , UpperCamelCase_: str , UpperCamelCase_: Optional[int] , UpperCamelCase_: str , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Tuple ): __lowerCamelCase = self.num_labels __lowerCamelCase = SqueezeBertForSequenceClassification(A__ ) model.to(A__ ) model.eval() __lowerCamelCase = model(A__ , attention_mask=A__ , labels=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Dict , UpperCamelCase_: Optional[int] , UpperCamelCase_: Any , UpperCamelCase_: List[str] , UpperCamelCase_: Any , UpperCamelCase_: Optional[int] ): __lowerCamelCase = self.num_labels __lowerCamelCase = SqueezeBertForTokenClassification(config=A__ ) model.to(A__ ) model.eval() __lowerCamelCase = model(A__ , attention_mask=A__ , labels=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: List[str] , UpperCamelCase_: List[str] , UpperCamelCase_: List[Any] , UpperCamelCase_: Optional[int] , UpperCamelCase_: Any , UpperCamelCase_: Dict ): __lowerCamelCase = self.num_choices __lowerCamelCase = SqueezeBertForMultipleChoice(config=A__ ) model.to(A__ ) model.eval() __lowerCamelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCamelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCamelCase = model( A__ , attention_mask=A__ , labels=A__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.prepare_config_and_inputs() ((__lowerCamelCase), (__lowerCamelCase), (__lowerCamelCase), (__lowerCamelCase), (__lowerCamelCase), (__lowerCamelCase)) = config_and_inputs __lowerCamelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowerCamelCase__( _lowerCamelCase , _lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : Any = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) UpperCAmelCase__ : Dict = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : List[str] = True UpperCAmelCase__ : str = False def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = SqueezeBertModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=A__ , dim=37 ) def lowerCAmelCase__ ( self: Any ): self.config_tester.run_common_tests() def lowerCAmelCase__ ( self: int ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*A__ ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*A__ ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*A__ ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*A__ ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*A__ ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*A__ ) @slow def lowerCAmelCase__ ( self: str ): for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = SqueezeBertModel.from_pretrained(A__ ) self.assertIsNotNone(A__ ) @require_sentencepiece @require_tokenizers @require_torch class lowerCamelCase__( unittest.TestCase): @slow def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = SqueezeBertForSequenceClassification.from_pretrained("""squeezebert/squeezebert-mnli""" ) __lowerCamelCase = torch.tensor([[1, 2_94_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 13, 15_88, 2]] ) __lowerCamelCase = model(A__ )[0] __lowerCamelCase = torch.Size((1, 3) ) self.assertEqual(output.shape , A__ ) __lowerCamelCase = torch.tensor([[0.6401, -0.0349, -0.6041]] ) self.assertTrue(torch.allclose(A__ , A__ , atol=1E-4 ) )
700
from ... import PretrainedConfig UpperCAmelCase_ = { 'sijunhe/nezha-cn-base': 'https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json', } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Dict = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP UpperCAmelCase__ : Dict = 'nezha' def __init__( self: Dict , UpperCamelCase_: Any=2_11_28 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Optional[int]=12 , UpperCamelCase_: List[str]=12 , UpperCamelCase_: Optional[int]=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: str=0.1 , UpperCamelCase_: Union[str, Any]=5_12 , UpperCamelCase_: Any=64 , UpperCamelCase_: Dict=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: Optional[Any]=1E-12 , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Any=0 , UpperCamelCase_: str=2 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: str=True , **UpperCamelCase_: Any , ): super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = max_relative_position __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = classifier_dropout __lowerCamelCase = use_cache
80
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'google/fnet-base': 'https://huggingface.co/google/fnet-base/resolve/main/config.json', 'google/fnet-large': 'https://huggingface.co/google/fnet-large/resolve/main/config.json' # See all FNet models at https://huggingface.co/models?filter=fnet } class lowerCamelCase__( _snake_case): UpperCAmelCase__ : Union[str, Any] = 'fnet' def __init__( self: Union[str, Any] , UpperCamelCase_: List[str]=3_20_00 , UpperCamelCase_: Tuple=7_68 , UpperCamelCase_: int=12 , UpperCamelCase_: int=30_72 , UpperCamelCase_: Optional[Any]="gelu_new" , UpperCamelCase_: Dict=0.1 , UpperCamelCase_: Any=5_12 , UpperCamelCase_: List[str]=4 , UpperCamelCase_: Union[str, Any]=0.02 , UpperCamelCase_: List[Any]=1E-12 , UpperCamelCase_: Dict=False , UpperCamelCase_: Dict=5_12 , UpperCamelCase_: str=3 , UpperCamelCase_: Tuple=1 , UpperCamelCase_: List[Any]=2 , **UpperCamelCase_: Optional[Any] , ): super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) __lowerCamelCase = vocab_size __lowerCamelCase = max_position_embeddings __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = type_vocab_size __lowerCamelCase = layer_norm_eps __lowerCamelCase = use_tpu_fourier_optimizations __lowerCamelCase = tpu_short_seq_length
701
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__: def __init__( self: Union[str, Any] , UpperCamelCase_: str = None , UpperCamelCase_: uuid.UUID = None , UpperCamelCase_: Dict=None , UpperCamelCase_: Any=None ): if not conversation_id: __lowerCamelCase = uuid.uuida() if past_user_inputs is None: __lowerCamelCase = [] if generated_responses is None: __lowerCamelCase = [] __lowerCamelCase = conversation_id __lowerCamelCase = past_user_inputs __lowerCamelCase = generated_responses __lowerCamelCase = text def __eq__( self: Optional[Any] , UpperCamelCase_: Union[str, Any] ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: str , UpperCamelCase_: bool = False ): if self.new_user_input: if overwrite: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' F'with: "{text}".' ) __lowerCamelCase = text else: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' F'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input' ) else: __lowerCamelCase = text def lowerCAmelCase__ ( self: List[str] ): if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __lowerCamelCase = None def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): self.generated_responses.append(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple ): for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self: Union[str, Any] ): __lowerCamelCase = F'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): __lowerCamelCase = """user""" if is_user else """bot""" output += F'{name} >> {text} \n' return output @add_end_docstrings( __lowerCamelCase , r'\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n ' , ) class lowerCamelCase__( __lowerCamelCase): def __init__( self: List[str] , *UpperCamelCase_: List[Any] , **UpperCamelCase_: str ): super().__init__(*UpperCamelCase_ , **UpperCamelCase_ ) if self.tokenizer.pad_token_id is None: __lowerCamelCase = self.tokenizer.eos_token def lowerCAmelCase__ ( self: str , UpperCamelCase_: int=None , UpperCamelCase_: Any=None , UpperCamelCase_: Union[str, Any]=None , **UpperCamelCase_: int ): __lowerCamelCase = {} __lowerCamelCase = {} __lowerCamelCase = {} if min_length_for_response is not None: __lowerCamelCase = min_length_for_response if minimum_tokens is not None: __lowerCamelCase = minimum_tokens if "max_length" in generate_kwargs: __lowerCamelCase = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __lowerCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCamelCase_ ) return preprocess_params, forward_params, postprocess_params def __call__( self: Any , UpperCamelCase_: Union[Conversation, List[Conversation]] , UpperCamelCase_: Optional[int]=0 , **UpperCamelCase_: Optional[int] ): __lowerCamelCase = super().__call__(UpperCamelCase_ , num_workers=UpperCamelCase_ , **UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and len(UpperCamelCase_ ) == 1: return outputs[0] return outputs def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Conversation , UpperCamelCase_: Optional[Any]=32 ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( F'Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ' """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): __lowerCamelCase = self.tokenizer._build_conversation_input_ids(UpperCamelCase_ ) else: # If the tokenizer cannot handle conversations, we default to only the old version __lowerCamelCase = self._legacy_parse_and_tokenize(UpperCamelCase_ ) if self.framework == "pt": __lowerCamelCase = torch.LongTensor([input_ids] ) elif self.framework == "tf": __lowerCamelCase = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str=10 , **UpperCamelCase_: List[str] ): __lowerCamelCase = generate_kwargs.get("""max_length""" , self.model.config.max_length ) __lowerCamelCase = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(F'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})' ) __lowerCamelCase = max_length - minimum_tokens __lowerCamelCase = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: __lowerCamelCase = model_inputs["""attention_mask"""][:, -trim:] __lowerCamelCase = model_inputs.pop("""conversation""" ) __lowerCamelCase = max_length __lowerCamelCase = self.model.generate(**UpperCamelCase_ , **UpperCamelCase_ ) if self.model.config.is_encoder_decoder: __lowerCamelCase = 1 else: __lowerCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Optional[Any] , UpperCamelCase_: int=True ): __lowerCamelCase = model_outputs["""output_ids"""] __lowerCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_ , ) __lowerCamelCase = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCamelCase_ ) return conversation def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Conversation ): __lowerCamelCase = self.tokenizer.eos_token_id __lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) ) if len(UpperCamelCase_ ) > self.tokenizer.model_max_length: __lowerCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
80
0
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class lowerCamelCase__( tf.keras.layers.Layer): def __init__( self: List[Any] , UpperCamelCase_: Dict[str, int] , UpperCamelCase_: List[str] , UpperCamelCase_: int = None , UpperCamelCase_: int = None ): super().__init__() __lowerCamelCase = pad_token_id __lowerCamelCase = max_length __lowerCamelCase = vocab __lowerCamelCase = merges __lowerCamelCase = BytePairTokenizer(UpperCAmelCase_ , UpperCAmelCase_ , sequence_length=UpperCAmelCase_ ) @classmethod def lowerCAmelCase__ ( cls: Tuple , UpperCamelCase_: GPTaTokenizer , *UpperCamelCase_: List[Any] , **UpperCamelCase_: List[Any] ): __lowerCamelCase = [""" """.join(UpperCAmelCase_ ) for m in tokenizer.bpe_ranks.keys()] __lowerCamelCase = tokenizer.get_vocab() return cls(UpperCAmelCase_ , UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_ ) @classmethod def lowerCAmelCase__ ( cls: Union[str, Any] , UpperCamelCase_: Union[str, os.PathLike] , *UpperCamelCase_: Optional[Any] , **UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = GPTaTokenizer.from_pretrained(UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_ ) return cls.from_tokenizer(UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_ ) @classmethod def lowerCAmelCase__ ( cls: List[str] , UpperCamelCase_: Union[str, Any] ): return cls(**UpperCAmelCase_ ) def lowerCAmelCase__ ( self: Dict ): return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Any , UpperCamelCase_: int = None ): __lowerCamelCase = self.tf_tokenizer(UpperCAmelCase_ ) __lowerCamelCase = tf.ones_like(UpperCAmelCase_ ) if self.pad_token_id is not None: # pad the tokens up to max length __lowerCamelCase = max_length if max_length is not None else self.max_length if max_length is not None: __lowerCamelCase, __lowerCamelCase = pad_model_inputs( UpperCAmelCase_ , max_seq_length=UpperCAmelCase_ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
702
import math def lowerCamelCase__ ( A__ : int ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = 2 __lowerCamelCase = int(math.sqrt(A__ ) ) # Size of every segment __lowerCamelCase = [True] * (end + 1) __lowerCamelCase = [] while start <= end: if temp[start] is True: in_prime.append(A__ ) for i in range(start * start , end + 1 , A__ ): __lowerCamelCase = False start += 1 prime += in_prime __lowerCamelCase = end + 1 __lowerCamelCase = min(2 * end , A__ ) while low <= n: __lowerCamelCase = [True] * (high - low + 1) for each in in_prime: __lowerCamelCase = math.floor(low / each ) * each if t < low: t += each for j in range(A__ , high + 1 , A__ ): __lowerCamelCase = False for j in range(len(A__ ) ): if temp[j] is True: prime.append(j + low ) __lowerCamelCase = high + 1 __lowerCamelCase = min(high + end , A__ ) return prime print(sieve(10**6))
80
0
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def lowerCamelCase__ ( A__ : Tuple , A__ : str , A__ : List[str] ): '''simple docstring''' __lowerCamelCase = ("""dense.weight""", """attention.self.query""", """attention.self.key""", """attention.self.value""") __lowerCamelCase = ( ("""layer.""", """layer_"""), ("""word_embeddings.weight""", """word_embeddings"""), ("""position_embeddings.weight""", """position_embeddings"""), ("""token_type_embeddings.weight""", """token_type_embeddings"""), (""".""", """/"""), ("""LayerNorm/weight""", """LayerNorm/gamma"""), ("""LayerNorm/bias""", """LayerNorm/beta"""), ("""weight""", """kernel"""), ) if not os.path.isdir(A__ ): os.makedirs(A__ ) __lowerCamelCase = model.state_dict() def to_tf_var_name(A__ : Dict ): for patt, repl in iter(A__ ): __lowerCamelCase = name.replace(A__ , A__ ) return f'bert/{name}' def create_tf_var(A__ : Optional[int] , A__ : int , A__ : int ): __lowerCamelCase = tf.dtypes.as_dtype(tensor.dtype ) __lowerCamelCase = tf.get_variable(dtype=A__ , shape=tensor.shape , name=A__ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(A__ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: __lowerCamelCase = to_tf_var_name(A__ ) __lowerCamelCase = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): __lowerCamelCase = torch_tensor.T __lowerCamelCase = create_tf_var(tensor=A__ , name=A__ , session=A__ ) tf.keras.backend.set_value(A__ , A__ ) __lowerCamelCase = session.run(A__ ) print(f'Successfully created {tf_name}: {np.allclose(A__ , A__ )}' ) __lowerCamelCase = tf.train.Saver(tf.trainable_variables() ) saver.save(A__ , os.path.join(A__ , model_name.replace("""-""" , """_""" ) + """.ckpt""" ) ) def lowerCamelCase__ ( A__ : List[str]=None ): '''simple docstring''' __lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--model_name""" , type=A__ , required=A__ , help="""model name e.g. bert-base-uncased""" ) parser.add_argument( """--cache_dir""" , type=A__ , default=A__ , required=A__ , help="""Directory containing pytorch model""" ) parser.add_argument("""--pytorch_model_path""" , type=A__ , required=A__ , help="""/path/to/<pytorch-model-name>.bin""" ) parser.add_argument("""--tf_cache_dir""" , type=A__ , required=A__ , help="""Directory in which to save tensorflow model""" ) __lowerCamelCase = parser.parse_args(A__ ) __lowerCamelCase = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=A__ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
703
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase_ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') class lowerCamelCase__( __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : int = BartphoTokenizer UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : List[str] = True def lowerCAmelCase__ ( self: Tuple ): super().setUp() __lowerCamelCase = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] __lowerCamelCase = dict(zip(UpperCamelCase_ , range(len(UpperCamelCase_ ) ) ) ) __lowerCamelCase = {"""unk_token""": """<unk>"""} __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F'{token} {vocab_tokens[token]}\n' ) __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase__ ( self: List[str] , **UpperCamelCase_: List[str] ): kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): __lowerCamelCase = """This is a là test""" __lowerCamelCase = """This is a<unk><unk> test""" return input_text, output_text def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) __lowerCamelCase = """This is a là test""" __lowerCamelCase = """▁This ▁is ▁a ▁l à ▁t est""".split() __lowerCamelCase = tokenizer.tokenize(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase_ ) , UpperCamelCase_ )
80
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCAmelCase_ = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
704
def lowerCamelCase__ ( A__ : dict ): '''simple docstring''' __lowerCamelCase = set() # To detect a back edge, keep track of vertices currently in the recursion stack __lowerCamelCase = set() return any( node not in visited and depth_first_search(A__ , A__ , A__ , A__ ) for node in graph ) def lowerCamelCase__ ( A__ : dict , A__ : int , A__ : set , A__ : set ): '''simple docstring''' visited.add(A__ ) rec_stk.add(A__ ) for node in graph[vertex]: if node not in visited: if depth_first_search(A__ , A__ , A__ , A__ ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(A__ ) return False if __name__ == "__main__": from doctest import testmod testmod()
80
0
import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Optional[Any] ): return F'gaussian_noise_s={seed}_shape={"_".join([str(__lowerCamelCase ) for s in shape] )}.npy' def lowerCAmelCase__ ( self: Tuple ): super().tearDown() gc.collect() def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: List[Any]=0 , UpperCamelCase_: List[Any]=(4, 4, 64, 64) , UpperCamelCase_: Dict=False ): __lowerCamelCase = jnp.bfloataa if fpaa else jnp.floataa __lowerCamelCase = jnp.array(load_hf_numpy(self.get_file_format(__lowerCamelCase , __lowerCamelCase ) ) , dtype=__lowerCamelCase ) return image def lowerCAmelCase__ ( self: str , UpperCamelCase_: Dict=False , UpperCamelCase_: Optional[Any]="CompVis/stable-diffusion-v1-4" ): __lowerCamelCase = jnp.bfloataa if fpaa else jnp.floataa __lowerCamelCase = "bf16" if fpaa else None __lowerCamelCase = FlaxUNetaDConditionModel.from_pretrained( __lowerCamelCase , subfolder="""unet""" , dtype=__lowerCamelCase , revision=__lowerCamelCase ) return model, params def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Tuple=0 , UpperCamelCase_: List[Any]=(4, 77, 7_68) , UpperCamelCase_: Dict=False ): __lowerCamelCase = jnp.bfloataa if fpaa else jnp.floataa __lowerCamelCase = jnp.array(load_hf_numpy(self.get_file_format(__lowerCamelCase , __lowerCamelCase ) ) , dtype=__lowerCamelCase ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2323, -0.1304, 0.0813, -0.3093, -0.0919, -0.1571, -0.1125, -0.5806]], [17, 0.55, [-0.0831, -0.2443, 0.0901, -0.0919, 0.3396, 0.0103, -0.3743, 0.0701]], [8, 0.89, [-0.4863, 0.0859, 0.0875, -0.1658, 0.9199, -0.0114, 0.4839, 0.4639]], [3, 10_00, [-0.5649, 0.2402, -0.5518, 0.1248, 1.1328, -0.2443, -0.0325, -1.0078]], # fmt: on ] ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: List[str] , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = self.get_unet_model(model_id="""CompVis/stable-diffusion-v1-4""" , fpaa=__lowerCamelCase ) __lowerCamelCase = self.get_latents(__lowerCamelCase , fpaa=__lowerCamelCase ) __lowerCamelCase = self.get_encoder_hidden_states(__lowerCamelCase , fpaa=__lowerCamelCase ) __lowerCamelCase = model.apply( {"""params""": params} , __lowerCamelCase , jnp.array(__lowerCamelCase , dtype=jnp.intaa ) , encoder_hidden_states=__lowerCamelCase , ).sample assert sample.shape == latents.shape __lowerCamelCase = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) __lowerCamelCase = jnp.array(__lowerCamelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(__lowerCamelCase , __lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1514, 0.0807, 0.1624, 0.1016, -0.1896, 0.0263, 0.0677, 0.2310]], [17, 0.55, [0.1164, -0.0216, 0.0170, 0.1589, -0.3120, 0.1005, -0.0581, -0.1458]], [8, 0.89, [-0.1758, -0.0169, 0.1004, -0.1411, 0.1312, 0.1103, -0.1996, 0.2139]], [3, 10_00, [0.1214, 0.0352, -0.0731, -0.1562, -0.0994, -0.0906, -0.2340, -0.0539]], # fmt: on ] ) def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: int , UpperCamelCase_: Dict , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = self.get_unet_model(model_id="""stabilityai/stable-diffusion-2""" , fpaa=__lowerCamelCase ) __lowerCamelCase = self.get_latents(__lowerCamelCase , shape=(4, 4, 96, 96) , fpaa=__lowerCamelCase ) __lowerCamelCase = self.get_encoder_hidden_states(__lowerCamelCase , shape=(4, 77, 10_24) , fpaa=__lowerCamelCase ) __lowerCamelCase = model.apply( {"""params""": params} , __lowerCamelCase , jnp.array(__lowerCamelCase , dtype=jnp.intaa ) , encoder_hidden_states=__lowerCamelCase , ).sample assert sample.shape == latents.shape __lowerCamelCase = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) __lowerCamelCase = jnp.array(__lowerCamelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(__lowerCamelCase , __lowerCamelCase , atol=1E-2 )
705
from __future__ import annotations def lowerCamelCase__ ( A__ : list[float] , A__ : list[float] ): '''simple docstring''' __lowerCamelCase = sorted(numsa + numsa ) __lowerCamelCase, __lowerCamelCase = divmod(len(A__ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ = [float(x) for x in input('Enter the elements of first array: ').split()] UpperCAmelCase_ = [float(x) for x in input('Enter the elements of second array: ').split()] print(f"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
80
0
def lowerCamelCase__ ( ): '''simple docstring''' return [list(range(1000 - i , -1000 - i , -1 ) ) for i in range(1000 )] UpperCAmelCase_ = generate_large_matrix() UpperCAmelCase_ = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def lowerCamelCase__ ( A__ : Union[str, Any] ): '''simple docstring''' assert all(row == sorted(lowerCamelCase_ , reverse=lowerCamelCase_ ) for row in grid ) assert all(list(lowerCamelCase_ ) == sorted(lowerCamelCase_ , reverse=lowerCamelCase_ ) for col in zip(*lowerCamelCase_ ) ) def lowerCamelCase__ ( A__ : Optional[int] ): '''simple docstring''' __lowerCamelCase = 0 __lowerCamelCase = len(lowerCamelCase_ ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: __lowerCamelCase = (left + right) // 2 __lowerCamelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: __lowerCamelCase = mid + 1 else: __lowerCamelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(lowerCamelCase_ ) def lowerCamelCase__ ( A__ : Any ): '''simple docstring''' __lowerCamelCase = 0 __lowerCamelCase = len(grid[0] ) for i in range(len(lowerCamelCase_ ) ): __lowerCamelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(lowerCamelCase_ ) * len(grid[0] )) - total def lowerCamelCase__ ( A__ : List[str] ): '''simple docstring''' return len([number for row in grid for number in row if number < 0] ) def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = 0 for row in grid: for i, number in enumerate(lowerCamelCase_ ): if number < 0: total += len(lowerCamelCase_ ) - i break return total def lowerCamelCase__ ( ): '''simple docstring''' from timeit import timeit print("""Running benchmarks""" ) __lowerCamelCase = ( 'from __main__ import count_negatives_binary_search, ' 'count_negatives_brute_force, count_negatives_brute_force_with_break, grid' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): __lowerCamelCase = timeit(f'{func}(grid=grid)' , setup=lowerCamelCase_ , number=500 ) print(f'{func}() took {time:0.4f} seconds' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
706
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) self.assertTrue(torch.allclose(gelu_python(UpperCamelCase_ ) , torch_builtin(UpperCamelCase_ ) ) ) self.assertFalse(torch.allclose(gelu_python(UpperCamelCase_ ) , gelu_new(UpperCamelCase_ ) ) ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = get_activation("""gelu_10""" ) __lowerCamelCase = torch_builtin(UpperCamelCase_ ) __lowerCamelCase = geluaa(UpperCamelCase_ ) __lowerCamelCase = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(UpperCamelCase_ ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def lowerCAmelCase__ ( self: str ): get_activation("""gelu""" ) get_activation("""gelu_10""" ) get_activation("""gelu_fast""" ) get_activation("""gelu_new""" ) get_activation("""gelu_python""" ) get_activation("""gelu_pytorch_tanh""" ) get_activation("""linear""" ) get_activation("""mish""" ) get_activation("""quick_gelu""" ) get_activation("""relu""" ) get_activation("""sigmoid""" ) get_activation("""silu""" ) get_activation("""swish""" ) get_activation("""tanh""" ) with self.assertRaises(UpperCamelCase_ ): get_activation("""bogus""" ) with self.assertRaises(UpperCamelCase_ ): get_activation(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = 1 __lowerCamelCase = get_activation("""gelu""" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(UpperCamelCase_ ): __lowerCamelCase = acta.a
80
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCamelCase__( UpperCAmelCase_ , unittest.TestCase): UpperCAmelCase__ : Dict = ShapEPipeline UpperCAmelCase__ : List[Any] = ['prompt'] UpperCAmelCase__ : Tuple = ['prompt'] UpperCAmelCase__ : Union[str, Any] = [ 'num_images_per_prompt', 'num_inference_steps', 'generator', 'latents', 'guidance_scale', 'frame_size', 'output_type', 'return_dict', ] UpperCAmelCase__ : int = False @property def lowerCAmelCase__ ( self: str ): return 32 @property def lowerCAmelCase__ ( self: str ): return 32 @property def lowerCAmelCase__ ( self: Tuple ): return self.time_input_dim * 4 @property def lowerCAmelCase__ ( self: Any ): return 8 @property def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) return tokenizer @property def lowerCAmelCase__ ( self: List[Any] ): torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModelWithProjection(_lowercase ) @property def lowerCAmelCase__ ( self: Tuple ): torch.manual_seed(0 ) __lowerCamelCase = { 'num_attention_heads': 2, 'attention_head_dim': 16, 'embedding_dim': self.time_input_dim, 'num_embeddings': 32, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } __lowerCamelCase = PriorTransformer(**_lowercase ) return model @property def lowerCAmelCase__ ( self: List[str] ): torch.manual_seed(0 ) __lowerCamelCase = { 'param_shapes': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 12, 'background': ( 0.1, 0.1, 0.1, ), } __lowerCamelCase = ShapERenderer(**_lowercase ) return model def lowerCAmelCase__ ( self: str ): __lowerCamelCase = self.dummy_prior __lowerCamelCase = self.dummy_text_encoder __lowerCamelCase = self.dummy_tokenizer __lowerCamelCase = self.dummy_renderer __lowerCamelCase = HeunDiscreteScheduler( beta_schedule="""exp""" , num_train_timesteps=10_24 , prediction_type="""sample""" , use_karras_sigmas=_lowercase , clip_sample=_lowercase , clip_sample_range=1.0 , ) __lowerCamelCase = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: List[Any]=0 ): if str(_lowercase ).startswith("""mps""" ): __lowerCamelCase = torch.manual_seed(_lowercase ) else: __lowerCamelCase = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) __lowerCamelCase = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 32, 'output_type': 'np', } return inputs def lowerCAmelCase__ ( self: int ): __lowerCamelCase = 'cpu' __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = self.pipeline_class(**_lowercase ) __lowerCamelCase = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __lowerCamelCase = pipe(**self.get_dummy_inputs(_lowercase ) ) __lowerCamelCase = output.images[0] __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __lowerCamelCase = np.array( [ 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase__ ( self: Any ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch_device == 'cpu' __lowerCamelCase = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowercase , relax_max_difference=_lowercase , ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = self.pipeline_class(**_lowercase ) __lowerCamelCase = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __lowerCamelCase = 1 __lowerCamelCase = 2 __lowerCamelCase = self.get_dummy_inputs(_lowercase ) for key in inputs.keys(): if key in self.batch_params: __lowerCamelCase = batch_size * [inputs[key]] __lowerCamelCase = pipe(**_lowercase , num_images_per_prompt=_lowercase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/shap_e/test_shap_e_np_out.npy""" ) __lowerCamelCase = ShapEPipeline.from_pretrained("""openai/shap-e""" ) __lowerCamelCase = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __lowerCamelCase = torch.Generator(device=_lowercase ).manual_seed(0 ) __lowerCamelCase = pipe( """a shark""" , generator=_lowercase , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type="""np""" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(_lowercase , _lowercase )
707
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class lowerCamelCase__( __lowerCamelCase): @slow @require_torch def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = EncoderDecoderModel.from_encoder_decoder_pretrained("""prajjwal1/bert-tiny""" , """prajjwal1/bert-tiny""" ) __lowerCamelCase = BertTokenizer.from_pretrained("""bert-base-uncased""" ) __lowerCamelCase = bertabert.config.encoder.vocab_size __lowerCamelCase = tokenizer.sep_token_id __lowerCamelCase = tokenizer.cls_token_id __lowerCamelCase = 1_28 __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""train[:1%]""" ) __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""validation[:1%]""" ) __lowerCamelCase = train_dataset.select(range(32 ) ) __lowerCamelCase = val_dataset.select(range(16 ) ) __lowerCamelCase = 4 def _map_to_encoder_decoder_inputs(UpperCamelCase_: List[Any] ): # Tokenizer will automatically set [BOS] <text> [EOS] __lowerCamelCase = tokenizer(batch["""article"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=5_12 ) __lowerCamelCase = tokenizer(batch["""highlights"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=1_28 ) __lowerCamelCase = inputs.input_ids __lowerCamelCase = inputs.attention_mask __lowerCamelCase = outputs.input_ids __lowerCamelCase = outputs.input_ids.copy() __lowerCamelCase = [ [-1_00 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["""labels"""] ] __lowerCamelCase = outputs.attention_mask assert all(len(UpperCamelCase_ ) == 5_12 for x in inputs.input_ids ) assert all(len(UpperCamelCase_ ) == 1_28 for x in outputs.input_ids ) return batch def _compute_metrics(UpperCamelCase_: int ): __lowerCamelCase = pred.label_ids __lowerCamelCase = pred.predictions # all unnecessary tokens are removed __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = sum([int(pred_str[i] == label_str[i] ) for i in range(len(UpperCamelCase_ ) )] ) / len(UpperCamelCase_ ) return {"accuracy": accuracy} # map train dataset __lowerCamelCase = train_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) train_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) # same for validation dataset __lowerCamelCase = val_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) val_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) __lowerCamelCase = self.get_auto_remove_tmp_dir() __lowerCamelCase = SeqaSeqTrainingArguments( output_dir=UpperCamelCase_ , per_device_train_batch_size=UpperCamelCase_ , per_device_eval_batch_size=UpperCamelCase_ , predict_with_generate=UpperCamelCase_ , evaluation_strategy="""steps""" , do_train=UpperCamelCase_ , do_eval=UpperCamelCase_ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer __lowerCamelCase = SeqaSeqTrainer( model=UpperCamelCase_ , args=UpperCamelCase_ , compute_metrics=_compute_metrics , train_dataset=UpperCamelCase_ , eval_dataset=UpperCamelCase_ , tokenizer=UpperCamelCase_ , ) # start training trainer.train()
80
0
def lowerCamelCase__ ( A__ : float , A__ : int ): if digit_amount > 0: return round(number - int(__A ) , __A ) return number - int(__A ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
708
class lowerCamelCase__: # Public class to implement a graph def __init__( self: Dict , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): __lowerCamelCase = row __lowerCamelCase = col __lowerCamelCase = graph def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): # Checking all 8 elements surrounding nth element __lowerCamelCase = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __lowerCamelCase = [-1, 0, 1, -1, 1, -1, 0, 1] __lowerCamelCase = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): # And finally, count all islands. __lowerCamelCase = [[False for j in range(self.COL )] for i in range(self.ROW )] __lowerCamelCase = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) count += 1 return count
80
0
import gc import unittest from diffusers import FlaxStableDiffusionInpaintPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Union[str, Any] ): # clean up the VRAM after each test super().tearDown() gc.collect() def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) __lowerCamelCase = """xvjiarui/stable-diffusion-2-inpainting""" __lowerCamelCase, __lowerCamelCase = FlaxStableDiffusionInpaintPipeline.from_pretrained(_lowerCAmelCase , safety_checker=_lowerCAmelCase ) __lowerCamelCase = """Face of a yellow cat, high resolution, sitting on a park bench""" __lowerCamelCase = jax.random.PRNGKey(0 ) __lowerCamelCase = 50 __lowerCamelCase = jax.device_count() __lowerCamelCase = num_samples * [prompt] __lowerCamelCase = num_samples * [init_image] __lowerCamelCase = num_samples * [mask_image] __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = pipeline.prepare_inputs(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # shard inputs and rng __lowerCamelCase = replicate(_lowerCAmelCase ) __lowerCamelCase = jax.random.split(_lowerCAmelCase , jax.device_count() ) __lowerCamelCase = shard(_lowerCAmelCase ) __lowerCamelCase = shard(_lowerCAmelCase ) __lowerCamelCase = shard(_lowerCAmelCase ) __lowerCamelCase = pipeline( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , jit=_lowerCAmelCase ) __lowerCamelCase = output.images.reshape(_lowerCAmelCase , 5_12 , 5_12 , 3 ) __lowerCamelCase = images[0, 2_53:2_56, 2_53:2_56, -1] __lowerCamelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) __lowerCamelCase = jnp.array( [0.361_1307, 0.3764_9736, 0.375_7408, 0.3821_3953, 0.3929_5167, 0.384_1631, 0.4155_4978, 0.413_7475, 0.421_7084] ) print(F'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
709
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = DPTConfig() if "large" in checkpoint_url: __lowerCamelCase = 1024 __lowerCamelCase = 4096 __lowerCamelCase = 24 __lowerCamelCase = 16 __lowerCamelCase = [5, 11, 17, 23] __lowerCamelCase = [256, 512, 1024, 1024] __lowerCamelCase = (1, 384, 384) if "ade" in checkpoint_url: __lowerCamelCase = True __lowerCamelCase = 150 __lowerCamelCase = """huggingface/label-files""" __lowerCamelCase = """ade20k-id2label.json""" __lowerCamelCase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type="""dataset""" ) ) , """r""" ) ) __lowerCamelCase = {int(A__ ): v for k, v in idalabel.items()} __lowerCamelCase = idalabel __lowerCamelCase = {v: k for k, v in idalabel.items()} __lowerCamelCase = [1, 150, 480, 480] return config, expected_shape def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' __lowerCamelCase = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.encoder""" ) if "pretrained.model" in name: __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.embeddings""" ) if "patch_embed" in name: __lowerCamelCase = name.replace("""patch_embed""" , """patch_embeddings""" ) if "pos_embed" in name: __lowerCamelCase = name.replace("""pos_embed""" , """position_embeddings""" ) if "attn.proj" in name: __lowerCamelCase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "proj" in name and "project" not in name: __lowerCamelCase = name.replace("""proj""" , """projection""" ) if "blocks" in name: __lowerCamelCase = name.replace("""blocks""" , """layer""" ) if "mlp.fc1" in name: __lowerCamelCase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: __lowerCamelCase = name.replace("""mlp.fc2""" , """output.dense""" ) if "norm1" in name: __lowerCamelCase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: __lowerCamelCase = name.replace("""norm2""" , """layernorm_after""" ) if "scratch.output_conv" in name: __lowerCamelCase = name.replace("""scratch.output_conv""" , """head""" ) if "scratch" in name: __lowerCamelCase = name.replace("""scratch""" , """neck""" ) if "layer1_rn" in name: __lowerCamelCase = name.replace("""layer1_rn""" , """convs.0""" ) if "layer2_rn" in name: __lowerCamelCase = name.replace("""layer2_rn""" , """convs.1""" ) if "layer3_rn" in name: __lowerCamelCase = name.replace("""layer3_rn""" , """convs.2""" ) if "layer4_rn" in name: __lowerCamelCase = name.replace("""layer4_rn""" , """convs.3""" ) if "refinenet" in name: __lowerCamelCase = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __lowerCamelCase = name.replace(f'refinenet{layer_idx}' , f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: __lowerCamelCase = name.replace("""out_conv""" , """projection""" ) if "resConfUnit1" in name: __lowerCamelCase = name.replace("""resConfUnit1""" , """residual_layer1""" ) if "resConfUnit2" in name: __lowerCamelCase = name.replace("""resConfUnit2""" , """residual_layer2""" ) if "conv1" in name: __lowerCamelCase = name.replace("""conv1""" , """convolution1""" ) if "conv2" in name: __lowerCamelCase = name.replace("""conv2""" , """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.0.project.0""" , """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.0.project.0""" , """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.0.project.0""" , """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.0.project.0""" , """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.3""" , """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.4""" , """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.3""" , """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.4""" , """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.3""" , """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.3""" , """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.4""" , """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: __lowerCamelCase = name.replace("""pretrained""" , """dpt""" ) if "bn" in name: __lowerCamelCase = name.replace("""bn""" , """batch_norm""" ) if "head" in name: __lowerCamelCase = name.replace("""head""" , """head.head""" ) if "encoder.norm" in name: __lowerCamelCase = name.replace("""encoder.norm""" , """layernorm""" ) if "auxlayer" in name: __lowerCamelCase = name.replace("""auxlayer""" , """auxiliary_head.head""" ) return name def lowerCamelCase__ ( A__ : Tuple , A__ : Any ): '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __lowerCamelCase = in_proj_weight[: config.hidden_size, :] __lowerCamelCase = in_proj_bias[: config.hidden_size] __lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] __lowerCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" __lowerCamelCase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( A__ : Optional[int] , A__ : Union[str, Any] , A__ : List[str] , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = get_dpt_config(A__ ) # load original state_dict from URL __lowerCamelCase = torch.hub.load_state_dict_from_url(A__ , map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(A__ ) # rename keys for key in state_dict.copy().keys(): __lowerCamelCase = state_dict.pop(A__ ) __lowerCamelCase = val # read in qkv matrices read_in_q_k_v(A__ , A__ ) # load HuggingFace model __lowerCamelCase = DPTForSemanticSegmentation(A__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(A__ ) model.load_state_dict(A__ ) model.eval() # Check outputs on an image __lowerCamelCase = 480 if """ade""" in checkpoint_url else 384 __lowerCamelCase = DPTImageProcessor(size=A__ ) __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(A__ , return_tensors="""pt""" ) # forward pass __lowerCamelCase = model(**A__ ).logits if """ade""" in checkpoint_url else model(**A__ ).predicted_depth # Assert logits __lowerCamelCase = torch.tensor([[6.3_199, 6.3_629, 6.4_148], [6.3_850, 6.3_615, 6.4_166], [6.3_519, 6.3_176, 6.3_575]] ) if "ade" in checkpoint_url: __lowerCamelCase = torch.tensor([[4.0_480, 4.2_420, 4.4_360], [4.3_124, 4.5_693, 4.8_261], [4.5_768, 4.8_965, 5.2_163]] ) assert outputs.shape == torch.Size(A__ ) assert ( torch.allclose(outputs[0, 0, :3, :3] , A__ , atol=1E-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3] , A__ ) ) Path(A__ ).mkdir(exist_ok=A__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(A__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=A__ , ) image_processor.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=A__ , ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) UpperCAmelCase_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
80
0
from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging UpperCAmelCase_ = logging.get_logger(__name__) def lowerCamelCase__ ( A__ : Any ): '''simple docstring''' if isinstance(lowerCamelCase__ , np.ndarray ): return list(tensor.shape ) __lowerCamelCase = tf.shape(lowerCamelCase__ ) if tensor.shape == tf.TensorShape(lowerCamelCase__ ): return dynamic __lowerCamelCase = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(lowerCamelCase__ )] def lowerCamelCase__ ( A__ : int , A__ : Optional[int] = None , A__ : Union[str, Any] = None ): '''simple docstring''' return tf.nn.softmax(logits=logits + 1E-9 , axis=lowerCamelCase__ , name=lowerCamelCase__ ) def lowerCamelCase__ ( A__ : str , A__ : Optional[Any] , A__ : Tuple , A__ : Optional[int]=1E-5 , A__ : Optional[int]=-1 ): '''simple docstring''' if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise NotImplementedError("""Only 1D weight and bias tensors are supported for now, with only a single axis.""" ) # Get mean and variance on the axis to be normalized __lowerCamelCase = tf.nn.moments(lowerCamelCase__ , axes=[axis] , keepdims=lowerCamelCase__ ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis __lowerCamelCase = [1] * inputs.shape.rank __lowerCamelCase = shape_list(lowerCamelCase__ )[axis] __lowerCamelCase = tf.reshape(lowerCamelCase__ , lowerCamelCase__ ) __lowerCamelCase = tf.reshape(lowerCamelCase__ , lowerCamelCase__ ) # Compute layer normalization using the batch_normalization # function. __lowerCamelCase = tf.nn.batch_normalization( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , offset=lowerCamelCase__ , scale=lowerCamelCase__ , variance_epsilon=lowerCamelCase__ , ) return outputs def lowerCamelCase__ ( A__ : str , A__ : int=0 , A__ : str=-1 ): '''simple docstring''' if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input __lowerCamelCase = tf.shape(lowerCamelCase__ ) __lowerCamelCase = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) __lowerCamelCase = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(lowerCamelCase__ , lowerCamelCase__ ) def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' if not isinstance(lowerCamelCase__ , tf.Tensor ): __lowerCamelCase = tf.convert_to_tensor(lowerCamelCase__ ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: __lowerCamelCase = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: __lowerCamelCase = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) __lowerCamelCase = ( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def lowerCamelCase__ ( A__ : int , A__ : Any , A__ : List[str] = "input_ids" ): '''simple docstring''' tf.debugging.assert_less( lowerCamelCase__ , tf.cast(lowerCamelCase__ , dtype=tensor.dtype ) , message=( f'The maximum value of {tensor_name} ({tf.math.reduce_max(lowerCamelCase__ )}) must be smaller than the embedding ' f'layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.' ) , ) def lowerCamelCase__ ( A__ : List[str] , A__ : str , A__ : Optional[int] ): '''simple docstring''' __lowerCamelCase = 64512 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. __lowerCamelCase = [x for x in data if len(lowerCamelCase__ ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( """The following attributes cannot be saved to HDF5 file because """ f'they are larger than {HDF5_OBJECT_HEADER_LIMIT} ' f'bytes: {bad_attributes}' ) __lowerCamelCase = np.asarray(lowerCamelCase__ ) __lowerCamelCase = 1 __lowerCamelCase = np.array_split(lowerCamelCase__ , lowerCamelCase__ ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 __lowerCamelCase = np.array_split(lowerCamelCase__ , lowerCamelCase__ ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(lowerCamelCase__ ): __lowerCamelCase = chunk_data else: __lowerCamelCase = data def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Union[str, Any] ): '''simple docstring''' if name in group.attrs: __lowerCamelCase = [n.decode("""utf8""" ) if hasattr(lowerCamelCase__ , """decode""" ) else n for n in group.attrs[name]] else: __lowerCamelCase = [] __lowerCamelCase = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode("""utf8""" ) if hasattr(lowerCamelCase__ , """decode""" ) else n for n in group.attrs["""%s%d""" % (name, chunk_id)]] ) chunk_id += 1 return data def lowerCamelCase__ ( A__ : str ): '''simple docstring''' def _expand_single_ad_tensor(A__ : str ): if isinstance(lowerCamelCase__ , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(lowerCamelCase__ , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , lowerCamelCase__ )
710
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
80
0
import collections import os import re from pathlib import Path UpperCAmelCase_ = 'src/transformers' # Matches is_xxx_available() UpperCAmelCase_ = re.compile(r'is\_([a-z_]*)_available()') # Catches a one-line _import_struct = {xxx} UpperCAmelCase_ = re.compile(r'^_import_structure\s+=\s+\{([^\}]+)\}') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] UpperCAmelCase_ = re.compile(r'\s+"\S*":\s+\[([^\]]*)\]') # Catches a line if not is_foo_available UpperCAmelCase_ = re.compile(r'^\s*if\s+not\s+is\_[a-z_]*\_available\(\)') # Catches a line _import_struct["bla"].append("foo") UpperCAmelCase_ = re.compile(r'^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] UpperCAmelCase_ = re.compile(r'^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]') # Catches a line with an object between quotes and a comma: "MyModel", UpperCAmelCase_ = re.compile(r'^\s+"([^"]+)",') # Catches a line with objects between brackets only: ["foo", "bar"], UpperCAmelCase_ = re.compile(r'^\s+\[([^\]]+)\]') # Catches a line with from foo import bar, bla, boo UpperCAmelCase_ = re.compile(r'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') # Catches a line with try: UpperCAmelCase_ = re.compile(r'^\s*try:') # Catches a line with else: UpperCAmelCase_ = re.compile(r'^\s*else:') def lowerCamelCase__ ( A__ : str ): '''simple docstring''' if _re_test_backend.search(_lowerCAmelCase ) is None: return None __lowerCamelCase = [b[0] for b in _re_backend.findall(_lowerCAmelCase )] backends.sort() return "_and_".join(_lowerCAmelCase ) def lowerCamelCase__ ( A__ : List[str] ): '''simple docstring''' with open(_lowerCAmelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: __lowerCamelCase = f.readlines() __lowerCamelCase = 0 while line_index < len(_lowerCAmelCase ) and not lines[line_index].startswith("""_import_structure = {""" ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(_lowerCAmelCase ): return None # First grab the objects without a specific backend in _import_structure __lowerCamelCase = [] while not lines[line_index].startswith("""if TYPE_CHECKING""" ) and find_backend(lines[line_index] ) is None: __lowerCamelCase = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(_lowerCAmelCase ): __lowerCamelCase = _re_one_line_import_struct.search(_lowerCAmelCase ).groups()[0] __lowerCamelCase = re.findall(R"""\[([^\]]+)\]""" , _lowerCAmelCase ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(""", """ )] ) line_index += 1 continue __lowerCamelCase = _re_import_struct_key_value.search(_lowerCAmelCase ) if single_line_import_search is not None: __lowerCamelCase = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(""", """ ) if len(_lowerCAmelCase ) > 0] objects.extend(_lowerCAmelCase ) elif line.startswith(""" """ * 8 + """\"""" ): objects.append(line[9:-3] ) line_index += 1 __lowerCamelCase = {"none": objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith("""if TYPE_CHECKING""" ): # If the line is an if not is_backend_available, we grab all objects associated. __lowerCamelCase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: __lowerCamelCase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 __lowerCamelCase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(""" """ * 4 ): __lowerCamelCase = lines[line_index] if _re_import_struct_add_one.search(_lowerCAmelCase ) is not None: objects.append(_re_import_struct_add_one.search(_lowerCAmelCase ).groups()[0] ) elif _re_import_struct_add_many.search(_lowerCAmelCase ) is not None: __lowerCamelCase = _re_import_struct_add_many.search(_lowerCAmelCase ).groups()[0].split(""", """ ) __lowerCamelCase = [obj[1:-1] for obj in imports if len(_lowerCAmelCase ) > 0] objects.extend(_lowerCAmelCase ) elif _re_between_brackets.search(_lowerCAmelCase ) is not None: __lowerCamelCase = _re_between_brackets.search(_lowerCAmelCase ).groups()[0].split(""", """ ) __lowerCamelCase = [obj[1:-1] for obj in imports if len(_lowerCAmelCase ) > 0] objects.extend(_lowerCAmelCase ) elif _re_quote_object.search(_lowerCAmelCase ) is not None: objects.append(_re_quote_object.search(_lowerCAmelCase ).groups()[0] ) elif line.startswith(""" """ * 8 + """\"""" ): objects.append(line[9:-3] ) elif line.startswith(""" """ * 12 + """\"""" ): objects.append(line[13:-3] ) line_index += 1 __lowerCamelCase = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend __lowerCamelCase = [] while ( line_index < len(_lowerCAmelCase ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith("""else""" ) ): __lowerCamelCase = lines[line_index] __lowerCamelCase = _re_import.search(_lowerCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(""", """ ) ) elif line.startswith(""" """ * 8 ): objects.append(line[8:-2] ) line_index += 1 __lowerCamelCase = {"none": objects} # Let's continue with backend-specific objects while line_index < len(_lowerCAmelCase ): # If the line is an if is_backend_available, we grab all objects associated. __lowerCamelCase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: __lowerCamelCase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 __lowerCamelCase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(""" """ * 8 ): __lowerCamelCase = lines[line_index] __lowerCamelCase = _re_import.search(_lowerCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(""", """ ) ) elif line.startswith(""" """ * 12 ): objects.append(line[12:-2] ) line_index += 1 __lowerCamelCase = objects else: line_index += 1 return import_dict_objects, type_hint_objects def lowerCamelCase__ ( A__ : List[Any] , A__ : List[Any] ): '''simple docstring''' def find_duplicates(A__ : List[Any] ): return [k for k, v in collections.Counter(_lowerCAmelCase ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] __lowerCamelCase = [] for key in import_dict_objects.keys(): __lowerCamelCase = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(f'Duplicate _import_structure definitions for: {duplicate_imports}' ) __lowerCamelCase = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(f'Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): __lowerCamelCase = "base imports" if key == "none" else f'{key} backend' errors.append(f'Differences for {name}:' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(f' {a} in TYPE_HINT but not in _import_structure.' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(f' {a} in _import_structure but not in TYPE_HINT.' ) return errors def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = [] for root, _, files in os.walk(_lowerCAmelCase ): if "__init__.py" in files: __lowerCamelCase = os.path.join(_lowerCAmelCase , """__init__.py""" ) __lowerCamelCase = parse_init(_lowerCAmelCase ) if objects is not None: __lowerCamelCase = analyze_results(*_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: __lowerCamelCase = f'Problem in {fname}, both halves do not define the same objects.\n{errors[0]}' failures.append("""\n""".join(_lowerCAmelCase ) ) if len(_lowerCAmelCase ) > 0: raise ValueError("""\n\n""".join(_lowerCAmelCase ) ) def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = [] for path, directories, files in os.walk(_lowerCAmelCase ): for folder in directories: # Ignore private modules if folder.startswith("""_""" ): directories.remove(_lowerCAmelCase ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(_lowerCAmelCase ) / folder).glob("""*.py""" ) ) ) == 0: continue __lowerCamelCase = str((Path(_lowerCAmelCase ) / folder).relative_to(_lowerCAmelCase ) ) __lowerCamelCase = short_path.replace(os.path.sep , """.""" ) submodules.append(_lowerCAmelCase ) for fname in files: if fname == "__init__.py": continue __lowerCamelCase = str((Path(_lowerCAmelCase ) / fname).relative_to(_lowerCAmelCase ) ) __lowerCamelCase = short_path.replace(""".py""" , """""" ).replace(os.path.sep , """.""" ) if len(submodule.split(""".""" ) ) == 1: submodules.append(_lowerCAmelCase ) return submodules UpperCAmelCase_ = [ 'convert_pytorch_checkpoint_to_tf2', 'modeling_flax_pytorch_utils', 'models.esm.openfold_utils', ] def lowerCamelCase__ ( ): '''simple docstring''' from transformers.utils import direct_transformers_import __lowerCamelCase = direct_transformers_import(_lowerCAmelCase ) __lowerCamelCase = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(_lowerCAmelCase , """__init__.py""" ) , """r""" ) as f: __lowerCamelCase = f.read() import_structure_keys.update(set(re.findall(R"""import_structure\[\"([^\"]*)\"\]""" , _lowerCAmelCase ) ) ) __lowerCamelCase = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(_lowerCAmelCase ) > 0: __lowerCamelCase = "\n".join(f'- {module}' for module in module_not_registered ) raise ValueError( """The following submodules are not properly registed in the main init of Transformers:\n""" f'{list_of_modules}\n' """Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.""" ) if __name__ == "__main__": check_all_inits() check_submodules()
711
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/config.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/config.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/config.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/config.json', 'bert-base-multilingual-uncased': 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json', 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/config.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/config.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-base-cased-finetuned-mrpc': 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json', 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json', 'bert-base-german-dbmdz-uncased': 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json', 'cl-tohoku/bert-base-japanese': 'https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json', 'cl-tohoku/bert-base-japanese-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json' ), 'wietsedv/bert-base-dutch-cased': 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json', # See all BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = 'bert' def __init__( self: List[str] , UpperCamelCase_: str=3_05_22 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Tuple=12 , UpperCamelCase_: int=12 , UpperCamelCase_: int=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[Any]=0.1 , UpperCamelCase_: Optional[int]=5_12 , UpperCamelCase_: List[Any]=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: List[str]=1E-12 , UpperCamelCase_: Dict=0 , UpperCamelCase_: List[Any]="absolute" , UpperCamelCase_: Tuple=True , UpperCamelCase_: Tuple=None , **UpperCamelCase_: Optional[Any] , ): super().__init__(pad_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = position_embedding_type __lowerCamelCase = use_cache __lowerCamelCase = classifier_dropout class lowerCamelCase__( __lowerCamelCase): @property def lowerCAmelCase__ ( self: Any ): if self.task == "multiple-choice": __lowerCamelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowerCamelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
80
0
from datetime import datetime as dt import os from github import Github UpperCAmelCase_ = [ "good first issue", "good second issue", "good difficult issue", "feature request", "new model", "wip", ] def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = Github(os.environ["""GITHUB_TOKEN"""] ) __lowerCamelCase = g.get_repo("""huggingface/transformers""" ) __lowerCamelCase = repo.get_issues(state="""open""" ) for issue in open_issues: __lowerCamelCase = sorted([comment for comment in issue.get_comments()] , key=lambda A__ : i.created_at , reverse=__lowerCAmelCase ) __lowerCamelCase = comments[0] if len(__lowerCAmelCase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="""closed""" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""" ) if __name__ == "__main__": main()
712
from __future__ import annotations from math import ceil, floor, sqrt def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0] __lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target __lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target __lowerCamelCase = 0 # an estimate of b, using the quadratic formula __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the triangle number corresponding to b_floor __lowerCamelCase = 42 # the triangle number corresponding to b_ceil __lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): __lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 __lowerCamelCase = floor(A__ ) __lowerCamelCase = ceil(A__ ) __lowerCamelCase = triangle_numbers[b_floor] __lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_first_guess * triangle_a __lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_second_guess * triangle_a __lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(f"""{solution() = }""")
80
0
'''simple docstring''' from functools import reduce UpperCAmelCase_ = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403098711121722383113' '62229893423380308135336276614282806444486645238749' '30358907296290491560440772390713810515859307960866' '70172427121883998797908792274921901699720888093776' '65727333001053367881220235421809751254540594752243' '52584907711670556013604839586446706324415722155397' '53697817977846174064955149290862569321978468622482' '83972241375657056057490261407972968652414535100474' '82166370484403199890008895243450658541227588666881' '16427171479924442928230863465674813919123162824586' '17866458359124566529476545682848912883142607690042' '24219022671055626321111109370544217506941658960408' '07198403850962455444362981230987879927244284909188' '84580156166097919133875499200524063689912560717606' '05886116467109405077541002256983155200055935729725' '71636269561882670428252483600823257530420752963450' ) def lowerCamelCase__ ( A__ : Optional[Any] = N ): '''simple docstring''' return max( # mypy cannot properly interpret reduce int(reduce(lambda A__ , A__ : str(int(__UpperCAmelCase ) * int(__UpperCAmelCase ) ) , n[i : i + 13] ) ) for i in range(len(__UpperCAmelCase ) - 12 ) ) if __name__ == "__main__": print(f"""{solution() = }""")
713
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[str] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet, attn in zip(self.resnets , self.attentions ): __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: str , UpperCamelCase_: Any , UpperCamelCase_: Optional[int] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet in self.resnets: __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: Tuple , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: List[Any]=True ): for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Optional[Any]=True ): for resnet in self.resnets: # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: int ): # there is always at least one resnet __lowerCamelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] __lowerCamelCase = [] for _ in range(self.num_layers ): __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions def __call__( self: int , UpperCamelCase_: Any , UpperCamelCase_: int , UpperCamelCase_: Dict , UpperCamelCase_: Optional[int]=True ): __lowerCamelCase = self.resnets[0](UpperCamelCase_ , UpperCamelCase_ ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) return hidden_states
80
0
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ = { 'configuration_informer': [ 'INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'InformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ 'INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'InformerForPrediction', 'InformerModel', 'InformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
714
import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging UpperCAmelCase_ = ['bart.large', 'bart.large.mnli', 'bart.large.cnn', 'bart_xsum/model.pt'] UpperCAmelCase_ = {'bart.large': BartModel, 'bart.large.mnli': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('0.9.0'): raise Exception('requires fairseq >= 0.9.0') logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = ' Hello world! cécé herlolip' UpperCAmelCase_ = [ ('model.classification_heads.mnli.dense.weight', 'classification_head.dense.weight'), ('model.classification_heads.mnli.dense.bias', 'classification_head.dense.bias'), ('model.classification_heads.mnli.out_proj.weight', 'classification_head.out_proj.weight'), ('model.classification_heads.mnli.out_proj.bias', 'classification_head.out_proj.bias'), ] def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """_float_tensor""", ] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Tuple , A__ : Any , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase = dct.pop(A__ ) __lowerCamelCase = val def lowerCamelCase__ ( A__ : Tuple ): '''simple docstring''' __lowerCamelCase = torch.load(A__ , map_location="""cpu""" ) __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , """bart.large.cnn""" ).eval() hub_interface.model.load_state_dict(sd["""model"""] ) return hub_interface def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = emb.weight.shape __lowerCamelCase = nn.Linear(A__ , A__ , bias=A__ ) __lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Optional[int] , A__ : Dict=None ): '''simple docstring''' if not os.path.exists(A__ ): __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , A__ ).eval() else: __lowerCamelCase = load_xsum_checkpoint(A__ ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: __lowerCamelCase = checkpoint_path.replace(""".""" , """-""" ) __lowerCamelCase = BartConfig.from_pretrained(A__ ) __lowerCamelCase = bart.encode(A__ ).unsqueeze(0 ) __lowerCamelCase = BartTokenizer.from_pretrained(A__ ).encode(A__ , return_tensors="""pt""" ).unsqueeze(0 ) if not torch.eq(A__ , A__ ).all(): raise ValueError( f'converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}' ) if checkpoint_path == "bart.large.mnli": __lowerCamelCase = bart.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""model.decoder.embed_tokens.weight"""] for src, dest in mnli_rename_keys: rename_key(A__ , A__ , A__ ) __lowerCamelCase = BartForSequenceClassification(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = bart.predict("""mnli""" , A__ , return_logits=A__ ) __lowerCamelCase = model(A__ )[0] # logits else: # no classification heads to worry about __lowerCamelCase = bart.model.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""decoder.embed_tokens.weight"""] __lowerCamelCase = bart.extract_features(A__ ) if hf_checkpoint_name == "facebook/bart-large": __lowerCamelCase = BartModel(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = model(A__ ).model[0] else: __lowerCamelCase = BartForConditionalGeneration(A__ ).eval() # an existing summarization ckpt model.model.load_state_dict(A__ ) if hasattr(A__ , """lm_head""" ): __lowerCamelCase = make_linear_from_emb(model.model.shared ) __lowerCamelCase = model.model(A__ )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f'`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}' ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError("""Some values in `fairseq_output` are different from `new_model_outputs`""" ) Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default=None, type=str, help='Which huggingface architecture to use: bart-large-xsum' ) UpperCAmelCase_ = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
80
0
from __future__ import annotations def lowerCamelCase__ ( A__ : Dict , A__ : List[str] , A__ : Dict ): '''simple docstring''' __lowerCamelCase = list(range(len(__A ) ) ) __lowerCamelCase = [v / w for v, w in zip(__A , __A )] index.sort(key=lambda A__ : ratio[i] , reverse=__A ) __lowerCamelCase = 0 __lowerCamelCase = [0] * len(__A ) for i in index: if weight[i] <= capacity: __lowerCamelCase = 1 max_value += value[i] capacity -= weight[i] else: __lowerCamelCase = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
715
import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class lowerCamelCase__: def __init__( self: Tuple , UpperCamelCase_: Any , UpperCamelCase_: List[Any]=14 , UpperCamelCase_: int=7 , UpperCamelCase_: Union[str, Any]=True , UpperCamelCase_: Dict=True , UpperCamelCase_: Optional[Any]=True , UpperCamelCase_: Tuple=True , UpperCamelCase_: List[str]=True , UpperCamelCase_: int=99 , UpperCamelCase_: str=32 , UpperCamelCase_: List[Any]=5 , UpperCamelCase_: Optional[int]=4 , UpperCamelCase_: List[Any]=37 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[str]=5_12 , UpperCamelCase_: Dict=16 , UpperCamelCase_: List[str]=2 , UpperCamelCase_: Optional[Any]=0.02 , UpperCamelCase_: List[str]=3 , UpperCamelCase_: Tuple=4 , UpperCamelCase_: Tuple=None , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_input_mask __lowerCamelCase = use_labels __lowerCamelCase = use_mc_token_ids __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = self.vocab_size - 1 def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None if self.use_mc_token_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = self.get_config() __lowerCamelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def lowerCAmelCase__ ( self: Dict ): return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: str , UpperCamelCase_: Dict , UpperCamelCase_: Tuple , UpperCamelCase_: Any , UpperCamelCase_: List[str] , *UpperCamelCase_: Optional[Any] ): __lowerCamelCase = CTRLModel(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , head_mask=UpperCamelCase_ ) model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ ) __lowerCamelCase = model(UpperCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Dict , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: List[Any] , *UpperCamelCase_: Tuple ): __lowerCamelCase = CTRLLMHeadModel(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ) = config_and_inputs __lowerCamelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask} return config, inputs_dict def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , *UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = self.num_labels __lowerCamelCase = CTRLForSequenceClassification(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : Any = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () UpperCAmelCase__ : Optional[Any] = (CTRLLMHeadModel,) if is_torch_available() else () UpperCAmelCase__ : int = ( { 'feature-extraction': CTRLModel, 'text-classification': CTRLForSequenceClassification, 'text-generation': CTRLLMHeadModel, 'zero-shot': CTRLForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ : List[str] = True UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : Optional[Any] = False def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Any , UpperCamelCase_: List[str] , UpperCamelCase_: Tuple , UpperCamelCase_: Tuple , UpperCamelCase_: List[str] ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = CTRLModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=UpperCamelCase_ , n_embd=37 ) def lowerCAmelCase__ ( self: Optional[int] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__ ( self: Optional[Any] ): self.config_tester.run_common_tests() def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*UpperCamelCase_ ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*UpperCamelCase_ ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCAmelCase__ ( self: List[Any] ): pass @slow def lowerCAmelCase__ ( self: Optional[Any] ): for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = CTRLModel.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) @unittest.skip("""The model doesn't support left padding""" ) # and it's not used enough to be worth fixing :) def lowerCAmelCase__ ( self: Optional[Any] ): pass @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: List[str] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = CTRLLMHeadModel.from_pretrained("""ctrl""" ) model.to(UpperCamelCase_ ) __lowerCamelCase = torch.tensor( [[1_18_59, 0, 16_11, 8]] , dtype=torch.long , device=UpperCamelCase_ ) # Legal the president is __lowerCamelCase = [ 1_18_59, 0, 16_11, 8, 5, 1_50, 2_64_49, 2, 19, 3_48, 4_69, 3, 25_95, 48, 2_07_40, 24_65_33, 24_65_33, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a __lowerCamelCase = model.generate(UpperCamelCase_ , do_sample=UpperCamelCase_ ) self.assertListEqual(output_ids[0].tolist() , UpperCamelCase_ )
80
0
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCamelCase__: def __init__( self: int , UpperCamelCase_: List[Any] , UpperCamelCase_: List[str]=13 , UpperCamelCase_: Any=32 , UpperCamelCase_: int=3 , UpperCamelCase_: List[str]=4 , UpperCamelCase_: Dict=[10, 20, 30, 40] , UpperCamelCase_: List[str]=[2, 2, 3, 2] , UpperCamelCase_: Tuple=True , UpperCamelCase_: Optional[Any]=True , UpperCamelCase_: Optional[Any]=37 , UpperCamelCase_: Tuple="gelu" , UpperCamelCase_: Optional[Any]=10 , UpperCamelCase_: int=0.02 , UpperCamelCase_: Optional[int]=["stage2", "stage3", "stage4"] , UpperCamelCase_: Any=[2, 3, 4] , UpperCamelCase_: Optional[int]=None , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = image_size __lowerCamelCase = num_channels __lowerCamelCase = num_stages __lowerCamelCase = hidden_sizes __lowerCamelCase = depths __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = num_labels __lowerCamelCase = initializer_range __lowerCamelCase = out_features __lowerCamelCase = out_indices __lowerCamelCase = scope def lowerCAmelCase__ ( self: str ): __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.num_labels ) __lowerCamelCase = self.get_config() return config, pixel_values, labels def lowerCAmelCase__ ( self: List[str] ): return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_A , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: int , UpperCamelCase_: List[str] , UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = ConvNextModel(config=_A ) model.to(_A ) model.eval() __lowerCamelCase = model(_A ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: str ): __lowerCamelCase = ConvNextForImageClassification(_A ) model.to(_A ) model.eval() __lowerCamelCase = model(_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Optional[int] , UpperCamelCase_: Dict , UpperCamelCase_: Tuple ): __lowerCamelCase = ConvNextBackbone(config=_A ) model.to(_A ) model.eval() __lowerCamelCase = model(_A ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __lowerCamelCase = None __lowerCamelCase = ConvNextBackbone(config=_A ) model.to(_A ) model.eval() __lowerCamelCase = model(_A ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase = config_and_inputs __lowerCamelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase__( a__ , a__ , unittest.TestCase): UpperCAmelCase__ : Any = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) UpperCAmelCase__ : List[str] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) UpperCAmelCase__ : Optional[int] = True UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : List[str] = False def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = ConvNextModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=_A , has_text_modality=_A , hidden_size=37 ) def lowerCAmelCase__ ( self: str ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCAmelCase__ ( self: List[Any] ): return @unittest.skip(reason="""ConvNext does not use inputs_embeds""" ) def lowerCAmelCase__ ( self: Optional[int] ): pass @unittest.skip(reason="""ConvNext does not support input and output embeddings""" ) def lowerCAmelCase__ ( self: List[str] ): pass @unittest.skip(reason="""ConvNext does not use feedforward chunking""" ) def lowerCAmelCase__ ( self: Union[str, Any] ): pass def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(_A ) __lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , _A ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_A ) def lowerCAmelCase__ ( self: List[str] ): def check_hidden_states_output(UpperCamelCase_: Dict , UpperCamelCase_: List[Any] , UpperCamelCase_: List[Any] ): __lowerCamelCase = model_class(_A ) model.to(_A ) model.eval() with torch.no_grad(): __lowerCamelCase = model(**self._prepare_for_class(_A , _A ) ) __lowerCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowerCamelCase = self.model_tester.num_stages self.assertEqual(len(_A ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = True check_hidden_states_output(_A , _A , _A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCamelCase = True check_hidden_states_output(_A , _A , _A ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_A ) @slow def lowerCAmelCase__ ( self: List[str] ): for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = ConvNextModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class lowerCamelCase__( unittest.TestCase): @cached_property def lowerCAmelCase__ ( self: Union[str, Any] ): return AutoImageProcessor.from_pretrained("""facebook/convnext-tiny-224""" ) if is_vision_available() else None @slow def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = ConvNextForImageClassification.from_pretrained("""facebook/convnext-tiny-224""" ).to(_A ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=_A , return_tensors="""pt""" ).to(_A ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**_A ) # verify the logits __lowerCamelCase = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , _A ) __lowerCamelCase = torch.tensor([-0.0260, -0.4739, 0.1911] ).to(_A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _A , atol=1E-4 ) ) @require_torch class lowerCamelCase__( unittest.TestCase , a__): UpperCAmelCase__ : Any = (ConvNextBackbone,) if is_torch_available() else () UpperCAmelCase__ : Any = ConvNextConfig UpperCAmelCase__ : List[str] = False def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = ConvNextModelTester(self )
716
def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0 for i in range(n + 1 )] __lowerCamelCase = 1 __lowerCamelCase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , A__ ): __lowerCamelCase = 1 __lowerCamelCase = 0 for i in range(A__ ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(f"""{solution() = }""")
80
0
import unittest from knapsack import greedy_knapsack as kp class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = [10, 20, 30, 40, 50, 60] __lowerCamelCase = [2, 4, 6, 8, 10, 12] __lowerCamelCase = 1_00 self.assertEqual(kp.calc_profit(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , 2_10 ) def lowerCAmelCase__ ( self: int ): self.assertRaisesRegex(__lowerCamelCase , """max_weight must greater than zero.""" ) def lowerCAmelCase__ ( self: Tuple ): self.assertRaisesRegex(__lowerCamelCase , """Weight can not be negative.""" ) def lowerCAmelCase__ ( self: int ): self.assertRaisesRegex(__lowerCamelCase , """Profit can not be negative.""" ) def lowerCAmelCase__ ( self: List[Any] ): self.assertRaisesRegex(__lowerCamelCase , """max_weight must greater than zero.""" ) def lowerCAmelCase__ ( self: List[Any] ): self.assertRaisesRegex( __lowerCamelCase , """The length of profit and weight must be same.""" ) if __name__ == "__main__": unittest.main()
717
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase): UpperCAmelCase__ : Dict = 1 @register_to_config def __init__( self: List[str] , UpperCamelCase_: int = 10_00 , UpperCamelCase_: Optional[Union[np.ndarray, List[float]]] = None ): # set `betas`, `alphas`, `timesteps` self.set_timesteps(UpperCamelCase_ ) # standard deviation of the initial noise distribution __lowerCamelCase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowerCamelCase = 4 # running values __lowerCamelCase = [] def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: int , UpperCamelCase_: Union[str, torch.device] = None ): __lowerCamelCase = num_inference_steps __lowerCamelCase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowerCamelCase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowerCamelCase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowerCamelCase = torch.sin(steps * math.pi / 2 ) ** 2 __lowerCamelCase = (1.0 - self.betas**2) ** 0.5 __lowerCamelCase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowerCamelCase = timesteps.to(UpperCamelCase_ ) __lowerCamelCase = [] def lowerCAmelCase__ ( self: int , UpperCamelCase_: torch.FloatTensor , UpperCamelCase_: int , UpperCamelCase_: torch.FloatTensor , UpperCamelCase_: bool = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowerCamelCase = (self.timesteps == timestep).nonzero().item() __lowerCamelCase = timestep_index + 1 __lowerCamelCase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(UpperCamelCase_ ) if len(self.ets ) == 1: __lowerCamelCase = self.ets[-1] elif len(self.ets ) == 2: __lowerCamelCase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowerCamelCase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowerCamelCase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowerCamelCase = self._get_prev_sample(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: torch.FloatTensor , *UpperCamelCase_: Dict , **UpperCamelCase_: Union[str, Any] ): return sample def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Any ): __lowerCamelCase = self.alphas[timestep_index] __lowerCamelCase = self.betas[timestep_index] __lowerCamelCase = self.alphas[prev_timestep_index] __lowerCamelCase = self.betas[prev_timestep_index] __lowerCamelCase = (sample - sigma * ets) / max(UpperCamelCase_ , 1E-8 ) __lowerCamelCase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self: List[Any] ): return self.config.num_train_timesteps
80
0
from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class lowerCamelCase__( _UpperCAmelCase): UpperCAmelCase__ : Union[str, Any] = 'openai/whisper-base' UpperCAmelCase__ : Tuple = ( 'This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the ' 'transcribed text.' ) UpperCAmelCase__ : Dict = 'transcriber' UpperCAmelCase__ : Tuple = WhisperProcessor UpperCAmelCase__ : Optional[int] = WhisperForConditionalGeneration UpperCAmelCase__ : str = ['audio'] UpperCAmelCase__ : Dict = ['text'] def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: List[Any] ): return self.pre_processor(lowerCamelCase_ , return_tensors="""pt""" ).input_features def lowerCAmelCase__ ( self: int , UpperCamelCase_: int ): return self.model.generate(inputs=lowerCamelCase_ ) def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: List[Any] ): return self.pre_processor.batch_decode(lowerCamelCase_ , skip_special_tokens=lowerCamelCase_ )[0]
718
import os from collections.abc import Iterator def lowerCamelCase__ ( A__ : str = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(A__ ): __lowerCamelCase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(A__ )[1] in (".py", ".ipynb"): yield os.path.join(A__ , A__ ).lstrip("""./""" ) def lowerCamelCase__ ( A__ : Optional[int] ): '''simple docstring''' return f'{i * " "}*' if i else "\n##" def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(A__ ) or old_parts[i] != new_part) and new_part: print(f'{md_prefix(A__ )} {new_part.replace("_" , " " ).title()}' ) return new_path def lowerCamelCase__ ( A__ : str = "." ): '''simple docstring''' __lowerCamelCase = """""" for filepath in sorted(good_file_paths(A__ ) ): __lowerCamelCase, __lowerCamelCase = os.path.split(A__ ) if filepath != old_path: __lowerCamelCase = print_path(A__ , A__ ) __lowerCamelCase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowerCamelCase = f'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowerCamelCase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(f'{md_prefix(A__ )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md('.')
80
0
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCamelCase__: def __init__( self: List[Any] , UpperCamelCase_: int , UpperCamelCase_: int=13 , UpperCamelCase_: str=32 , UpperCamelCase_: List[Any]=3 , UpperCamelCase_: Union[str, Any]=4 , UpperCamelCase_: Tuple=[10, 20, 30, 40] , UpperCamelCase_: Optional[int]=[2, 2, 3, 2] , UpperCamelCase_: Optional[Any]=True , UpperCamelCase_: Any=True , UpperCamelCase_: Dict=37 , UpperCamelCase_: Optional[Any]="gelu" , UpperCamelCase_: str=10 , UpperCamelCase_: List[Any]=0.02 , UpperCamelCase_: Optional[int]=["stage2", "stage3", "stage4"] , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: str=None , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = image_size __lowerCamelCase = num_channels __lowerCamelCase = num_stages __lowerCamelCase = hidden_sizes __lowerCamelCase = depths __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = out_features __lowerCamelCase = num_labels __lowerCamelCase = scope __lowerCamelCase = num_stages def lowerCAmelCase__ ( self: int ): __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = self.get_config() return config, pixel_values, labels def lowerCAmelCase__ ( self: List[str] ): return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def lowerCAmelCase__ ( self: Optional[int] ): return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=5_12 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=__UpperCamelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=2_56 , auxiliary_num_convs=1 , auxiliary_concat_input=__UpperCamelCase , loss_ignore_index=2_55 , num_labels=self.num_labels , ) def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: List[Any] ): __lowerCamelCase = UperNetForSemanticSegmentation(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() __lowerCamelCase = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ) = config_and_inputs __lowerCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase__( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase): UpperCAmelCase__ : int = (UperNetForSemanticSegmentation,) if is_torch_available() else () UpperCAmelCase__ : List[Any] = {'image-segmentation': UperNetForSemanticSegmentation} if is_torch_available() else {} UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : Dict = False UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Optional[Any] = False def lowerCAmelCase__ ( self: str ): __lowerCamelCase = UperNetModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def lowerCAmelCase__ ( self: List[str] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCAmelCase__ ( self: Dict ): return def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(__UpperCamelCase ) __lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__UpperCamelCase ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def lowerCAmelCase__ ( self: Optional[int] ): pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def lowerCAmelCase__ ( self: Optional[int] ): pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowerCAmelCase__ ( self: Dict ): pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowerCAmelCase__ ( self: int ): pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowerCAmelCase__ ( self: Dict ): pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCAmelCase__ ( self: Tuple ): pass def lowerCAmelCase__ ( self: Dict ): def check_hidden_states_output(UpperCamelCase_: int , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Dict ): __lowerCamelCase = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): __lowerCamelCase = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) __lowerCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowerCamelCase = self.model_tester.num_stages self.assertEqual(len(__UpperCamelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCamelCase = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase, __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = _config_zero_init(__UpperCamelCase ) __lowerCamelCase = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: __lowerCamelCase = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def lowerCAmelCase__ ( self: int ): pass @slow def lowerCAmelCase__ ( self: Optional[Any] ): for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = UperNetForSemanticSegmentation.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""" ) __lowerCamelCase = Image.open(_A ).convert("""RGB""" ) return image @require_torch @require_vision @slow class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) __lowerCamelCase = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(__UpperCamelCase ) __lowerCamelCase = prepare_img() __lowerCamelCase = processor(images=__UpperCamelCase , return_tensors="""pt""" ).to(__UpperCamelCase ) with torch.no_grad(): __lowerCamelCase = model(**__UpperCamelCase ) __lowerCamelCase = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) __lowerCamelCase = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) __lowerCamelCase = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(__UpperCamelCase ) __lowerCamelCase = prepare_img() __lowerCamelCase = processor(images=__UpperCamelCase , return_tensors="""pt""" ).to(__UpperCamelCase ) with torch.no_grad(): __lowerCamelCase = model(**__UpperCamelCase ) __lowerCamelCase = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) __lowerCamelCase = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , __UpperCamelCase , atol=1E-4 ) )
719
from __future__ import annotations def lowerCamelCase__ ( A__ : list ): '''simple docstring''' if not nums: raise ValueError("""List is empty""" ) return sum(A__ ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
80
0
from copy import deepcopy class lowerCamelCase__: def __init__( self: Tuple , UpperCamelCase_: Union[str, Any] = None , UpperCamelCase_: str = None ): if arr is None and size is not None: __lowerCamelCase = size __lowerCamelCase = [0] * size elif arr is not None: self.init(snake_case_ ) else: raise ValueError("""Either arr or size must be specified""" ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = len(snake_case_ ) __lowerCamelCase = deepcopy(snake_case_ ) for i in range(1 , self.size ): __lowerCamelCase = self.next_(snake_case_ ) if j < self.size: self.tree[j] += self.tree[i] def lowerCAmelCase__ ( self: str ): __lowerCamelCase = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): __lowerCamelCase = self.next_(snake_case_ ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def lowerCAmelCase__ ( UpperCamelCase_: str ): return index + (index & (-index)) @staticmethod def lowerCAmelCase__ ( UpperCamelCase_: Optional[int] ): return index - (index & (-index)) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Optional[int] , UpperCamelCase_: str ): if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value __lowerCamelCase = self.next_(snake_case_ ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: Optional[Any] , UpperCamelCase_: int ): self.add(snake_case_ , value - self.get(snake_case_ ) ) def lowerCAmelCase__ ( self: Any , UpperCamelCase_: int ): if right == 0: return 0 __lowerCamelCase = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] __lowerCamelCase = self.prev(snake_case_ ) return result def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: Tuple , UpperCamelCase_: Dict ): return self.prefix(snake_case_ ) - self.prefix(snake_case_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Any ): return self.query(snake_case_ , index + 1 ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: Any ): value -= self.tree[0] if value < 0: return -1 __lowerCamelCase = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 __lowerCamelCase = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
720
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase): UpperCAmelCase__ : Any = 'maskformer-swin' UpperCAmelCase__ : List[Any] = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Any , UpperCamelCase_: Any=2_24 , UpperCamelCase_: List[str]=4 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: Optional[int]=96 , UpperCamelCase_: List[str]=[2, 2, 6, 2] , UpperCamelCase_: Optional[Any]=[3, 6, 12, 24] , UpperCamelCase_: str=7 , UpperCamelCase_: int=4.0 , UpperCamelCase_: Optional[int]=True , UpperCamelCase_: Union[str, Any]=0.0 , UpperCamelCase_: Optional[int]=0.0 , UpperCamelCase_: Any=0.1 , UpperCamelCase_: Union[str, Any]="gelu" , UpperCamelCase_: int=False , UpperCamelCase_: Optional[int]=0.02 , UpperCamelCase_: Optional[Any]=1E-5 , UpperCamelCase_: Optional[int]=None , UpperCamelCase_: List[Any]=None , **UpperCamelCase_: Union[str, Any] , ): super().__init__(**UpperCamelCase_ ) __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = embed_dim __lowerCamelCase = depths __lowerCamelCase = len(UpperCamelCase_ ) __lowerCamelCase = num_heads __lowerCamelCase = window_size __lowerCamelCase = mlp_ratio __lowerCamelCase = qkv_bias __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_absolute_embeddings __lowerCamelCase = layer_norm_eps __lowerCamelCase = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCamelCase = int(embed_dim * 2 ** (len(UpperCamelCase_ ) - 1) ) __lowerCamelCase = ["""stem"""] + [F'stage{idx}' for idx in range(1 , len(UpperCamelCase_ ) + 1 )] __lowerCamelCase, __lowerCamelCase = get_aligned_output_features_output_indices( out_features=UpperCamelCase_ , out_indices=UpperCamelCase_ , stage_names=self.stage_names )
80
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, is_vision_available, ) UpperCAmelCase_ = {'''configuration_vit''': ['''VIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTConfig''', '''ViTOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['''ViTFeatureExtractor'''] UpperCAmelCase_ = ['''ViTImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ '''VIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTForImageClassification''', '''ViTForMaskedImageModeling''', '''ViTModel''', '''ViTPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ '''TFViTForImageClassification''', '''TFViTModel''', '''TFViTPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ '''FlaxViTForImageClassification''', '''FlaxViTModel''', '''FlaxViTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit import VIT_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTConfig, ViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_vit import ViTFeatureExtractor from .image_processing_vit import ViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit import ( VIT_PRETRAINED_MODEL_ARCHIVE_LIST, ViTForImageClassification, ViTForMaskedImageModeling, ViTModel, ViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit import TFViTForImageClassification, TFViTModel, TFViTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel, FlaxViTPreTrainedModel else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
721
from __future__ import annotations def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): __lowerCamelCase, __lowerCamelCase = array[indexa], array[indexa] def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if length > 1: __lowerCamelCase = int(length / 2 ) for i in range(A__ , low + middle ): comp_and_swap(A__ , A__ , i + middle , A__ ) bitonic_merge(A__ , A__ , A__ , A__ ) bitonic_merge(A__ , low + middle , A__ , A__ ) def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if length > 1: __lowerCamelCase = int(length / 2 ) bitonic_sort(A__ , A__ , A__ , 1 ) bitonic_sort(A__ , low + middle , A__ , 0 ) bitonic_merge(A__ , A__ , A__ , A__ ) if __name__ == "__main__": UpperCAmelCase_ = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase_ = [int(item.strip()) for item in user_input.split(',')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('\nSorted array in ascending order is: ', end='') print(*unsorted, sep=', ') bitonic_merge(unsorted, 0, len(unsorted), 0) print('Sorted array in descending order is: ', end='') print(*unsorted, sep=', ')
80
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig UpperCAmelCase_ = { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/config.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/config.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/config.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/config.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/config.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/config.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/config.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/config.json', } class lowerCamelCase__( _lowercase): UpperCAmelCase__ : Tuple = '''albert''' def __init__( self: Any , UpperCamelCase_: List[str]=3_00_00 , UpperCamelCase_: Tuple=1_28 , UpperCamelCase_: Tuple=40_96 , UpperCamelCase_: Optional[int]=12 , UpperCamelCase_: Dict=1 , UpperCamelCase_: Optional[int]=64 , UpperCamelCase_: str=1_63_84 , UpperCamelCase_: Union[str, Any]=1 , UpperCamelCase_: List[Any]="gelu_new" , UpperCamelCase_: Optional[Any]=0 , UpperCamelCase_: Any=0 , UpperCamelCase_: Dict=5_12 , UpperCamelCase_: Dict=2 , UpperCamelCase_: Optional[Any]=0.02 , UpperCamelCase_: Optional[int]=1E-12 , UpperCamelCase_: Dict=0.1 , UpperCamelCase_: List[Any]="absolute" , UpperCamelCase_: Optional[int]=0 , UpperCamelCase_: Optional[Any]=2 , UpperCamelCase_: List[str]=3 , **UpperCamelCase_: List[Any] , ): super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ ) __lowerCamelCase = vocab_size __lowerCamelCase = embedding_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_hidden_groups __lowerCamelCase = num_attention_heads __lowerCamelCase = inner_group_num __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = classifier_dropout_prob __lowerCamelCase = position_embedding_type class lowerCamelCase__( _lowercase): @property def lowerCAmelCase__ ( self: List[str] ): if self.task == "multiple-choice": __lowerCamelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowerCamelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
700
from ... import PretrainedConfig UpperCAmelCase_ = { 'sijunhe/nezha-cn-base': 'https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json', } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Dict = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP UpperCAmelCase__ : Dict = 'nezha' def __init__( self: Dict , UpperCamelCase_: Any=2_11_28 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Optional[int]=12 , UpperCamelCase_: List[str]=12 , UpperCamelCase_: Optional[int]=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: str=0.1 , UpperCamelCase_: Union[str, Any]=5_12 , UpperCamelCase_: Any=64 , UpperCamelCase_: Dict=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: Optional[Any]=1E-12 , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Any=0 , UpperCamelCase_: str=2 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: str=True , **UpperCamelCase_: Any , ): super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = max_relative_position __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = classifier_dropout __lowerCamelCase = use_cache
80
0
'''simple docstring''' import pickle import numpy as np from matplotlib import pyplot as plt class lowerCamelCase__: def __init__( self: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: Optional[int] , UpperCamelCase_: Optional[Any]=0.2 , UpperCamelCase_: Tuple=0.2 ): __lowerCamelCase = bp_numa __lowerCamelCase = bp_numa __lowerCamelCase = bp_numa __lowerCamelCase = conva_get[:2] __lowerCamelCase = conva_get[2] __lowerCamelCase = size_pa __lowerCamelCase = rate_w __lowerCamelCase = rate_t __lowerCamelCase = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] __lowerCamelCase = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) __lowerCamelCase = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) __lowerCamelCase = -2 * np.random.rand(self.conva[1] ) + 1 __lowerCamelCase = -2 * np.random.rand(self.num_bpa ) + 1 __lowerCamelCase = -2 * np.random.rand(self.num_bpa ) + 1 def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Union[str, Any] ): # save model dict with pickle __lowerCamelCase = { """num_bp1""": self.num_bpa, """num_bp2""": self.num_bpa, """num_bp3""": self.num_bpa, """conv1""": self.conva, """step_conv1""": self.step_conva, """size_pooling1""": self.size_poolinga, """rate_weight""": self.rate_weight, """rate_thre""": self.rate_thre, """w_conv1""": self.w_conva, """wkj""": self.wkj, """vji""": self.vji, """thre_conv1""": self.thre_conva, """thre_bp2""": self.thre_bpa, """thre_bp3""": self.thre_bpa, } with open(UpperCamelCase_ , """wb""" ) as f: pickle.dump(UpperCamelCase_ , UpperCamelCase_ ) print(F'Model saved: {save_path}' ) @classmethod def lowerCAmelCase__ ( cls: Optional[int] , UpperCamelCase_: List[Any] ): # read saved model with open(UpperCamelCase_ , """rb""" ) as f: __lowerCamelCase = pickle.load(UpperCamelCase_ ) # noqa: S301 __lowerCamelCase = model_dic.get("""conv1""" ) conv_get.append(model_dic.get("""step_conv1""" ) ) __lowerCamelCase = model_dic.get("""size_pooling1""" ) __lowerCamelCase = model_dic.get("""num_bp1""" ) __lowerCamelCase = model_dic.get("""num_bp2""" ) __lowerCamelCase = model_dic.get("""num_bp3""" ) __lowerCamelCase = model_dic.get("""rate_weight""" ) __lowerCamelCase = model_dic.get("""rate_thre""" ) # create model instance __lowerCamelCase = CNN(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # modify model parameter __lowerCamelCase = model_dic.get("""w_conv1""" ) __lowerCamelCase = model_dic.get("""wkj""" ) __lowerCamelCase = model_dic.get("""vji""" ) __lowerCamelCase = model_dic.get("""thre_conv1""" ) __lowerCamelCase = model_dic.get("""thre_bp2""" ) __lowerCamelCase = model_dic.get("""thre_bp3""" ) return conv_ins def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Union[str, Any] ): return 1 / (1 + np.exp(-1 * x )) def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: int ): return round(UpperCamelCase_ , 3 ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Optional[int] , UpperCamelCase_: Optional[int] , UpperCamelCase_: Any , UpperCamelCase_: List[Any] ): # convolution process __lowerCamelCase = convs[0] __lowerCamelCase = convs[1] __lowerCamelCase = np.shape(UpperCamelCase_ )[0] # get the data slice of original image data, data_focus __lowerCamelCase = [] for i_focus in range(0 , size_data - size_conv + 1 , UpperCamelCase_ ): for j_focus in range(0 , size_data - size_conv + 1 , UpperCamelCase_ ): __lowerCamelCase = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(UpperCamelCase_ ) # calculate the feature map of every single kernel, and saved as list of matrix __lowerCamelCase = [] __lowerCamelCase = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(UpperCamelCase_ ): __lowerCamelCase = [] for i_focus in range(len(UpperCamelCase_ ) ): __lowerCamelCase = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(UpperCamelCase_ ) ) __lowerCamelCase = np.asmatrix(UpperCamelCase_ ).reshape( UpperCamelCase_ , UpperCamelCase_ ) data_featuremap.append(UpperCamelCase_ ) # expanding the data slice to One dimenssion __lowerCamelCase = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(UpperCamelCase_ ) ) __lowerCamelCase = np.asarray(UpperCamelCase_ ) return focus_list, data_featuremap def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: List[str] , UpperCamelCase_: Optional[int] , UpperCamelCase_: Optional[int]="average_pool" ): # pooling process __lowerCamelCase = len(featuremaps[0] ) __lowerCamelCase = int(size_map / size_pooling ) __lowerCamelCase = [] for i_map in range(len(UpperCamelCase_ ) ): __lowerCamelCase = featuremaps[i_map] __lowerCamelCase = [] for i_focus in range(0 , UpperCamelCase_ , UpperCamelCase_ ): for j_focus in range(0 , UpperCamelCase_ , UpperCamelCase_ ): __lowerCamelCase = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(UpperCamelCase_ ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(UpperCamelCase_ ) ) __lowerCamelCase = np.asmatrix(UpperCamelCase_ ).reshape(UpperCamelCase_ , UpperCamelCase_ ) featuremap_pooled.append(UpperCamelCase_ ) return featuremap_pooled def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: List[Any] ): # expanding three dimension data to one dimension list __lowerCamelCase = [] for i in range(len(UpperCamelCase_ ) ): __lowerCamelCase = np.shape(data[i] ) __lowerCamelCase = data[i].reshape(1 , shapes[0] * shapes[1] ) __lowerCamelCase = data_listed.getA().tolist()[0] data_expanded.extend(UpperCamelCase_ ) __lowerCamelCase = np.asarray(UpperCamelCase_ ) return data_expanded def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Tuple ): # expanding matrix to one dimension list __lowerCamelCase = np.asarray(UpperCamelCase_ ) __lowerCamelCase = np.shape(UpperCamelCase_ ) __lowerCamelCase = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Optional[int] , UpperCamelCase_: List[str] , UpperCamelCase_: Any , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Any ): __lowerCamelCase = [] __lowerCamelCase = 0 for i_map in range(UpperCamelCase_ ): __lowerCamelCase = np.ones((size_map, size_map) ) for i in range(0 , UpperCamelCase_ , UpperCamelCase_ ): for j in range(0 , UpperCamelCase_ , UpperCamelCase_ ): __lowerCamelCase = pd_pool[ i_pool ] __lowerCamelCase = i_pool + 1 __lowerCamelCase = np.multiply( UpperCamelCase_ , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(UpperCamelCase_ ) return pd_all def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Any , UpperCamelCase_: int , UpperCamelCase_: Dict , UpperCamelCase_: Tuple , UpperCamelCase_: int , UpperCamelCase_: List[str]=bool ): # model traning print("""----------------------Start Training-------------------------""" ) print((""" - - Shape: Train_Data """, np.shape(UpperCamelCase_ )) ) print((""" - - Shape: Teach_Data """, np.shape(UpperCamelCase_ )) ) __lowerCamelCase = 0 __lowerCamelCase = [] __lowerCamelCase = 1_00_00 while rp < n_repeat and mse >= error_accuracy: __lowerCamelCase = 0 print(F'-------------Learning Time {rp}--------------' ) for p in range(len(UpperCamelCase_ ) ): # print('------------Learning Image: %d--------------'%p) __lowerCamelCase = np.asmatrix(datas_train[p] ) __lowerCamelCase = np.asarray(datas_teach[p] ) __lowerCamelCase, __lowerCamelCase = self.convolute( UpperCamelCase_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) __lowerCamelCase = self.pooling(UpperCamelCase_ , self.size_poolinga ) __lowerCamelCase = np.shape(UpperCamelCase_ ) __lowerCamelCase = self._expand(UpperCamelCase_ ) __lowerCamelCase = data_bp_input __lowerCamelCase = np.dot(UpperCamelCase_ , self.vji.T ) - self.thre_bpa __lowerCamelCase = self.sig(UpperCamelCase_ ) __lowerCamelCase = np.dot(UpperCamelCase_ , self.wkj.T ) - self.thre_bpa __lowerCamelCase = self.sig(UpperCamelCase_ ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- __lowerCamelCase = np.multiply( (data_teach - bp_outa) , np.multiply(UpperCamelCase_ , (1 - bp_outa) ) ) __lowerCamelCase = np.multiply( np.dot(UpperCamelCase_ , self.wkj ) , np.multiply(UpperCamelCase_ , (1 - bp_outa) ) ) __lowerCamelCase = np.dot(UpperCamelCase_ , self.vji ) __lowerCamelCase = pd_i_all / (self.size_poolinga * self.size_poolinga) __lowerCamelCase = pd_conva_pooled.T.getA().tolist() __lowerCamelCase = self._calculate_gradient_from_pool( UpperCamelCase_ , UpperCamelCase_ , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): __lowerCamelCase = self._expand_mat(pd_conva_all[k_conv] ) __lowerCamelCase = self.rate_weight * np.dot(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) __lowerCamelCase = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer __lowerCamelCase = self.wkj + pd_k_all.T * bp_outa * self.rate_weight __lowerCamelCase = self.vji + pd_j_all.T * bp_outa * self.rate_weight __lowerCamelCase = self.thre_bpa - pd_k_all * self.rate_thre __lowerCamelCase = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image __lowerCamelCase = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) __lowerCamelCase = rp + 1 __lowerCamelCase = error_count / patterns all_mse.append(UpperCamelCase_ ) def draw_error(): __lowerCamelCase = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(UpperCamelCase_ , """+-""" ) plt.plot(UpperCamelCase_ , """r--""" ) plt.xlabel("""Learning Times""" ) plt.ylabel("""All_mse""" ) plt.grid(UpperCamelCase_ , alpha=0.5 ) plt.show() print("""------------------Training Complished---------------------""" ) print((""" - - Training epoch: """, rp, F' - - Mse: {mse:.6f}') ) if draw_e: draw_error() return mse def lowerCAmelCase__ ( self: Any , UpperCamelCase_: List[Any] ): # model predict __lowerCamelCase = [] print("""-------------------Start Testing-------------------------""" ) print((""" - - Shape: Test_Data """, np.shape(UpperCamelCase_ )) ) for p in range(len(UpperCamelCase_ ) ): __lowerCamelCase = np.asmatrix(datas_test[p] ) __lowerCamelCase, __lowerCamelCase = self.convolute( UpperCamelCase_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) __lowerCamelCase = self.pooling(UpperCamelCase_ , self.size_poolinga ) __lowerCamelCase = self._expand(UpperCamelCase_ ) __lowerCamelCase = data_bp_input __lowerCamelCase = bp_outa * self.vji.T - self.thre_bpa __lowerCamelCase = self.sig(UpperCamelCase_ ) __lowerCamelCase = bp_outa * self.wkj.T - self.thre_bpa __lowerCamelCase = self.sig(UpperCamelCase_ ) produce_out.extend(bp_outa.getA().tolist() ) __lowerCamelCase = [list(map(self.do_round , UpperCamelCase_ ) ) for each in produce_out] return np.asarray(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Dict ): # return the data of image after convoluting process so we can check it out __lowerCamelCase = np.asmatrix(UpperCamelCase_ ) __lowerCamelCase, __lowerCamelCase = self.convolute( UpperCamelCase_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) __lowerCamelCase = self.pooling(UpperCamelCase_ , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
701
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__: def __init__( self: Union[str, Any] , UpperCamelCase_: str = None , UpperCamelCase_: uuid.UUID = None , UpperCamelCase_: Dict=None , UpperCamelCase_: Any=None ): if not conversation_id: __lowerCamelCase = uuid.uuida() if past_user_inputs is None: __lowerCamelCase = [] if generated_responses is None: __lowerCamelCase = [] __lowerCamelCase = conversation_id __lowerCamelCase = past_user_inputs __lowerCamelCase = generated_responses __lowerCamelCase = text def __eq__( self: Optional[Any] , UpperCamelCase_: Union[str, Any] ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: str , UpperCamelCase_: bool = False ): if self.new_user_input: if overwrite: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' F'with: "{text}".' ) __lowerCamelCase = text else: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' F'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input' ) else: __lowerCamelCase = text def lowerCAmelCase__ ( self: List[str] ): if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __lowerCamelCase = None def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): self.generated_responses.append(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple ): for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self: Union[str, Any] ): __lowerCamelCase = F'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): __lowerCamelCase = """user""" if is_user else """bot""" output += F'{name} >> {text} \n' return output @add_end_docstrings( __lowerCamelCase , r'\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n ' , ) class lowerCamelCase__( __lowerCamelCase): def __init__( self: List[str] , *UpperCamelCase_: List[Any] , **UpperCamelCase_: str ): super().__init__(*UpperCamelCase_ , **UpperCamelCase_ ) if self.tokenizer.pad_token_id is None: __lowerCamelCase = self.tokenizer.eos_token def lowerCAmelCase__ ( self: str , UpperCamelCase_: int=None , UpperCamelCase_: Any=None , UpperCamelCase_: Union[str, Any]=None , **UpperCamelCase_: int ): __lowerCamelCase = {} __lowerCamelCase = {} __lowerCamelCase = {} if min_length_for_response is not None: __lowerCamelCase = min_length_for_response if minimum_tokens is not None: __lowerCamelCase = minimum_tokens if "max_length" in generate_kwargs: __lowerCamelCase = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __lowerCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCamelCase_ ) return preprocess_params, forward_params, postprocess_params def __call__( self: Any , UpperCamelCase_: Union[Conversation, List[Conversation]] , UpperCamelCase_: Optional[int]=0 , **UpperCamelCase_: Optional[int] ): __lowerCamelCase = super().__call__(UpperCamelCase_ , num_workers=UpperCamelCase_ , **UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and len(UpperCamelCase_ ) == 1: return outputs[0] return outputs def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Conversation , UpperCamelCase_: Optional[Any]=32 ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( F'Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ' """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): __lowerCamelCase = self.tokenizer._build_conversation_input_ids(UpperCamelCase_ ) else: # If the tokenizer cannot handle conversations, we default to only the old version __lowerCamelCase = self._legacy_parse_and_tokenize(UpperCamelCase_ ) if self.framework == "pt": __lowerCamelCase = torch.LongTensor([input_ids] ) elif self.framework == "tf": __lowerCamelCase = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str=10 , **UpperCamelCase_: List[str] ): __lowerCamelCase = generate_kwargs.get("""max_length""" , self.model.config.max_length ) __lowerCamelCase = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(F'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})' ) __lowerCamelCase = max_length - minimum_tokens __lowerCamelCase = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: __lowerCamelCase = model_inputs["""attention_mask"""][:, -trim:] __lowerCamelCase = model_inputs.pop("""conversation""" ) __lowerCamelCase = max_length __lowerCamelCase = self.model.generate(**UpperCamelCase_ , **UpperCamelCase_ ) if self.model.config.is_encoder_decoder: __lowerCamelCase = 1 else: __lowerCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Optional[Any] , UpperCamelCase_: int=True ): __lowerCamelCase = model_outputs["""output_ids"""] __lowerCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_ , ) __lowerCamelCase = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCamelCase_ ) return conversation def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Conversation ): __lowerCamelCase = self.tokenizer.eos_token_id __lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) ) if len(UpperCamelCase_ ) > self.tokenizer.model_max_length: __lowerCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
80
0
def lowerCamelCase__ ( A__ : int , A__ : int ): '''simple docstring''' while b: __lowerCamelCase, __lowerCamelCase = b, a % b return a def lowerCamelCase__ ( A__ : int , A__ : int ): '''simple docstring''' return a if b == 0 else euclidean_gcd_recursive(A__ , a % b ) def lowerCamelCase__ ( ): '''simple docstring''' print(f'euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}' ) print(f'euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}' ) print(f'euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}' ) print(f'euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}' ) print(f'euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}' ) print(f'euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}' ) print(f'euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}' ) print(f'euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}' ) print(f'euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}' ) print(f'euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}' ) if __name__ == "__main__": main()
702
import math def lowerCamelCase__ ( A__ : int ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = 2 __lowerCamelCase = int(math.sqrt(A__ ) ) # Size of every segment __lowerCamelCase = [True] * (end + 1) __lowerCamelCase = [] while start <= end: if temp[start] is True: in_prime.append(A__ ) for i in range(start * start , end + 1 , A__ ): __lowerCamelCase = False start += 1 prime += in_prime __lowerCamelCase = end + 1 __lowerCamelCase = min(2 * end , A__ ) while low <= n: __lowerCamelCase = [True] * (high - low + 1) for each in in_prime: __lowerCamelCase = math.floor(low / each ) * each if t < low: t += each for j in range(A__ , high + 1 , A__ ): __lowerCamelCase = False for j in range(len(A__ ) ): if temp[j] is True: prime.append(j + low ) __lowerCamelCase = high + 1 __lowerCamelCase = min(high + end , A__ ) return prime print(sieve(10**6))
80
0
def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = 0 for ch in input_str: __lowerCamelCase = ord(snake_case_ ) __lowerCamelCase = pow(2 , snake_case_ ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
703
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase_ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') class lowerCamelCase__( __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : int = BartphoTokenizer UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : List[str] = True def lowerCAmelCase__ ( self: Tuple ): super().setUp() __lowerCamelCase = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] __lowerCamelCase = dict(zip(UpperCamelCase_ , range(len(UpperCamelCase_ ) ) ) ) __lowerCamelCase = {"""unk_token""": """<unk>"""} __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F'{token} {vocab_tokens[token]}\n' ) __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase__ ( self: List[str] , **UpperCamelCase_: List[str] ): kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): __lowerCamelCase = """This is a là test""" __lowerCamelCase = """This is a<unk><unk> test""" return input_text, output_text def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) __lowerCamelCase = """This is a là test""" __lowerCamelCase = """▁This ▁is ▁a ▁l à ▁t est""".split() __lowerCamelCase = tokenizer.tokenize(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase_ ) , UpperCamelCase_ )
80
0
import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( A__ : Optional[Any] , A__ : str , A__ : List[str] ): '''simple docstring''' if gpta_config_file == "": __lowerCamelCase = GPTaConfig() else: __lowerCamelCase = GPTaConfig.from_json_file(_lowerCamelCase ) __lowerCamelCase = GPTaModel(_lowerCamelCase ) # Load weights from numpy load_tf_weights_in_gpta(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Save pytorch-model __lowerCamelCase = pytorch_dump_folder_path + "/" + WEIGHTS_NAME __lowerCamelCase = pytorch_dump_folder_path + "/" + CONFIG_NAME print(f'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() , _lowerCamelCase ) print(f'Save configuration file to {pytorch_config_dump_path}' ) with open(_lowerCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--gpt2_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--gpt2_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained OpenAI model. \n' 'This specifies the model architecture.' ), ) UpperCAmelCase_ = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
704
def lowerCamelCase__ ( A__ : dict ): '''simple docstring''' __lowerCamelCase = set() # To detect a back edge, keep track of vertices currently in the recursion stack __lowerCamelCase = set() return any( node not in visited and depth_first_search(A__ , A__ , A__ , A__ ) for node in graph ) def lowerCamelCase__ ( A__ : dict , A__ : int , A__ : set , A__ : set ): '''simple docstring''' visited.add(A__ ) rec_stk.add(A__ ) for node in graph[vertex]: if node not in visited: if depth_first_search(A__ , A__ , A__ , A__ ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(A__ ) return False if __name__ == "__main__": from doctest import testmod testmod()
80
0
def lowerCamelCase__ ( A__ : int ): if a < 0: raise ValueError("""Input value must be a positive integer""" ) elif isinstance(A__ , A__ ): raise TypeError("""Input value must be a 'int' type""" ) return bin(A__ ).count("""1""" ) if __name__ == "__main__": import doctest doctest.testmod()
705
from __future__ import annotations def lowerCamelCase__ ( A__ : list[float] , A__ : list[float] ): '''simple docstring''' __lowerCamelCase = sorted(numsa + numsa ) __lowerCamelCase, __lowerCamelCase = divmod(len(A__ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ = [float(x) for x in input('Enter the elements of first array: ').split()] UpperCAmelCase_ = [float(x) for x in input('Enter the elements of second array: ').split()] print(f"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
80
0
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class lowerCamelCase__: def __init__( self: Optional[Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int=1_00 , UpperCamelCase_: int=13 , UpperCamelCase_: List[Any]=30 , UpperCamelCase_: str=2 , UpperCamelCase_: Optional[Any]=3 , UpperCamelCase_: Optional[int]=True , UpperCamelCase_: Any=True , UpperCamelCase_: Optional[Any]=32 , UpperCamelCase_: Tuple=4 , UpperCamelCase_: str=4 , UpperCamelCase_: List[Any]=37 , UpperCamelCase_: List[str]="gelu" , UpperCamelCase_: str=0.1 , UpperCamelCase_: Optional[int]=0.1 , UpperCamelCase_: Any=10 , UpperCamelCase_: List[str]=0.02 , UpperCamelCase_: Dict=3 , UpperCamelCase_: str=None , UpperCamelCase_: Optional[int]=[0, 1, 2, 3] , ): __lowerCamelCase = parent __lowerCamelCase = 1_00 __lowerCamelCase = batch_size __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = scope __lowerCamelCase = out_indices __lowerCamelCase = num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCamelCase = (image_size // patch_size) ** 2 __lowerCamelCase = num_patches + 1 def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __lowerCamelCase = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCAmelCase__ ( self: List[Any] ): return BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=a_ , initializer_range=self.initializer_range , out_indices=self.out_indices , ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Optional[Any] , UpperCamelCase_: List[str] ): __lowerCamelCase = BeitModel(config=a_ ) model.to(a_ ) model.eval() __lowerCamelCase = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[int] , UpperCamelCase_: List[Any] ): __lowerCamelCase = BeitForMaskedImageModeling(config=a_ ) model.to(a_ ) model.eval() __lowerCamelCase = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def lowerCAmelCase__ ( self: Any , UpperCamelCase_: List[str] , UpperCamelCase_: Any , UpperCamelCase_: List[Any] , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = self.type_sequence_label_size __lowerCamelCase = BeitForImageClassification(a_ ) model.to(a_ ) model.eval() __lowerCamelCase = model(a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCamelCase = 1 __lowerCamelCase = BeitForImageClassification(a_ ) model.to(a_ ) model.eval() __lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCamelCase = model(a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Optional[int] , UpperCamelCase_: List[Any] , UpperCamelCase_: str , UpperCamelCase_: int ): __lowerCamelCase = self.num_labels __lowerCamelCase = BeitForSemanticSegmentation(a_ ) model.to(a_ ) model.eval() __lowerCamelCase = model(a_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) __lowerCamelCase = model(a_ , labels=a_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase = config_and_inputs __lowerCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase__( __a , __a , unittest.TestCase): UpperCAmelCase__ : Union[str, Any] = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) UpperCAmelCase__ : List[str] = ( { 'feature-extraction': BeitModel, 'image-classification': BeitForImageClassification, 'image-segmentation': BeitForSemanticSegmentation, } if is_torch_available() else {} ) UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : Union[str, Any] = False UpperCAmelCase__ : Any = False def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = BeitModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=a_ , has_text_modality=a_ , hidden_size=37 ) def lowerCAmelCase__ ( self: str ): self.config_tester.run_common_tests() @unittest.skip(reason="""BEiT does not use inputs_embeds""" ) def lowerCAmelCase__ ( self: Optional[int] ): pass @require_torch_multi_gpu @unittest.skip(reason="""BEiT has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowerCAmelCase__ ( self: str ): pass def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(a_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_ , nn.Linear ) ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(a_ ) __lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , a_ ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*a_ ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*a_ ) def lowerCAmelCase__ ( self: int ): if not self.model_tester.is_training: return __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(a_ ), BeitForMaskedImageModeling]: continue __lowerCamelCase = model_class(a_ ) model.to(a_ ) model.train() __lowerCamelCase = self._prepare_for_class(a_ , a_ , return_labels=a_ ) __lowerCamelCase = model(**a_ ).loss loss.backward() def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __lowerCamelCase = False __lowerCamelCase = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(a_ ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue __lowerCamelCase = model_class(a_ ) model.gradient_checkpointing_enable() model.to(a_ ) model.train() __lowerCamelCase = self._prepare_for_class(a_ , a_ , return_labels=a_ ) __lowerCamelCase = model(**a_ ).loss loss.backward() def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = _config_zero_init(a_ ) for model_class in self.all_model_classes: __lowerCamelCase = model_class(config=a_ ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @slow def lowerCAmelCase__ ( self: int ): for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = BeitModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class lowerCamelCase__( unittest.TestCase): @cached_property def lowerCAmelCase__ ( self: Dict ): return BeitImageProcessor.from_pretrained("""microsoft/beit-base-patch16-224""" ) if is_vision_available() else None @slow def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = BeitForMaskedImageModeling.from_pretrained("""microsoft/beit-base-patch16-224-pt22k""" ).to(a_ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=a_ , return_tensors="""pt""" ).pixel_values.to(a_ ) # prepare bool_masked_pos __lowerCamelCase = torch.ones((1, 1_96) , dtype=torch.bool ).to(a_ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(pixel_values=a_ , bool_masked_pos=a_ ) __lowerCamelCase = outputs.logits # verify the logits __lowerCamelCase = torch.Size((1, 1_96, 81_92) ) self.assertEqual(logits.shape , a_ ) __lowerCamelCase = torch.tensor( [[-3.2437, 0.5072, -13.9174], [-3.2456, 0.4948, -13.9401], [-3.2033, 0.5121, -13.8550]] ).to(a_ ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] , a_ , atol=1E-2 ) ) @slow def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = BeitForImageClassification.from_pretrained("""microsoft/beit-base-patch16-224""" ).to(a_ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=a_ , return_tensors="""pt""" ).to(a_ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**a_ ) __lowerCamelCase = outputs.logits # verify the logits __lowerCamelCase = torch.Size((1, 10_00) ) self.assertEqual(logits.shape , a_ ) __lowerCamelCase = torch.tensor([-1.2385, -1.0987, -1.0108] ).to(a_ ) self.assertTrue(torch.allclose(logits[0, :3] , a_ , atol=1E-4 ) ) __lowerCamelCase = 2_81 self.assertEqual(logits.argmax(-1 ).item() , a_ ) @slow def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = BeitForImageClassification.from_pretrained("""microsoft/beit-large-patch16-224-pt22k-ft22k""" ).to( a_ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=a_ , return_tensors="""pt""" ).to(a_ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**a_ ) __lowerCamelCase = outputs.logits # verify the logits __lowerCamelCase = torch.Size((1, 2_18_41) ) self.assertEqual(logits.shape , a_ ) __lowerCamelCase = torch.tensor([1.6881, -0.2787, 0.5901] ).to(a_ ) self.assertTrue(torch.allclose(logits[0, :3] , a_ , atol=1E-4 ) ) __lowerCamelCase = 23_96 self.assertEqual(logits.argmax(-1 ).item() , a_ ) @slow def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = BeitForSemanticSegmentation.from_pretrained("""microsoft/beit-base-finetuned-ade-640-640""" ) __lowerCamelCase = model.to(a_ ) __lowerCamelCase = BeitImageProcessor(do_resize=a_ , size=6_40 , do_center_crop=a_ ) __lowerCamelCase = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) __lowerCamelCase = Image.open(ds[0]["""file"""] ) __lowerCamelCase = image_processor(images=a_ , return_tensors="""pt""" ).to(a_ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**a_ ) __lowerCamelCase = outputs.logits # verify the logits __lowerCamelCase = torch.Size((1, 1_50, 1_60, 1_60) ) self.assertEqual(logits.shape , a_ ) __lowerCamelCase = version.parse(PIL.__version__ ) < version.parse("""9.0.0""" ) if is_pillow_less_than_a: __lowerCamelCase = torch.tensor( [ [[-4.9225, -2.3954, -3.0522], [-2.8822, -1.0046, -1.7561], [-2.9549, -1.3228, -2.1347]], [[-5.8168, -3.4129, -4.0778], [-3.8651, -2.2214, -3.0277], [-3.8356, -2.4643, -3.3535]], [[-0.0078, 3.9952, 4.0754], [2.9856, 4.6944, 5.0035], [3.2413, 4.7813, 4.9969]], ] , device=a_ , ) else: __lowerCamelCase = torch.tensor( [ [[-4.8960, -2.3688, -3.0355], [-2.8478, -0.9836, -1.7418], [-2.9449, -1.3332, -2.1456]], [[-5.8081, -3.4124, -4.1006], [-3.8561, -2.2081, -3.0323], [-3.8365, -2.4601, -3.3669]], [[-0.0309, 3.9868, 4.0540], [2.9640, 4.6877, 4.9976], [3.2081, 4.7690, 4.9942]], ] , device=a_ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , a_ , atol=1E-4 ) ) @slow def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = BeitForSemanticSegmentation.from_pretrained("""microsoft/beit-base-finetuned-ade-640-640""" ) __lowerCamelCase = model.to(a_ ) __lowerCamelCase = BeitImageProcessor(do_resize=a_ , size=6_40 , do_center_crop=a_ ) __lowerCamelCase = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) __lowerCamelCase = Image.open(ds[0]["""file"""] ) __lowerCamelCase = image_processor(images=a_ , return_tensors="""pt""" ).to(a_ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**a_ ) __lowerCamelCase = outputs.logits.detach().cpu() __lowerCamelCase = image_processor.post_process_semantic_segmentation(outputs=a_ , target_sizes=[(5_00, 3_00)] ) __lowerCamelCase = torch.Size((5_00, 3_00) ) self.assertEqual(segmentation[0].shape , a_ ) __lowerCamelCase = image_processor.post_process_semantic_segmentation(outputs=a_ ) __lowerCamelCase = torch.Size((1_60, 1_60) ) self.assertEqual(segmentation[0].shape , a_ )
706
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) self.assertTrue(torch.allclose(gelu_python(UpperCamelCase_ ) , torch_builtin(UpperCamelCase_ ) ) ) self.assertFalse(torch.allclose(gelu_python(UpperCamelCase_ ) , gelu_new(UpperCamelCase_ ) ) ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = get_activation("""gelu_10""" ) __lowerCamelCase = torch_builtin(UpperCamelCase_ ) __lowerCamelCase = geluaa(UpperCamelCase_ ) __lowerCamelCase = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(UpperCamelCase_ ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def lowerCAmelCase__ ( self: str ): get_activation("""gelu""" ) get_activation("""gelu_10""" ) get_activation("""gelu_fast""" ) get_activation("""gelu_new""" ) get_activation("""gelu_python""" ) get_activation("""gelu_pytorch_tanh""" ) get_activation("""linear""" ) get_activation("""mish""" ) get_activation("""quick_gelu""" ) get_activation("""relu""" ) get_activation("""sigmoid""" ) get_activation("""silu""" ) get_activation("""swish""" ) get_activation("""tanh""" ) with self.assertRaises(UpperCamelCase_ ): get_activation("""bogus""" ) with self.assertRaises(UpperCamelCase_ ): get_activation(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = 1 __lowerCamelCase = get_activation("""gelu""" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(UpperCamelCase_ ): __lowerCamelCase = acta.a
80
0
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : Optional[Any] = StableDiffusionDiffEditPipeline UpperCAmelCase__ : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'height', 'width', 'image'} | {'image_latents'} UpperCAmelCase__ : Optional[Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {'image'} | {'image_latents'} UpperCAmelCase__ : List[Any] = frozenset( []) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess UpperCAmelCase__ : Union[str, Any] = frozenset([]) def lowerCAmelCase__ ( self: Dict ): torch.manual_seed(0 ) __lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=__lowerCAmelCase , ) __lowerCamelCase = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=__lowerCAmelCase , set_alpha_to_one=__lowerCAmelCase , ) __lowerCamelCase = DDIMInverseScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=__lowerCAmelCase , set_alpha_to_zero=__lowerCAmelCase , ) torch.manual_seed(0 ) __lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) __lowerCamelCase = CLIPTextModel(__lowerCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) __lowerCamelCase = { """unet""": unet, """scheduler""": scheduler, """inverse_scheduler""": inverse_scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Optional[int] , UpperCamelCase_: str=0 ): __lowerCamelCase = floats_tensor((1, 16, 16) , rng=random.Random(__lowerCAmelCase ) ).to(__lowerCAmelCase ) __lowerCamelCase = floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(__lowerCAmelCase ) ).to(__lowerCAmelCase ) if str(__lowerCAmelCase ).startswith("""mps""" ): __lowerCamelCase = torch.manual_seed(__lowerCAmelCase ) else: __lowerCamelCase = torch.Generator(device=__lowerCAmelCase ).manual_seed(__lowerCAmelCase ) __lowerCamelCase = { """prompt""": """a dog and a newt""", """mask_image""": mask, """image_latents""": latents, """generator""": generator, """num_inference_steps""": 2, """inpaint_strength""": 1.0, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: List[str] , UpperCamelCase_: Optional[int]=0 ): __lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowerCAmelCase ) ).to(__lowerCAmelCase ) __lowerCamelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCamelCase = Image.fromarray(np.uinta(__lowerCAmelCase ) ).convert("""RGB""" ) if str(__lowerCAmelCase ).startswith("""mps""" ): __lowerCamelCase = torch.manual_seed(__lowerCAmelCase ) else: __lowerCamelCase = torch.Generator(device=__lowerCAmelCase ).manual_seed(__lowerCAmelCase ) __lowerCamelCase = { """image""": image, """source_prompt""": """a cat and a frog""", """target_prompt""": """a dog and a newt""", """generator""": generator, """num_inference_steps""": 2, """num_maps_per_mask""": 2, """mask_encode_strength""": 1.0, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[Any]=0 ): __lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowerCAmelCase ) ).to(__lowerCAmelCase ) __lowerCamelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCamelCase = Image.fromarray(np.uinta(__lowerCAmelCase ) ).convert("""RGB""" ) if str(__lowerCAmelCase ).startswith("""mps""" ): __lowerCamelCase = torch.manual_seed(__lowerCAmelCase ) else: __lowerCamelCase = torch.Generator(device=__lowerCAmelCase ).manual_seed(__lowerCAmelCase ) __lowerCamelCase = { """image""": image, """prompt""": """a cat and a frog""", """generator""": generator, """num_inference_steps""": 2, """inpaint_strength""": 1.0, """guidance_scale""": 6.0, """decode_latents""": True, """output_type""": """numpy""", } return inputs def lowerCAmelCase__ ( self: Optional[int] ): if not hasattr(self.pipeline_class , """_optional_components""" ): return __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) __lowerCamelCase = self.get_dummy_inputs(__lowerCAmelCase ) __lowerCamelCase = pipe(**__lowerCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__lowerCAmelCase ) __lowerCamelCase = self.pipeline_class.from_pretrained(__lowerCAmelCase ) pipe_loaded.to(__lowerCAmelCase ) pipe_loaded.set_progress_bar_config(disable=__lowerCAmelCase ) for optional_component in pipe._optional_components: self.assertTrue( getattr(__lowerCAmelCase , __lowerCAmelCase ) is None , F'`{optional_component}` did not stay set to None after loading.' , ) __lowerCamelCase = self.get_dummy_inputs(__lowerCAmelCase ) __lowerCamelCase = pipe_loaded(**__lowerCAmelCase )[0] __lowerCamelCase = np.abs(output - output_loaded ).max() self.assertLess(__lowerCAmelCase , 1E-4 ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = """cpu""" __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) __lowerCamelCase = self.get_dummy_mask_inputs(__lowerCAmelCase ) __lowerCamelCase = pipe.generate_mask(**__lowerCAmelCase ) __lowerCamelCase = mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) __lowerCamelCase = np.array([0] * 9 ) __lowerCamelCase = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(__lowerCAmelCase , 1E-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = """cpu""" __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) __lowerCamelCase = self.get_dummy_inversion_inputs(__lowerCAmelCase ) __lowerCamelCase = pipe.invert(**__lowerCAmelCase ).images __lowerCamelCase = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __lowerCamelCase = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.5_1050, 0.5015, 0.4407, 0.4799] , ) __lowerCamelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__lowerCAmelCase , 1E-3 ) def lowerCAmelCase__ ( self: Optional[int] ): super().test_inference_batch_single_identical(expected_max_diff=5E-3 ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = """cpu""" __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = {"""beta_start""": 0.0_0085, """beta_end""": 0.012, """beta_schedule""": """scaled_linear"""} __lowerCamelCase = DPMSolverMultistepScheduler(**__lowerCAmelCase ) __lowerCamelCase = DPMSolverMultistepInverseScheduler(**__lowerCAmelCase ) __lowerCamelCase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) __lowerCamelCase = self.get_dummy_inversion_inputs(__lowerCAmelCase ) __lowerCamelCase = pipe.invert(**__lowerCAmelCase ).images __lowerCamelCase = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __lowerCamelCase = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.5_1050, 0.5015, 0.4407, 0.4799] , ) __lowerCamelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__lowerCAmelCase , 1E-3 ) @require_torch_gpu @slow class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Union[str, Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def lowerCAmelCase__ ( cls: List[Any] ): __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png""" ) __lowerCamelCase = raw_image.convert("""RGB""" ).resize((7_68, 7_68) ) __lowerCamelCase = raw_image def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = StableDiffusionDiffEditPipeline.from_pretrained( """stabilityai/stable-diffusion-2-1""" , safety_checker=__lowerCAmelCase , torch_dtype=torch.floataa ) __lowerCamelCase = DDIMScheduler.from_config(pipe.scheduler.config ) __lowerCamelCase = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__lowerCAmelCase ) __lowerCamelCase = """a bowl of fruit""" __lowerCamelCase = """a bowl of pears""" __lowerCamelCase = pipe.generate_mask( image=self.raw_image , source_prompt=__lowerCAmelCase , target_prompt=__lowerCAmelCase , generator=__lowerCAmelCase , ) __lowerCamelCase = pipe.invert( prompt=__lowerCAmelCase , image=self.raw_image , inpaint_strength=0.7 , generator=__lowerCAmelCase ).latents __lowerCamelCase = pipe( prompt=__lowerCAmelCase , mask_image=__lowerCAmelCase , image_latents=__lowerCAmelCase , generator=__lowerCAmelCase , negative_prompt=__lowerCAmelCase , inpaint_strength=0.7 , output_type="""numpy""" , ).images[0] __lowerCamelCase = ( np.array( load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/diffedit/pears.png""" ).resize((7_68, 7_68) ) ) / 2_55 ) assert np.abs((expected_image - image).max() ) < 5E-1 def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = StableDiffusionDiffEditPipeline.from_pretrained( """stabilityai/stable-diffusion-2-1""" , safety_checker=__lowerCAmelCase , torch_dtype=torch.floataa ) __lowerCamelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) __lowerCamelCase = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__lowerCAmelCase ) __lowerCamelCase = """a bowl of fruit""" __lowerCamelCase = """a bowl of pears""" __lowerCamelCase = pipe.generate_mask( image=self.raw_image , source_prompt=__lowerCAmelCase , target_prompt=__lowerCAmelCase , generator=__lowerCAmelCase , ) __lowerCamelCase = pipe.invert( prompt=__lowerCAmelCase , image=self.raw_image , inpaint_strength=0.7 , generator=__lowerCAmelCase , num_inference_steps=25 , ).latents __lowerCamelCase = pipe( prompt=__lowerCAmelCase , mask_image=__lowerCAmelCase , image_latents=__lowerCAmelCase , generator=__lowerCAmelCase , negative_prompt=__lowerCAmelCase , inpaint_strength=0.7 , num_inference_steps=25 , output_type="""numpy""" , ).images[0] __lowerCamelCase = ( np.array( load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/diffedit/pears.png""" ).resize((7_68, 7_68) ) ) / 2_55 ) assert np.abs((expected_image - image).max() ) < 5E-1
707
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class lowerCamelCase__( __lowerCamelCase): @slow @require_torch def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = EncoderDecoderModel.from_encoder_decoder_pretrained("""prajjwal1/bert-tiny""" , """prajjwal1/bert-tiny""" ) __lowerCamelCase = BertTokenizer.from_pretrained("""bert-base-uncased""" ) __lowerCamelCase = bertabert.config.encoder.vocab_size __lowerCamelCase = tokenizer.sep_token_id __lowerCamelCase = tokenizer.cls_token_id __lowerCamelCase = 1_28 __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""train[:1%]""" ) __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""validation[:1%]""" ) __lowerCamelCase = train_dataset.select(range(32 ) ) __lowerCamelCase = val_dataset.select(range(16 ) ) __lowerCamelCase = 4 def _map_to_encoder_decoder_inputs(UpperCamelCase_: List[Any] ): # Tokenizer will automatically set [BOS] <text> [EOS] __lowerCamelCase = tokenizer(batch["""article"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=5_12 ) __lowerCamelCase = tokenizer(batch["""highlights"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=1_28 ) __lowerCamelCase = inputs.input_ids __lowerCamelCase = inputs.attention_mask __lowerCamelCase = outputs.input_ids __lowerCamelCase = outputs.input_ids.copy() __lowerCamelCase = [ [-1_00 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["""labels"""] ] __lowerCamelCase = outputs.attention_mask assert all(len(UpperCamelCase_ ) == 5_12 for x in inputs.input_ids ) assert all(len(UpperCamelCase_ ) == 1_28 for x in outputs.input_ids ) return batch def _compute_metrics(UpperCamelCase_: int ): __lowerCamelCase = pred.label_ids __lowerCamelCase = pred.predictions # all unnecessary tokens are removed __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = sum([int(pred_str[i] == label_str[i] ) for i in range(len(UpperCamelCase_ ) )] ) / len(UpperCamelCase_ ) return {"accuracy": accuracy} # map train dataset __lowerCamelCase = train_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) train_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) # same for validation dataset __lowerCamelCase = val_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) val_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) __lowerCamelCase = self.get_auto_remove_tmp_dir() __lowerCamelCase = SeqaSeqTrainingArguments( output_dir=UpperCamelCase_ , per_device_train_batch_size=UpperCamelCase_ , per_device_eval_batch_size=UpperCamelCase_ , predict_with_generate=UpperCamelCase_ , evaluation_strategy="""steps""" , do_train=UpperCamelCase_ , do_eval=UpperCamelCase_ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer __lowerCamelCase = SeqaSeqTrainer( model=UpperCamelCase_ , args=UpperCamelCase_ , compute_metrics=_compute_metrics , train_dataset=UpperCamelCase_ , eval_dataset=UpperCamelCase_ , tokenizer=UpperCamelCase_ , ) # start training trainer.train()
80
0
def lowerCamelCase__ ( A__ : int ): if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): raise TypeError("""only integers accepted as input""" ) else: __lowerCamelCase = str(abs(UpperCAmelCase__ ) ) __lowerCamelCase = [list(UpperCAmelCase__ ) for char in range(len(UpperCAmelCase__ ) )] for index in range(len(UpperCAmelCase__ ) ): num_transpositions[index].pop(UpperCAmelCase__ ) return max( int("""""".join(list(UpperCAmelCase__ ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__('doctest').testmod()
708
class lowerCamelCase__: # Public class to implement a graph def __init__( self: Dict , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): __lowerCamelCase = row __lowerCamelCase = col __lowerCamelCase = graph def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): # Checking all 8 elements surrounding nth element __lowerCamelCase = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __lowerCamelCase = [-1, 0, 1, -1, 1, -1, 0, 1] __lowerCamelCase = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): # And finally, count all islands. __lowerCamelCase = [[False for j in range(self.COL )] for i in range(self.ROW )] __lowerCamelCase = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) count += 1 return count
80
0
import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = '''▁''' UpperCAmelCase_ = { '''vocab_file''': '''vocab.json''', '''spm_file''': '''sentencepiece.bpe.model''', } UpperCAmelCase_ = { '''vocab_file''': { '''facebook/s2t-small-librispeech-asr''': ( '''https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/vocab.json''' ), }, '''spm_file''': { '''facebook/s2t-small-librispeech-asr''': ( '''https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/sentencepiece.bpe.model''' ) }, } UpperCAmelCase_ = { '''facebook/s2t-small-librispeech-asr''': 1_024, } UpperCAmelCase_ = ['''pt''', '''fr''', '''ru''', '''nl''', '''ro''', '''it''', '''es''', '''de'''] UpperCAmelCase_ = {'''mustc''': MUSTC_LANGS} class lowerCamelCase__( a__): UpperCAmelCase__ : Optional[Any] = VOCAB_FILES_NAMES UpperCAmelCase__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : List[Any] = MAX_MODEL_INPUT_SIZES UpperCAmelCase__ : List[str] = ["""input_ids""", """attention_mask"""] UpperCAmelCase__ : List[int] = [] def __init__( self: Union[str, Any] , UpperCamelCase_: List[str] , UpperCamelCase_: Dict , UpperCamelCase_: int="<s>" , UpperCamelCase_: Union[str, Any]="</s>" , UpperCamelCase_: Union[str, Any]="<pad>" , UpperCamelCase_: List[Any]="<unk>" , UpperCamelCase_: Tuple=False , UpperCamelCase_: str=False , UpperCamelCase_: Dict=None , UpperCamelCase_: List[str]=None , UpperCamelCase_: Dict = None , **UpperCamelCase_: List[Any] , ): __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , do_upper_case=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , lang_codes=lowerCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase__ , ) __lowerCamelCase = do_upper_case __lowerCamelCase = do_lower_case __lowerCamelCase = load_json(lowerCAmelCase__ ) __lowerCamelCase = {v: k for k, v in self.encoder.items()} __lowerCamelCase = spm_file __lowerCamelCase = load_spm(lowerCAmelCase__ , self.sp_model_kwargs ) if lang_codes is not None: __lowerCamelCase = lang_codes __lowerCamelCase = LANGUAGES[lang_codes] __lowerCamelCase = [F'<lang:{lang}>' for lang in self.langs] __lowerCamelCase = {lang: self.sp_model.PieceToId(F'<lang:{lang}>' ) for lang in self.langs} __lowerCamelCase = self.lang_tokens __lowerCamelCase = tgt_lang if tgt_lang is not None else self.langs[0] self.set_tgt_lang_special_tokens(self._tgt_lang ) else: __lowerCamelCase = {} @property def lowerCAmelCase__ ( self: int ): return len(self.encoder ) @property def lowerCAmelCase__ ( self: List[str] ): return self._tgt_lang @tgt_lang.setter def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: List[str] ): __lowerCamelCase = new_tgt_lang self.set_tgt_lang_special_tokens(lowerCAmelCase__ ) def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: int ): __lowerCamelCase = self.lang_code_to_id[tgt_lang] __lowerCamelCase = [lang_code_id] def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: List[str] ): return self.sp_model.encode(lowerCAmelCase__ , out_type=lowerCAmelCase__ ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: Optional[int] ): return self.encoder.get(lowerCAmelCase__ , self.encoder[self.unk_token] ) def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Optional[int] ): return self.decoder.get(lowerCAmelCase__ , self.unk_token ) def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Tuple ): __lowerCamelCase = [] __lowerCamelCase = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: __lowerCamelCase = self.sp_model.decode(lowerCAmelCase__ ) out_string += (decoded.upper() if self.do_upper_case else decoded) + token + " " __lowerCamelCase = [] else: current_sub_tokens.append(lowerCAmelCase__ ) __lowerCamelCase = self.sp_model.decode(lowerCAmelCase__ ) out_string += decoded.upper() if self.do_upper_case else decoded return out_string.strip() def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Any , UpperCamelCase_: Optional[int]=None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + [self.eos_token_id] def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: Tuple , UpperCamelCase_: List[Any] = None , UpperCamelCase_: Dict = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__ ) __lowerCamelCase = [1] * len(self.prefix_tokens ) __lowerCamelCase = [1] if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase__ )) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase__ )) + ([0] * len(lowerCAmelCase__ )) + suffix_ones def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = self.encoder.copy() vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self: Union[str, Any] ): __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self: Optional[int] , UpperCamelCase_: str ): __lowerCamelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __lowerCamelCase = {} __lowerCamelCase = load_spm(self.spm_file , self.sp_model_kwargs ) def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: Optional[int] , UpperCamelCase_: int = None ): __lowerCamelCase = Path(lowerCAmelCase__ ) assert save_dir.is_dir(), F'{save_directory} should be a directory' __lowerCamelCase = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["vocab_file"] ) __lowerCamelCase = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["spm_file"] ) save_json(self.encoder , lowerCAmelCase__ ) if os.path.abspath(self.spm_file ) != os.path.abspath(lowerCAmelCase__ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , lowerCAmelCase__ ) elif not os.path.isfile(self.spm_file ): with open(lowerCAmelCase__ , """wb""" ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__ ) return (str(lowerCAmelCase__ ), str(lowerCAmelCase__ )) def lowerCamelCase__ ( A__ : str , A__ : Dict[str, Any] ): '''simple docstring''' __lowerCamelCase = sentencepiece.SentencePieceProcessor(**_lowercase ) spm.Load(str(_lowercase ) ) return spm def lowerCamelCase__ ( A__ : str ): '''simple docstring''' with open(_lowercase , """r""" ) as f: return json.load(_lowercase ) def lowerCamelCase__ ( A__ : Dict , A__ : str ): '''simple docstring''' with open(_lowercase , """w""" ) as f: json.dump(_lowercase , _lowercase , indent=2 )
709
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = DPTConfig() if "large" in checkpoint_url: __lowerCamelCase = 1024 __lowerCamelCase = 4096 __lowerCamelCase = 24 __lowerCamelCase = 16 __lowerCamelCase = [5, 11, 17, 23] __lowerCamelCase = [256, 512, 1024, 1024] __lowerCamelCase = (1, 384, 384) if "ade" in checkpoint_url: __lowerCamelCase = True __lowerCamelCase = 150 __lowerCamelCase = """huggingface/label-files""" __lowerCamelCase = """ade20k-id2label.json""" __lowerCamelCase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type="""dataset""" ) ) , """r""" ) ) __lowerCamelCase = {int(A__ ): v for k, v in idalabel.items()} __lowerCamelCase = idalabel __lowerCamelCase = {v: k for k, v in idalabel.items()} __lowerCamelCase = [1, 150, 480, 480] return config, expected_shape def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' __lowerCamelCase = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.encoder""" ) if "pretrained.model" in name: __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.embeddings""" ) if "patch_embed" in name: __lowerCamelCase = name.replace("""patch_embed""" , """patch_embeddings""" ) if "pos_embed" in name: __lowerCamelCase = name.replace("""pos_embed""" , """position_embeddings""" ) if "attn.proj" in name: __lowerCamelCase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "proj" in name and "project" not in name: __lowerCamelCase = name.replace("""proj""" , """projection""" ) if "blocks" in name: __lowerCamelCase = name.replace("""blocks""" , """layer""" ) if "mlp.fc1" in name: __lowerCamelCase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: __lowerCamelCase = name.replace("""mlp.fc2""" , """output.dense""" ) if "norm1" in name: __lowerCamelCase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: __lowerCamelCase = name.replace("""norm2""" , """layernorm_after""" ) if "scratch.output_conv" in name: __lowerCamelCase = name.replace("""scratch.output_conv""" , """head""" ) if "scratch" in name: __lowerCamelCase = name.replace("""scratch""" , """neck""" ) if "layer1_rn" in name: __lowerCamelCase = name.replace("""layer1_rn""" , """convs.0""" ) if "layer2_rn" in name: __lowerCamelCase = name.replace("""layer2_rn""" , """convs.1""" ) if "layer3_rn" in name: __lowerCamelCase = name.replace("""layer3_rn""" , """convs.2""" ) if "layer4_rn" in name: __lowerCamelCase = name.replace("""layer4_rn""" , """convs.3""" ) if "refinenet" in name: __lowerCamelCase = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __lowerCamelCase = name.replace(f'refinenet{layer_idx}' , f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: __lowerCamelCase = name.replace("""out_conv""" , """projection""" ) if "resConfUnit1" in name: __lowerCamelCase = name.replace("""resConfUnit1""" , """residual_layer1""" ) if "resConfUnit2" in name: __lowerCamelCase = name.replace("""resConfUnit2""" , """residual_layer2""" ) if "conv1" in name: __lowerCamelCase = name.replace("""conv1""" , """convolution1""" ) if "conv2" in name: __lowerCamelCase = name.replace("""conv2""" , """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.0.project.0""" , """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.0.project.0""" , """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.0.project.0""" , """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.0.project.0""" , """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.3""" , """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.4""" , """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.3""" , """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.4""" , """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.3""" , """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.3""" , """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.4""" , """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: __lowerCamelCase = name.replace("""pretrained""" , """dpt""" ) if "bn" in name: __lowerCamelCase = name.replace("""bn""" , """batch_norm""" ) if "head" in name: __lowerCamelCase = name.replace("""head""" , """head.head""" ) if "encoder.norm" in name: __lowerCamelCase = name.replace("""encoder.norm""" , """layernorm""" ) if "auxlayer" in name: __lowerCamelCase = name.replace("""auxlayer""" , """auxiliary_head.head""" ) return name def lowerCamelCase__ ( A__ : Tuple , A__ : Any ): '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __lowerCamelCase = in_proj_weight[: config.hidden_size, :] __lowerCamelCase = in_proj_bias[: config.hidden_size] __lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] __lowerCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" __lowerCamelCase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( A__ : Optional[int] , A__ : Union[str, Any] , A__ : List[str] , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = get_dpt_config(A__ ) # load original state_dict from URL __lowerCamelCase = torch.hub.load_state_dict_from_url(A__ , map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(A__ ) # rename keys for key in state_dict.copy().keys(): __lowerCamelCase = state_dict.pop(A__ ) __lowerCamelCase = val # read in qkv matrices read_in_q_k_v(A__ , A__ ) # load HuggingFace model __lowerCamelCase = DPTForSemanticSegmentation(A__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(A__ ) model.load_state_dict(A__ ) model.eval() # Check outputs on an image __lowerCamelCase = 480 if """ade""" in checkpoint_url else 384 __lowerCamelCase = DPTImageProcessor(size=A__ ) __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(A__ , return_tensors="""pt""" ) # forward pass __lowerCamelCase = model(**A__ ).logits if """ade""" in checkpoint_url else model(**A__ ).predicted_depth # Assert logits __lowerCamelCase = torch.tensor([[6.3_199, 6.3_629, 6.4_148], [6.3_850, 6.3_615, 6.4_166], [6.3_519, 6.3_176, 6.3_575]] ) if "ade" in checkpoint_url: __lowerCamelCase = torch.tensor([[4.0_480, 4.2_420, 4.4_360], [4.3_124, 4.5_693, 4.8_261], [4.5_768, 4.8_965, 5.2_163]] ) assert outputs.shape == torch.Size(A__ ) assert ( torch.allclose(outputs[0, 0, :3, :3] , A__ , atol=1E-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3] , A__ ) ) Path(A__ ).mkdir(exist_ok=A__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(A__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=A__ , ) image_processor.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=A__ , ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) UpperCAmelCase_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
80
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = [[1, 2, 4], [1, 2, 3, 4]] __lowerCamelCase = DisjunctiveConstraint(__lowerCAmelCase ) self.assertTrue(isinstance(dc.token_ids , __lowerCAmelCase ) ) with self.assertRaises(__lowerCAmelCase ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(__lowerCAmelCase ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(__lowerCAmelCase ): DisjunctiveConstraint(__lowerCAmelCase ) # fails here def lowerCAmelCase__ ( self: int ): __lowerCamelCase = [[1, 2, 3], [1, 2, 4]] __lowerCamelCase = DisjunctiveConstraint(__lowerCAmelCase ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(1 ) __lowerCamelCase = stepped is True and completed is False and reset is False self.assertTrue(__lowerCAmelCase ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(2 ) __lowerCamelCase = stepped is True and completed is False and reset is False self.assertTrue(__lowerCAmelCase ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(3 ) __lowerCamelCase = stepped is True and completed is True and reset is False self.assertTrue(__lowerCAmelCase ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] __lowerCamelCase = DisjunctiveConstraint(__lowerCAmelCase ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
710
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
80
0
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def lowerCamelCase__ ( A__ : Any ): '''simple docstring''' __lowerCamelCase = args.pruning_method __lowerCamelCase = args.threshold __lowerCamelCase = args.model_name_or_path.rstrip("""/""" ) __lowerCamelCase = args.target_model_path print(f'Load fine-pruned model from {model_name_or_path}' ) __lowerCamelCase = torch.load(os.path.join(_snake_case , """pytorch_model.bin""" ) ) __lowerCamelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: __lowerCamelCase = tensor print(f'Copied layer {name}' ) elif "classifier" in name or "qa_output" in name: __lowerCamelCase = tensor print(f'Copied layer {name}' ) elif "bias" in name: __lowerCamelCase = tensor print(f'Copied layer {name}' ) else: if pruning_method == "magnitude": __lowerCamelCase = MagnitudeBinarizer.apply(inputs=_snake_case , threshold=_snake_case ) __lowerCamelCase = tensor * mask print(f'Pruned layer {name}' ) elif pruning_method == "topK": if "mask_scores" in name: continue __lowerCamelCase = name[:-6] __lowerCamelCase = model[f'{prefix_}mask_scores'] __lowerCamelCase = TopKBinarizer.apply(_snake_case , _snake_case ) __lowerCamelCase = tensor * mask print(f'Pruned layer {name}' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue __lowerCamelCase = name[:-6] __lowerCamelCase = model[f'{prefix_}mask_scores'] __lowerCamelCase = ThresholdBinarizer.apply(_snake_case , _snake_case , _snake_case ) __lowerCamelCase = tensor * mask print(f'Pruned layer {name}' ) elif pruning_method == "l0": if "mask_scores" in name: continue __lowerCamelCase = name[:-6] __lowerCamelCase = model[f'{prefix_}mask_scores'] __lowerCamelCase, __lowerCamelCase = -0.1, 1.1 __lowerCamelCase = torch.sigmoid(_snake_case ) __lowerCamelCase = s * (r - l) + l __lowerCamelCase = s_bar.clamp(min=0.0 , max=1.0 ) __lowerCamelCase = tensor * mask print(f'Pruned layer {name}' ) else: raise ValueError("""Unknown pruning method""" ) if target_model_path is None: __lowerCamelCase = os.path.join( os.path.dirname(_snake_case ) , f'bertarized_{os.path.basename(_snake_case )}' ) if not os.path.isdir(_snake_case ): shutil.copytree(_snake_case , _snake_case ) print(f'\nCreated folder {target_model_path}' ) torch.save(_snake_case , os.path.join(_snake_case , """pytorch_model.bin""" ) ) print("""\nPruned model saved! See you later!""" ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) UpperCAmelCase_ = parser.parse_args() main(args)
711
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/config.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/config.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/config.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/config.json', 'bert-base-multilingual-uncased': 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json', 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/config.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/config.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-base-cased-finetuned-mrpc': 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json', 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json', 'bert-base-german-dbmdz-uncased': 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json', 'cl-tohoku/bert-base-japanese': 'https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json', 'cl-tohoku/bert-base-japanese-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json' ), 'wietsedv/bert-base-dutch-cased': 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json', # See all BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = 'bert' def __init__( self: List[str] , UpperCamelCase_: str=3_05_22 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Tuple=12 , UpperCamelCase_: int=12 , UpperCamelCase_: int=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[Any]=0.1 , UpperCamelCase_: Optional[int]=5_12 , UpperCamelCase_: List[Any]=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: List[str]=1E-12 , UpperCamelCase_: Dict=0 , UpperCamelCase_: List[Any]="absolute" , UpperCamelCase_: Tuple=True , UpperCamelCase_: Tuple=None , **UpperCamelCase_: Optional[Any] , ): super().__init__(pad_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = position_embedding_type __lowerCamelCase = use_cache __lowerCamelCase = classifier_dropout class lowerCamelCase__( __lowerCamelCase): @property def lowerCAmelCase__ ( self: Any ): if self.task == "multiple-choice": __lowerCamelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowerCamelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
80
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCAmelCase_ = { 'configuration_maskformer': ['MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MaskFormerConfig'], 'configuration_maskformer_swin': ['MaskFormerSwinConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['MaskFormerFeatureExtractor'] UpperCAmelCase_ = ['MaskFormerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ 'MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'MaskFormerForInstanceSegmentation', 'MaskFormerModel', 'MaskFormerPreTrainedModel', ] UpperCAmelCase_ = [ 'MaskFormerSwinBackbone', 'MaskFormerSwinModel', 'MaskFormerSwinPreTrainedModel', ] if TYPE_CHECKING: from .configuration_maskformer import MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskFormerConfig from .configuration_maskformer_swin import MaskFormerSwinConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_maskformer import MaskFormerFeatureExtractor from .image_processing_maskformer import MaskFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskformer import ( MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskFormerForInstanceSegmentation, MaskFormerModel, MaskFormerPreTrainedModel, ) from .modeling_maskformer_swin import ( MaskFormerSwinBackbone, MaskFormerSwinModel, MaskFormerSwinPreTrainedModel, ) else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure)
712
from __future__ import annotations from math import ceil, floor, sqrt def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0] __lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target __lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target __lowerCamelCase = 0 # an estimate of b, using the quadratic formula __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the triangle number corresponding to b_floor __lowerCamelCase = 42 # the triangle number corresponding to b_ceil __lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): __lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 __lowerCamelCase = floor(A__ ) __lowerCamelCase = ceil(A__ ) __lowerCamelCase = triangle_numbers[b_floor] __lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_first_guess * triangle_a __lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_second_guess * triangle_a __lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(f"""{solution() = }""")
80
0
'''simple docstring''' import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = np.inf def set_batch_size(A__ : List[Any] ) -> None: nonlocal batch_size if isinstance(__UpperCamelCase , __UpperCamelCase ): __lowerCamelCase = min(__UpperCamelCase , config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__UpperCamelCase , __UpperCamelCase ): __lowerCamelCase = min(__UpperCamelCase , config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__UpperCamelCase , __UpperCamelCase ) and feature.dtype == "binary": __lowerCamelCase = min(__UpperCamelCase , config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__UpperCamelCase , __UpperCamelCase ) return None if batch_size is np.inf else batch_size class lowerCamelCase__( __lowerCamelCase): def __init__( self: str , UpperCamelCase_: Dict , UpperCamelCase_: int = None , UpperCamelCase_: Union[str, Any] = None , UpperCamelCase_: Any = None , UpperCamelCase_: str = False , UpperCamelCase_: Union[str, Any] = False , UpperCamelCase_: Tuple = None , **UpperCamelCase_: Tuple , ): super().__init__( _SCREAMING_SNAKE_CASE , split=_SCREAMING_SNAKE_CASE , features=_SCREAMING_SNAKE_CASE , cache_dir=_SCREAMING_SNAKE_CASE , keep_in_memory=_SCREAMING_SNAKE_CASE , streaming=_SCREAMING_SNAKE_CASE , num_proc=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) __lowerCamelCase = path_or_paths if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else {self.split: path_or_paths} __lowerCamelCase = _PACKAGED_DATASETS_MODULES["""parquet"""][1] __lowerCamelCase = Parquet( cache_dir=_SCREAMING_SNAKE_CASE , data_files=_SCREAMING_SNAKE_CASE , features=_SCREAMING_SNAKE_CASE , hash=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) def lowerCAmelCase__ ( self: Any ): if self.streaming: __lowerCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None self.builder.download_and_prepare( download_config=_SCREAMING_SNAKE_CASE , download_mode=_SCREAMING_SNAKE_CASE , verification_mode=_SCREAMING_SNAKE_CASE , base_path=_SCREAMING_SNAKE_CASE , num_proc=self.num_proc , ) __lowerCamelCase = self.builder.as_dataset( split=self.split , verification_mode=_SCREAMING_SNAKE_CASE , in_memory=self.keep_in_memory ) return dataset class lowerCamelCase__: def __init__( self: Tuple , UpperCamelCase_: int , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] = None , **UpperCamelCase_: str , ): __lowerCamelCase = dataset __lowerCamelCase = path_or_buf __lowerCamelCase = batch_size or get_writer_batch_size(dataset.features ) __lowerCamelCase = parquet_writer_kwargs def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with open(self.path_or_buf , """wb+""" ) as buffer: __lowerCamelCase = self._write(file_obj=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE , **self.parquet_writer_kwargs ) else: __lowerCamelCase = self._write(file_obj=self.path_or_buf , batch_size=_SCREAMING_SNAKE_CASE , **self.parquet_writer_kwargs ) return written def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: str , UpperCamelCase_: Dict , **UpperCamelCase_: int ): __lowerCamelCase = 0 __lowerCamelCase = parquet_writer_kwargs.pop("""path_or_buf""" , _SCREAMING_SNAKE_CASE ) __lowerCamelCase = self.dataset.features.arrow_schema __lowerCamelCase = pq.ParquetWriter(_SCREAMING_SNAKE_CASE , schema=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) for offset in logging.tqdm( range(0 , len(self.dataset ) , _SCREAMING_SNAKE_CASE ) , unit="""ba""" , disable=not logging.is_progress_bar_enabled() , desc="""Creating parquet from Arrow format""" , ): __lowerCamelCase = query_table( table=self.dataset._data , key=slice(_SCREAMING_SNAKE_CASE , offset + batch_size ) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(_SCREAMING_SNAKE_CASE ) written += batch.nbytes writer.close() return written
713
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[str] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet, attn in zip(self.resnets , self.attentions ): __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: str , UpperCamelCase_: Any , UpperCamelCase_: Optional[int] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet in self.resnets: __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: Tuple , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: List[Any]=True ): for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Optional[Any]=True ): for resnet in self.resnets: # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: int ): # there is always at least one resnet __lowerCamelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] __lowerCamelCase = [] for _ in range(self.num_layers ): __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions def __call__( self: int , UpperCamelCase_: Any , UpperCamelCase_: int , UpperCamelCase_: Dict , UpperCamelCase_: Optional[int]=True ): __lowerCamelCase = self.resnets[0](UpperCamelCase_ , UpperCamelCase_ ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) return hidden_states
80
0
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline UpperCAmelCase_ = argparse.ArgumentParser('Stable Diffusion script with intel optimization', add_help=False) parser.add_argument('--dpm', action='store_true', help='Enable DPMSolver or not') parser.add_argument('--steps', default=None, type=int, help='Num inference steps') UpperCAmelCase_ = parser.parse_args() UpperCAmelCase_ = 'cpu' UpperCAmelCase_ = 'a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings' UpperCAmelCase_ = 'path-to-your-trained-model' UpperCAmelCase_ = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: UpperCAmelCase_ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) UpperCAmelCase_ = pipe.to(device) # to channels last UpperCAmelCase_ = pipe.unet.to(memory_format=torch.channels_last) UpperCAmelCase_ = pipe.vae.to(memory_format=torch.channels_last) UpperCAmelCase_ = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: UpperCAmelCase_ = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex UpperCAmelCase_ = torch.randn(2, 4, 64, 64) UpperCAmelCase_ = torch.rand(1) * 999 UpperCAmelCase_ = torch.randn(2, 77, 768) UpperCAmelCase_ = (sample, timestep, encoder_hidden_status) try: UpperCAmelCase_ = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: UpperCAmelCase_ = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) UpperCAmelCase_ = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) UpperCAmelCase_ = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: UpperCAmelCase_ = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute UpperCAmelCase_ = 666 UpperCAmelCase_ = torch.Generator(device).manual_seed(seed) UpperCAmelCase_ = {'generator': generator} if args.steps is not None: UpperCAmelCase_ = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): UpperCAmelCase_ = pipe(prompt, **generate_kwargs).images[0] # save image image.save('generated.png')
714
import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging UpperCAmelCase_ = ['bart.large', 'bart.large.mnli', 'bart.large.cnn', 'bart_xsum/model.pt'] UpperCAmelCase_ = {'bart.large': BartModel, 'bart.large.mnli': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('0.9.0'): raise Exception('requires fairseq >= 0.9.0') logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = ' Hello world! cécé herlolip' UpperCAmelCase_ = [ ('model.classification_heads.mnli.dense.weight', 'classification_head.dense.weight'), ('model.classification_heads.mnli.dense.bias', 'classification_head.dense.bias'), ('model.classification_heads.mnli.out_proj.weight', 'classification_head.out_proj.weight'), ('model.classification_heads.mnli.out_proj.bias', 'classification_head.out_proj.bias'), ] def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """_float_tensor""", ] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Tuple , A__ : Any , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase = dct.pop(A__ ) __lowerCamelCase = val def lowerCamelCase__ ( A__ : Tuple ): '''simple docstring''' __lowerCamelCase = torch.load(A__ , map_location="""cpu""" ) __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , """bart.large.cnn""" ).eval() hub_interface.model.load_state_dict(sd["""model"""] ) return hub_interface def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = emb.weight.shape __lowerCamelCase = nn.Linear(A__ , A__ , bias=A__ ) __lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Optional[int] , A__ : Dict=None ): '''simple docstring''' if not os.path.exists(A__ ): __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , A__ ).eval() else: __lowerCamelCase = load_xsum_checkpoint(A__ ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: __lowerCamelCase = checkpoint_path.replace(""".""" , """-""" ) __lowerCamelCase = BartConfig.from_pretrained(A__ ) __lowerCamelCase = bart.encode(A__ ).unsqueeze(0 ) __lowerCamelCase = BartTokenizer.from_pretrained(A__ ).encode(A__ , return_tensors="""pt""" ).unsqueeze(0 ) if not torch.eq(A__ , A__ ).all(): raise ValueError( f'converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}' ) if checkpoint_path == "bart.large.mnli": __lowerCamelCase = bart.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""model.decoder.embed_tokens.weight"""] for src, dest in mnli_rename_keys: rename_key(A__ , A__ , A__ ) __lowerCamelCase = BartForSequenceClassification(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = bart.predict("""mnli""" , A__ , return_logits=A__ ) __lowerCamelCase = model(A__ )[0] # logits else: # no classification heads to worry about __lowerCamelCase = bart.model.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""decoder.embed_tokens.weight"""] __lowerCamelCase = bart.extract_features(A__ ) if hf_checkpoint_name == "facebook/bart-large": __lowerCamelCase = BartModel(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = model(A__ ).model[0] else: __lowerCamelCase = BartForConditionalGeneration(A__ ).eval() # an existing summarization ckpt model.model.load_state_dict(A__ ) if hasattr(A__ , """lm_head""" ): __lowerCamelCase = make_linear_from_emb(model.model.shared ) __lowerCamelCase = model.model(A__ )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f'`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}' ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError("""Some values in `fairseq_output` are different from `new_model_outputs`""" ) Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default=None, type=str, help='Which huggingface architecture to use: bart-large-xsum' ) UpperCAmelCase_ = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
80
0
from __future__ import annotations from typing import Generic, TypeVar UpperCAmelCase_ = TypeVar('T') class lowerCamelCase__( Generic[T]): def __init__( self: List[Any] , UpperCamelCase_: T ): __lowerCamelCase = data __lowerCamelCase = self __lowerCamelCase = 0 class lowerCamelCase__( Generic[T]): def __init__( self: str ): # map from node name to the node object __lowerCamelCase = {} def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: T ): # create a new set with x as its member __lowerCamelCase = DisjointSetTreeNode(__lowercase ) def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: T ): # find the set x belongs to (with path-compression) __lowerCamelCase = self.map[data] if elem_ref != elem_ref.parent: __lowerCamelCase = self.find_set(elem_ref.parent.data ) return elem_ref.parent def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: DisjointSetTreeNode[T] , UpperCamelCase_: DisjointSetTreeNode[T] ): # helper function for union operation if nodea.rank > nodea.rank: __lowerCamelCase = nodea else: __lowerCamelCase = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: T , UpperCamelCase_: T ): # merge 2 disjoint sets self.link(self.find_set(__lowercase ) , self.find_set(__lowercase ) ) class lowerCamelCase__( Generic[T]): def __init__( self: int ): # connections: map from the node to the neighbouring nodes (with weights) __lowerCamelCase = {} def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: T ): # add a node ONLY if its not present in the graph if node not in self.connections: __lowerCamelCase = {} def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: T , UpperCamelCase_: T , UpperCamelCase_: int ): # add an edge with the given weight self.add_node(__lowercase ) self.add_node(__lowercase ) __lowerCamelCase = weight __lowerCamelCase = weight def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = [] __lowerCamelCase = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda UpperCamelCase_ : x[2] ) # creating the disjoint set __lowerCamelCase = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(__lowercase ) # MST generation __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = edges[index] index += 1 __lowerCamelCase = disjoint_set.find_set(__lowercase ) __lowerCamelCase = disjoint_set.find_set(__lowercase ) if parent_u != parent_v: num_edges += 1 graph.add_edge(__lowercase , __lowercase , __lowercase ) disjoint_set.union(__lowercase , __lowercase ) return graph
715
import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class lowerCamelCase__: def __init__( self: Tuple , UpperCamelCase_: Any , UpperCamelCase_: List[Any]=14 , UpperCamelCase_: int=7 , UpperCamelCase_: Union[str, Any]=True , UpperCamelCase_: Dict=True , UpperCamelCase_: Optional[Any]=True , UpperCamelCase_: Tuple=True , UpperCamelCase_: List[str]=True , UpperCamelCase_: int=99 , UpperCamelCase_: str=32 , UpperCamelCase_: List[Any]=5 , UpperCamelCase_: Optional[int]=4 , UpperCamelCase_: List[Any]=37 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[str]=5_12 , UpperCamelCase_: Dict=16 , UpperCamelCase_: List[str]=2 , UpperCamelCase_: Optional[Any]=0.02 , UpperCamelCase_: List[str]=3 , UpperCamelCase_: Tuple=4 , UpperCamelCase_: Tuple=None , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_input_mask __lowerCamelCase = use_labels __lowerCamelCase = use_mc_token_ids __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = self.vocab_size - 1 def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None if self.use_mc_token_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = self.get_config() __lowerCamelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def lowerCAmelCase__ ( self: Dict ): return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: str , UpperCamelCase_: Dict , UpperCamelCase_: Tuple , UpperCamelCase_: Any , UpperCamelCase_: List[str] , *UpperCamelCase_: Optional[Any] ): __lowerCamelCase = CTRLModel(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , head_mask=UpperCamelCase_ ) model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ ) __lowerCamelCase = model(UpperCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Dict , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: List[Any] , *UpperCamelCase_: Tuple ): __lowerCamelCase = CTRLLMHeadModel(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ) = config_and_inputs __lowerCamelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask} return config, inputs_dict def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , *UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = self.num_labels __lowerCamelCase = CTRLForSequenceClassification(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : Any = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () UpperCAmelCase__ : Optional[Any] = (CTRLLMHeadModel,) if is_torch_available() else () UpperCAmelCase__ : int = ( { 'feature-extraction': CTRLModel, 'text-classification': CTRLForSequenceClassification, 'text-generation': CTRLLMHeadModel, 'zero-shot': CTRLForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ : List[str] = True UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : Optional[Any] = False def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Any , UpperCamelCase_: List[str] , UpperCamelCase_: Tuple , UpperCamelCase_: Tuple , UpperCamelCase_: List[str] ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = CTRLModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=UpperCamelCase_ , n_embd=37 ) def lowerCAmelCase__ ( self: Optional[int] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__ ( self: Optional[Any] ): self.config_tester.run_common_tests() def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*UpperCamelCase_ ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*UpperCamelCase_ ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCAmelCase__ ( self: List[Any] ): pass @slow def lowerCAmelCase__ ( self: Optional[Any] ): for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = CTRLModel.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) @unittest.skip("""The model doesn't support left padding""" ) # and it's not used enough to be worth fixing :) def lowerCAmelCase__ ( self: Optional[Any] ): pass @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: List[str] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = CTRLLMHeadModel.from_pretrained("""ctrl""" ) model.to(UpperCamelCase_ ) __lowerCamelCase = torch.tensor( [[1_18_59, 0, 16_11, 8]] , dtype=torch.long , device=UpperCamelCase_ ) # Legal the president is __lowerCamelCase = [ 1_18_59, 0, 16_11, 8, 5, 1_50, 2_64_49, 2, 19, 3_48, 4_69, 3, 25_95, 48, 2_07_40, 24_65_33, 24_65_33, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a __lowerCamelCase = model.generate(UpperCamelCase_ , do_sample=UpperCamelCase_ ) self.assertListEqual(output_ids[0].tolist() , UpperCamelCase_ )
80
0
import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger('transformers.models.encodec') UpperCAmelCase_ = { """quantizer.vq.layers.*._codebook.inited""": """quantizer.layers.*.codebook.inited""", """quantizer.vq.layers.*._codebook.cluster_size""": """quantizer.layers.*.codebook.cluster_size""", """quantizer.vq.layers.*._codebook.embed""": """quantizer.layers.*.codebook.embed""", """quantizer.vq.layers.*._codebook.embed_avg""": """quantizer.layers.*.codebook.embed_avg""", } UpperCAmelCase_ = { """encoder.model.0.conv.conv""": """encoder.layers.0.conv""", """encoder.model.1.block.1.conv.conv""": """encoder.layers.1.block.1.conv""", """encoder.model.1.block.3.conv.conv""": """encoder.layers.1.block.3.conv""", """encoder.model.1.shortcut.conv.conv""": """encoder.layers.1.shortcut.conv""", """encoder.model.3.conv.conv""": """encoder.layers.3.conv""", """encoder.model.4.block.1.conv.conv""": """encoder.layers.4.block.1.conv""", """encoder.model.4.block.3.conv.conv""": """encoder.layers.4.block.3.conv""", """encoder.model.4.shortcut.conv.conv""": """encoder.layers.4.shortcut.conv""", """encoder.model.6.conv.conv""": """encoder.layers.6.conv""", """encoder.model.7.block.1.conv.conv""": """encoder.layers.7.block.1.conv""", """encoder.model.7.block.3.conv.conv""": """encoder.layers.7.block.3.conv""", """encoder.model.7.shortcut.conv.conv""": """encoder.layers.7.shortcut.conv""", """encoder.model.9.conv.conv""": """encoder.layers.9.conv""", """encoder.model.10.block.1.conv.conv""": """encoder.layers.10.block.1.conv""", """encoder.model.10.block.3.conv.conv""": """encoder.layers.10.block.3.conv""", """encoder.model.10.shortcut.conv.conv""": """encoder.layers.10.shortcut.conv""", """encoder.model.12.conv.conv""": """encoder.layers.12.conv""", """encoder.model.13.lstm""": """encoder.layers.13.lstm""", """encoder.model.15.conv.conv""": """encoder.layers.15.conv""", } UpperCAmelCase_ = { """encoder.model.0.conv.norm""": """encoder.layers.0.norm""", """encoder.model.1.block.1.conv.norm""": """encoder.layers.1.block.1.norm""", """encoder.model.1.block.3.conv.norm""": """encoder.layers.1.block.3.norm""", """encoder.model.1.shortcut.conv.norm""": """encoder.layers.1.shortcut.norm""", """encoder.model.3.conv.norm""": """encoder.layers.3.norm""", """encoder.model.4.block.1.conv.norm""": """encoder.layers.4.block.1.norm""", """encoder.model.4.block.3.conv.norm""": """encoder.layers.4.block.3.norm""", """encoder.model.4.shortcut.conv.norm""": """encoder.layers.4.shortcut.norm""", """encoder.model.6.conv.norm""": """encoder.layers.6.norm""", """encoder.model.7.block.1.conv.norm""": """encoder.layers.7.block.1.norm""", """encoder.model.7.block.3.conv.norm""": """encoder.layers.7.block.3.norm""", """encoder.model.7.shortcut.conv.norm""": """encoder.layers.7.shortcut.norm""", """encoder.model.9.conv.norm""": """encoder.layers.9.norm""", """encoder.model.10.block.1.conv.norm""": """encoder.layers.10.block.1.norm""", """encoder.model.10.block.3.conv.norm""": """encoder.layers.10.block.3.norm""", """encoder.model.10.shortcut.conv.norm""": """encoder.layers.10.shortcut.norm""", """encoder.model.12.conv.norm""": """encoder.layers.12.norm""", """encoder.model.15.conv.norm""": """encoder.layers.15.norm""", } UpperCAmelCase_ = { """decoder.model.0.conv.conv""": """decoder.layers.0.conv""", """decoder.model.1.lstm""": """decoder.layers.1.lstm""", """decoder.model.3.convtr.convtr""": """decoder.layers.3.conv""", """decoder.model.4.block.1.conv.conv""": """decoder.layers.4.block.1.conv""", """decoder.model.4.block.3.conv.conv""": """decoder.layers.4.block.3.conv""", """decoder.model.4.shortcut.conv.conv""": """decoder.layers.4.shortcut.conv""", """decoder.model.6.convtr.convtr""": """decoder.layers.6.conv""", """decoder.model.7.block.1.conv.conv""": """decoder.layers.7.block.1.conv""", """decoder.model.7.block.3.conv.conv""": """decoder.layers.7.block.3.conv""", """decoder.model.7.shortcut.conv.conv""": """decoder.layers.7.shortcut.conv""", """decoder.model.9.convtr.convtr""": """decoder.layers.9.conv""", """decoder.model.10.block.1.conv.conv""": """decoder.layers.10.block.1.conv""", """decoder.model.10.block.3.conv.conv""": """decoder.layers.10.block.3.conv""", """decoder.model.10.shortcut.conv.conv""": """decoder.layers.10.shortcut.conv""", """decoder.model.12.convtr.convtr""": """decoder.layers.12.conv""", """decoder.model.13.block.1.conv.conv""": """decoder.layers.13.block.1.conv""", """decoder.model.13.block.3.conv.conv""": """decoder.layers.13.block.3.conv""", """decoder.model.13.shortcut.conv.conv""": """decoder.layers.13.shortcut.conv""", """decoder.model.15.conv.conv""": """decoder.layers.15.conv""", } UpperCAmelCase_ = { """decoder.model.0.conv.norm""": """decoder.layers.0.norm""", """decoder.model.3.convtr.norm""": """decoder.layers.3.norm""", """decoder.model.4.block.1.conv.norm""": """decoder.layers.4.block.1.norm""", """decoder.model.4.block.3.conv.norm""": """decoder.layers.4.block.3.norm""", """decoder.model.4.shortcut.conv.norm""": """decoder.layers.4.shortcut.norm""", """decoder.model.6.convtr.norm""": """decoder.layers.6.norm""", """decoder.model.7.block.1.conv.norm""": """decoder.layers.7.block.1.norm""", """decoder.model.7.block.3.conv.norm""": """decoder.layers.7.block.3.norm""", """decoder.model.7.shortcut.conv.norm""": """decoder.layers.7.shortcut.norm""", """decoder.model.9.convtr.norm""": """decoder.layers.9.norm""", """decoder.model.10.block.1.conv.norm""": """decoder.layers.10.block.1.norm""", """decoder.model.10.block.3.conv.norm""": """decoder.layers.10.block.3.norm""", """decoder.model.10.shortcut.conv.norm""": """decoder.layers.10.shortcut.norm""", """decoder.model.12.convtr.norm""": """decoder.layers.12.norm""", """decoder.model.13.block.1.conv.norm""": """decoder.layers.13.block.1.norm""", """decoder.model.13.block.3.conv.norm""": """decoder.layers.13.block.3.norm""", """decoder.model.13.shortcut.conv.norm""": """decoder.layers.13.shortcut.norm""", """decoder.model.15.conv.norm""": """decoder.layers.15.norm""", } UpperCAmelCase_ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } UpperCAmelCase_ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } UpperCAmelCase_ = [] UpperCAmelCase_ = [] def lowerCamelCase__ ( A__ : int , A__ : List[Any] , A__ : Dict , A__ : Optional[Any] , A__ : List[Any] ): '''simple docstring''' for attribute in key.split(""".""" ): __lowerCamelCase = getattr(_lowerCAmelCase , _lowerCAmelCase ) if weight_type is not None: __lowerCamelCase = getattr(_lowerCAmelCase , _lowerCAmelCase ).shape else: __lowerCamelCase = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": __lowerCamelCase = value elif weight_type == "weight_g": __lowerCamelCase = value elif weight_type == "weight_v": __lowerCamelCase = value elif weight_type == "bias": __lowerCamelCase = value elif weight_type == "running_mean": __lowerCamelCase = value elif weight_type == "running_var": __lowerCamelCase = value elif weight_type == "num_batches_tracked": __lowerCamelCase = value elif weight_type == "weight_ih_l0": __lowerCamelCase = value elif weight_type == "weight_hh_l0": __lowerCamelCase = value elif weight_type == "bias_ih_l0": __lowerCamelCase = value elif weight_type == "bias_hh_l0": __lowerCamelCase = value elif weight_type == "weight_ih_l1": __lowerCamelCase = value elif weight_type == "weight_hh_l1": __lowerCamelCase = value elif weight_type == "bias_ih_l1": __lowerCamelCase = value elif weight_type == "bias_hh_l1": __lowerCamelCase = value else: __lowerCamelCase = value logger.info(f'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def lowerCamelCase__ ( A__ : Optional[int] , A__ : Any ): '''simple docstring''' for key in ignore_keys: if key.endswith(""".*""" ): if name.startswith(key[:-1] ): return True elif ".*." in key: __lowerCamelCase = key.split(""".*.""" ) if prefix in name and suffix in name: return True elif key in name: return True return False def lowerCamelCase__ ( A__ : Tuple , A__ : int , A__ : List[str] ): '''simple docstring''' __lowerCamelCase = [] if model_name == "encodec_24khz" or "encodec_32khz": __lowerCamelCase = MAPPING_24K elif model_name == "encodec_48khz": __lowerCamelCase = MAPPING_48K else: raise ValueError(f'Unsupported model: {model_name}' ) for name, value in orig_dict.items(): if should_ignore(_lowerCAmelCase , _lowerCAmelCase ): logger.info(f'{name} was ignored' ) continue __lowerCamelCase = False for key, mapped_key in MAPPING.items(): if "*" in key: __lowerCamelCase = key.split(""".*.""" ) if prefix in name and suffix in name: __lowerCamelCase = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith("""embed""" ) and name.endswith("""embed_avg""" ): continue __lowerCamelCase = True if "*" in mapped_key: __lowerCamelCase = name.split(_lowerCAmelCase )[0].split(""".""" )[-2] __lowerCamelCase = mapped_key.replace("""*""" , _lowerCAmelCase ) if "weight_g" in name: __lowerCamelCase = "weight_g" elif "weight_v" in name: __lowerCamelCase = "weight_v" elif "weight_ih_l0" in name: __lowerCamelCase = "weight_ih_l0" elif "weight_hh_l0" in name: __lowerCamelCase = "weight_hh_l0" elif "bias_ih_l0" in name: __lowerCamelCase = "bias_ih_l0" elif "bias_hh_l0" in name: __lowerCamelCase = "bias_hh_l0" elif "weight_ih_l1" in name: __lowerCamelCase = "weight_ih_l1" elif "weight_hh_l1" in name: __lowerCamelCase = "weight_hh_l1" elif "bias_ih_l1" in name: __lowerCamelCase = "bias_ih_l1" elif "bias_hh_l1" in name: __lowerCamelCase = "bias_hh_l1" elif "bias" in name: __lowerCamelCase = "bias" elif "weight" in name: __lowerCamelCase = "weight" elif "running_mean" in name: __lowerCamelCase = "running_mean" elif "running_var" in name: __lowerCamelCase = "running_var" elif "num_batches_tracked" in name: __lowerCamelCase = "num_batches_tracked" else: __lowerCamelCase = None set_recursively(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) continue if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(f'Unused weights: {unused_weights}' ) @torch.no_grad() def lowerCamelCase__ ( A__ : Optional[Any] , A__ : int , A__ : Any , A__ : Union[str, Any]=None , A__ : Optional[Any]=None , ): '''simple docstring''' if config_path is not None: __lowerCamelCase = EncodecConfig.from_pretrained(_lowerCAmelCase ) else: __lowerCamelCase = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": __lowerCamelCase = [8, 5, 4, 4] __lowerCamelCase = [2.2] __lowerCamelCase = 64 __lowerCamelCase = 32000 __lowerCamelCase = 2048 __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False elif model_name == "encodec_48khz": __lowerCamelCase = [8, 5, 4, 2] __lowerCamelCase = [3.0, 6.0, 12.0, 24.0] __lowerCamelCase = 48000 __lowerCamelCase = 2 __lowerCamelCase = False __lowerCamelCase = "time_group_norm" __lowerCamelCase = True __lowerCamelCase = 1.0 __lowerCamelCase = 0.01 else: raise ValueError(f'Unknown model name: {model_name}' ) __lowerCamelCase = EncodecModel(_lowerCAmelCase ) __lowerCamelCase = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(_lowerCAmelCase ) __lowerCamelCase = torch.load(_lowerCAmelCase ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights __lowerCamelCase = original_checkpoint["best_state"] recursively_load_weights(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) model.save_pretrained(_lowerCAmelCase ) if repo_id: print("""Pushing to the hub...""" ) feature_extractor.push_to_hub(_lowerCAmelCase ) model.push_to_hub(_lowerCAmelCase ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument( '--model', default='encodec_24khz', type=str, help='The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) UpperCAmelCase_ = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
716
def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0 for i in range(n + 1 )] __lowerCamelCase = 1 __lowerCamelCase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , A__ ): __lowerCamelCase = 1 __lowerCamelCase = 0 for i in range(A__ ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(f"""{solution() = }""")
80
0
def lowerCamelCase__ ( A__ : list ): '''simple docstring''' if any(not isinstance(A__ , A__ ) or x < 0 for x in sequence ): raise TypeError("""Sequence must be list of non-negative integers""" ) for _ in range(len(A__ ) ): for i, (rod_upper, rod_lower) in enumerate(zip(A__ , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
717
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase): UpperCAmelCase__ : Dict = 1 @register_to_config def __init__( self: List[str] , UpperCamelCase_: int = 10_00 , UpperCamelCase_: Optional[Union[np.ndarray, List[float]]] = None ): # set `betas`, `alphas`, `timesteps` self.set_timesteps(UpperCamelCase_ ) # standard deviation of the initial noise distribution __lowerCamelCase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowerCamelCase = 4 # running values __lowerCamelCase = [] def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: int , UpperCamelCase_: Union[str, torch.device] = None ): __lowerCamelCase = num_inference_steps __lowerCamelCase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowerCamelCase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowerCamelCase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowerCamelCase = torch.sin(steps * math.pi / 2 ) ** 2 __lowerCamelCase = (1.0 - self.betas**2) ** 0.5 __lowerCamelCase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowerCamelCase = timesteps.to(UpperCamelCase_ ) __lowerCamelCase = [] def lowerCAmelCase__ ( self: int , UpperCamelCase_: torch.FloatTensor , UpperCamelCase_: int , UpperCamelCase_: torch.FloatTensor , UpperCamelCase_: bool = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowerCamelCase = (self.timesteps == timestep).nonzero().item() __lowerCamelCase = timestep_index + 1 __lowerCamelCase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(UpperCamelCase_ ) if len(self.ets ) == 1: __lowerCamelCase = self.ets[-1] elif len(self.ets ) == 2: __lowerCamelCase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowerCamelCase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowerCamelCase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowerCamelCase = self._get_prev_sample(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: torch.FloatTensor , *UpperCamelCase_: Dict , **UpperCamelCase_: Union[str, Any] ): return sample def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Any ): __lowerCamelCase = self.alphas[timestep_index] __lowerCamelCase = self.betas[timestep_index] __lowerCamelCase = self.alphas[prev_timestep_index] __lowerCamelCase = self.betas[prev_timestep_index] __lowerCamelCase = (sample - sigma * ets) / max(UpperCamelCase_ , 1E-8 ) __lowerCamelCase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self: List[Any] ): return self.config.num_train_timesteps
80
0
UpperCAmelCase_ = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] UpperCAmelCase_ = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] UpperCAmelCase_ = { 0: 'Sunday', 1: 'Monday', 2: 'Tuesday', 3: 'Wednesday', 4: 'Thursday', 5: 'Friday', 6: 'Saturday', } def lowerCamelCase__ ( A__ : List[Any] , A__ : str , A__ : Optional[int] ): '''simple docstring''' assert len(str(__lowerCAmelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: __lowerCamelCase = year // 100 __lowerCamelCase = (5 * (century % 4) + 2) % 7 __lowerCamelCase = year % 100 __lowerCamelCase = centurian % 12 __lowerCamelCase = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 __lowerCamelCase = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) __lowerCamelCase = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
718
import os from collections.abc import Iterator def lowerCamelCase__ ( A__ : str = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(A__ ): __lowerCamelCase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(A__ )[1] in (".py", ".ipynb"): yield os.path.join(A__ , A__ ).lstrip("""./""" ) def lowerCamelCase__ ( A__ : Optional[int] ): '''simple docstring''' return f'{i * " "}*' if i else "\n##" def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(A__ ) or old_parts[i] != new_part) and new_part: print(f'{md_prefix(A__ )} {new_part.replace("_" , " " ).title()}' ) return new_path def lowerCamelCase__ ( A__ : str = "." ): '''simple docstring''' __lowerCamelCase = """""" for filepath in sorted(good_file_paths(A__ ) ): __lowerCamelCase, __lowerCamelCase = os.path.split(A__ ) if filepath != old_path: __lowerCamelCase = print_path(A__ , A__ ) __lowerCamelCase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowerCamelCase = f'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowerCamelCase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(f'{md_prefix(A__ )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md('.')
80
0
import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch UpperCAmelCase_ = random.Random() def lowerCamelCase__ ( A__ : Any , A__ : Dict=1.0 , A__ : List[Any]=None , A__ : Union[str, Any]=None ): '''simple docstring''' if rng is None: __lowerCamelCase = global_rng __lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class lowerCamelCase__( unittest.TestCase): def __init__( self: str , UpperCamelCase_: List[Any] , UpperCamelCase_: Dict=7 , UpperCamelCase_: List[str]=4_00 , UpperCamelCase_: Optional[int]=20_00 , UpperCamelCase_: Union[str, Any]=1 , UpperCamelCase_: Tuple=0.0 , UpperCamelCase_: Optional[Any]=1_60_00 , UpperCamelCase_: str=True , UpperCamelCase_: Tuple=80 , UpperCamelCase_: Any=16 , UpperCamelCase_: int=64 , UpperCamelCase_: Optional[int]="hann_window" , UpperCamelCase_: Optional[Any]=80 , UpperCamelCase_: List[Any]=76_00 , UpperCamelCase_: List[str]=1E-10 , UpperCamelCase_: Union[str, Any]=True , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = min_seq_length __lowerCamelCase = max_seq_length __lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCamelCase = feature_size __lowerCamelCase = padding_value __lowerCamelCase = sampling_rate __lowerCamelCase = do_normalize __lowerCamelCase = num_mel_bins __lowerCamelCase = hop_length __lowerCamelCase = win_length __lowerCamelCase = win_function __lowerCamelCase = fmin __lowerCamelCase = fmax __lowerCamelCase = mel_floor __lowerCamelCase = return_attention_mask def lowerCAmelCase__ ( self: int ): return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Any=False , UpperCamelCase_: List[Any]=False ): def _flatten(UpperCamelCase_: Tuple ): return list(itertools.chain(*lowerCAmelCase_ ) ) if equal_length: __lowerCamelCase = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __lowerCamelCase = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCamelCase = [np.asarray(lowerCAmelCase_ ) for x in speech_inputs] return speech_inputs def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: str=False , UpperCamelCase_: int=False ): if equal_length: __lowerCamelCase = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowerCamelCase = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCamelCase = [np.asarray(lowerCAmelCase_ ) for x in speech_inputs] return speech_inputs @require_torch class lowerCamelCase__( __a , unittest.TestCase): UpperCAmelCase__ : int = SpeechTaFeatureExtractor def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = SpeechTaFeatureExtractionTester(self ) def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: List[Any] ): self.assertTrue(np.all(np.mean(lowerCAmelCase_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase_ , axis=0 ) - 1 ) < 1E-3 ) ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCamelCase = [np.asarray(lowerCAmelCase_ ) for speech_input in speech_inputs] # Test not batched input __lowerCamelCase = feat_extract(speech_inputs[0] , return_tensors="""np""" ).input_values __lowerCamelCase = feat_extract(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) # Test batched __lowerCamelCase = feat_extract(lowerCAmelCase_ , return_tensors="""np""" ).input_values __lowerCamelCase = feat_extract(lowerCAmelCase_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase_ , lowerCAmelCase_ ): self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCamelCase = ["""longest""", """max_length""", """do_not_pad"""] __lowerCamelCase = [None, 16_00, None] for max_length, padding in zip(lowerCAmelCase_ , lowerCAmelCase_ ): __lowerCamelCase = feat_extract(lowerCAmelCase_ , padding=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors="""np""" ) __lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00] ) self.assertTrue(input_values[0][8_00:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:10_00] ) self.assertTrue(input_values[0][10_00:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:12_00] ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCamelCase = range(8_00 , 14_00 , 2_00 ) __lowerCamelCase = [floats_list((1, x) )[0] for x in lengths] __lowerCamelCase = ["""longest""", """max_length""", """do_not_pad"""] __lowerCamelCase = [None, 16_00, None] for max_length, padding in zip(lowerCAmelCase_ , lowerCAmelCase_ ): __lowerCamelCase = feat_extract(lowerCAmelCase_ , max_length=lowerCAmelCase_ , padding=lowerCAmelCase_ ) __lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00] ) self._check_zero_mean_unit_variance(input_values[1][:10_00] ) self._check_zero_mean_unit_variance(input_values[2][:12_00] ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCamelCase = feat_extract( lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=10_00 , padding="""max_length""" , return_tensors="""np""" ) __lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCamelCase = feat_extract( lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=10_00 , padding="""longest""" , return_tensors="""np""" ) __lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1, :10_00] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00) ) __lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCamelCase = feat_extract( lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=20_00 , padding="""longest""" , return_tensors="""np""" ) __lowerCamelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00] ) self._check_zero_mean_unit_variance(input_values[1, :10_00] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00) ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCamelCase = np.random.rand(1_00 ).astype(np.floataa ) __lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCamelCase = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __lowerCamelCase = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCamelCase = [np.asarray(lowerCAmelCase_ ) for speech_input in speech_inputs] # Test feature size __lowerCamelCase = feature_extractor(audio_target=lowerCAmelCase_ , padding=lowerCAmelCase_ , return_tensors="""np""" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input __lowerCamelCase = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_values __lowerCamelCase = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) # Test batched __lowerCamelCase = feature_extractor(lowerCAmelCase_ , return_tensors="""np""" ).input_values __lowerCamelCase = feature_extractor(lowerCAmelCase_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase_ , lowerCAmelCase_ ): self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __lowerCamelCase = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] __lowerCamelCase = np.asarray(lowerCAmelCase_ ) __lowerCamelCase = feature_extractor(lowerCAmelCase_ , return_tensors="""np""" ).input_values __lowerCamelCase = feature_extractor(lowerCAmelCase_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase_ , lowerCAmelCase_ ): self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1E-3 ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = self.feat_extract_tester.prepare_inputs_for_target() __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCamelCase = feat_extract.model_input_names[0] __lowerCamelCase = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase_ ) == len(lowerCAmelCase_ ) for x, y in zip(lowerCAmelCase_ , processed_features[input_name] ) ) ) __lowerCamelCase = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase_ ) __lowerCamelCase = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) __lowerCamelCase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowerCamelCase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase_ ) __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCamelCase = feat_extract.model_input_names[0] __lowerCamelCase = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) __lowerCamelCase = processed_features[input_name] if len(batch_features_input.shape ) < 3: __lowerCamelCase = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def lowerCAmelCase__ ( self: int ): __lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) __lowerCamelCase = self.feat_extract_tester.prepare_inputs_for_target() __lowerCamelCase = feat_extract.model_input_names[0] __lowerCamelCase = BatchFeature({input_name: speech_inputs} ) __lowerCamelCase = feat_extract.num_mel_bins # hack! __lowerCamelCase = feat_extract.pad(lowerCAmelCase_ , padding="""longest""" , return_tensors="""np""" )[input_name] __lowerCamelCase = feat_extract.pad(lowerCAmelCase_ , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = self.feat_extract_dict __lowerCamelCase = True __lowerCamelCase = self.feature_extraction_class(**lowerCAmelCase_ ) __lowerCamelCase = self.feat_extract_tester.prepare_inputs_for_target() __lowerCamelCase = [len(lowerCAmelCase_ ) for x in speech_inputs] __lowerCamelCase = feat_extract.model_input_names[0] __lowerCamelCase = BatchFeature({input_name: speech_inputs} ) __lowerCamelCase = feat_extract.num_mel_bins # hack! __lowerCamelCase = feat_extract.pad(lowerCAmelCase_ , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , lowerCAmelCase_ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = self.feat_extract_dict __lowerCamelCase = True __lowerCamelCase = self.feature_extraction_class(**lowerCAmelCase_ ) __lowerCamelCase = self.feat_extract_tester.prepare_inputs_for_target() __lowerCamelCase = [len(lowerCAmelCase_ ) for x in speech_inputs] __lowerCamelCase = feat_extract.model_input_names[0] __lowerCamelCase = BatchFeature({input_name: speech_inputs} ) __lowerCamelCase = min(lowerCAmelCase_ ) __lowerCamelCase = feat_extract.num_mel_bins # hack! __lowerCamelCase = feat_extract.pad( lowerCAmelCase_ , padding="""max_length""" , max_length=lowerCAmelCase_ , truncation=lowerCAmelCase_ , return_tensors="""np""" ) self.assertIn("""attention_mask""" , lowerCAmelCase_ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: int ): from datasets import load_dataset __lowerCamelCase = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __lowerCamelCase = ds.sort("""id""" ).select(range(lowerCAmelCase_ ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = torch.tensor( [2.3804E-03, 2.0752E-03, 1.9836E-03, 2.1057E-03, 1.6174E-03, 3.0518E-04, 9.1553E-05, 3.3569E-04, 9.7656E-04, 1.8311E-03, 2.0142E-03, 2.1057E-03, 1.7395E-03, 4.5776E-04, -3.9673E-04, 4.5776E-04, 1.0071E-03, 9.1553E-05, 4.8828E-04, 1.1597E-03, 7.3242E-04, 9.4604E-04, 1.8005E-03, 1.8311E-03, 8.8501E-04, 4.2725E-04, 4.8828E-04, 7.3242E-04, 1.0986E-03, 2.1057E-03] ) # fmt: on __lowerCamelCase = self._load_datasamples(1 ) __lowerCamelCase = SpeechTaFeatureExtractor() __lowerCamelCase = feature_extractor(lowerCAmelCase_ , return_tensors="""pt""" ).input_values self.assertEquals(input_values.shape , (1, 9_36_80) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase_ , atol=1E-6 ) ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = torch.tensor( [-2.6870, -3.0104, -3.1356, -3.5352, -3.0044, -3.0353, -3.4719, -3.6777, -3.1520, -2.9435, -2.6553, -2.8795, -2.9944, -2.5921, -3.0279, -3.0386, -3.0864, -3.1291, -3.2353, -2.7444, -2.6831, -2.7287, -3.1761, -3.1571, -3.2726, -3.0582, -3.1007, -3.4533, -3.4695, -3.0998] ) # fmt: on __lowerCamelCase = self._load_datasamples(1 ) __lowerCamelCase = SpeechTaFeatureExtractor() __lowerCamelCase = feature_extractor(audio_target=lowerCAmelCase_ , return_tensors="""pt""" ).input_values self.assertEquals(input_values.shape , (1, 3_66, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase_ , atol=1E-4 ) )
719
from __future__ import annotations def lowerCamelCase__ ( A__ : list ): '''simple docstring''' if not nums: raise ValueError("""List is empty""" ) return sum(A__ ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
80
0
import string from math import logaa def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = document.translate( str.maketrans("""""" , """""" , string.punctuation ) ).replace("""\n""" , """""" ) __lowerCamelCase = document_without_punctuation.split(""" """ ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = corpus.lower().translate( str.maketrans("""""" , """""" , string.punctuation ) ) # strip all punctuation and replace it with '' __lowerCamelCase = corpus_without_punctuation.split("""\n""" ) __lowerCamelCase = term.lower() return (len([doc for doc in docs if term in doc] ), len(UpperCamelCase__ )) def lowerCamelCase__ ( A__ : int , A__ : int , A__ : Any=False ): '''simple docstring''' if smoothing: if n == 0: raise ValueError("""log10(0) is undefined.""" ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError("""df must be > 0""" ) elif n == 0: raise ValueError("""log10(0) is undefined.""" ) return round(logaa(n / df ) , 3 ) def lowerCamelCase__ ( A__ : int , A__ : int ): '''simple docstring''' return round(tf * idf , 3 )
720
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase): UpperCAmelCase__ : Any = 'maskformer-swin' UpperCAmelCase__ : List[Any] = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Any , UpperCamelCase_: Any=2_24 , UpperCamelCase_: List[str]=4 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: Optional[int]=96 , UpperCamelCase_: List[str]=[2, 2, 6, 2] , UpperCamelCase_: Optional[Any]=[3, 6, 12, 24] , UpperCamelCase_: str=7 , UpperCamelCase_: int=4.0 , UpperCamelCase_: Optional[int]=True , UpperCamelCase_: Union[str, Any]=0.0 , UpperCamelCase_: Optional[int]=0.0 , UpperCamelCase_: Any=0.1 , UpperCamelCase_: Union[str, Any]="gelu" , UpperCamelCase_: int=False , UpperCamelCase_: Optional[int]=0.02 , UpperCamelCase_: Optional[Any]=1E-5 , UpperCamelCase_: Optional[int]=None , UpperCamelCase_: List[Any]=None , **UpperCamelCase_: Union[str, Any] , ): super().__init__(**UpperCamelCase_ ) __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = embed_dim __lowerCamelCase = depths __lowerCamelCase = len(UpperCamelCase_ ) __lowerCamelCase = num_heads __lowerCamelCase = window_size __lowerCamelCase = mlp_ratio __lowerCamelCase = qkv_bias __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_absolute_embeddings __lowerCamelCase = layer_norm_eps __lowerCamelCase = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCamelCase = int(embed_dim * 2 ** (len(UpperCamelCase_ ) - 1) ) __lowerCamelCase = ["""stem"""] + [F'stage{idx}' for idx in range(1 , len(UpperCamelCase_ ) + 1 )] __lowerCamelCase, __lowerCamelCase = get_aligned_output_features_output_indices( out_features=UpperCamelCase_ , out_indices=UpperCamelCase_ , stage_names=self.stage_names )
80
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { "google/mobilenet_v1_1.0_224": "https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json", "google/mobilenet_v1_0.75_192": "https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json", # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : str = 'mobilenet_v1' def __init__( self: str , UpperCamelCase_: Tuple=3 , UpperCamelCase_: List[Any]=2_24 , UpperCamelCase_: Any=1.0 , UpperCamelCase_: List[Any]=8 , UpperCamelCase_: Optional[int]="relu6" , UpperCamelCase_: List[Any]=True , UpperCamelCase_: Tuple=0.999 , UpperCamelCase_: Optional[Any]=0.02 , UpperCamelCase_: List[Any]=0.001 , **UpperCamelCase_: Optional[Any] , ): super().__init__(**UpperCAmelCase__ ) if depth_multiplier <= 0: raise ValueError("""depth_multiplier must be greater than zero.""" ) __lowerCamelCase = num_channels __lowerCamelCase = image_size __lowerCamelCase = depth_multiplier __lowerCamelCase = min_depth __lowerCamelCase = hidden_act __lowerCamelCase = tf_padding __lowerCamelCase = classifier_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : List[Any] = version.parse('1.11') @property def lowerCAmelCase__ ( self: int ): return OrderedDict([("""pixel_values""", {0: """batch"""})] ) @property def lowerCAmelCase__ ( self: Optional[Any] ): if self.task == "image-classification": return OrderedDict([("""logits""", {0: """batch"""})] ) else: return OrderedDict([("""last_hidden_state""", {0: """batch"""}), ("""pooler_output""", {0: """batch"""})] ) @property def lowerCAmelCase__ ( self: int ): return 1E-4
721
from __future__ import annotations def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): __lowerCamelCase, __lowerCamelCase = array[indexa], array[indexa] def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if length > 1: __lowerCamelCase = int(length / 2 ) for i in range(A__ , low + middle ): comp_and_swap(A__ , A__ , i + middle , A__ ) bitonic_merge(A__ , A__ , A__ , A__ ) bitonic_merge(A__ , low + middle , A__ , A__ ) def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if length > 1: __lowerCamelCase = int(length / 2 ) bitonic_sort(A__ , A__ , A__ , 1 ) bitonic_sort(A__ , low + middle , A__ , 0 ) bitonic_merge(A__ , A__ , A__ , A__ ) if __name__ == "__main__": UpperCAmelCase_ = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase_ = [int(item.strip()) for item in user_input.split(',')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('\nSorted array in ascending order is: ', end='') print(*unsorted, sep=', ') bitonic_merge(unsorted, 0, len(unsorted), 0) print('Sorted array in descending order is: ', end='') print(*unsorted, sep=', ')
80
0
from __future__ import annotations from scipy.special import comb # type: ignore class lowerCamelCase__: def __init__( self: List[str] , UpperCamelCase_: Optional[int] ): __lowerCamelCase = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. __lowerCamelCase = len(A__ ) - 1 def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Any ): assert 0 <= t <= 1, "Time t must be between 0 and 1." __lowerCamelCase = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , A__ ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(A__ ) , 5 ) == 1 return output_values def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: int ): assert 0 <= t <= 1, "Time t must be between 0 and 1." __lowerCamelCase = self.basis_function(A__ ) __lowerCamelCase = 0.0 __lowerCamelCase = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: Any = 0.01 ): from matplotlib import pyplot as plt # type: ignore __lowerCamelCase = [] # x coordinates of points to plot __lowerCamelCase = [] # y coordinates of points to plot __lowerCamelCase = 0.0 while t <= 1: __lowerCamelCase = self.bezier_curve_function(A__ ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size __lowerCamelCase = [i[0] for i in self.list_of_points] __lowerCamelCase = [i[1] for i in self.list_of_points] plt.plot( A__ , A__ , color="""blue""" , label="""Curve of Degree """ + str(self.degree ) , ) plt.scatter(A__ , A__ , color="""red""" , label="""Control Points""" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
700
from ... import PretrainedConfig UpperCAmelCase_ = { 'sijunhe/nezha-cn-base': 'https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json', } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Dict = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP UpperCAmelCase__ : Dict = 'nezha' def __init__( self: Dict , UpperCamelCase_: Any=2_11_28 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Optional[int]=12 , UpperCamelCase_: List[str]=12 , UpperCamelCase_: Optional[int]=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: str=0.1 , UpperCamelCase_: Union[str, Any]=5_12 , UpperCamelCase_: Any=64 , UpperCamelCase_: Dict=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: Optional[Any]=1E-12 , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Any=0 , UpperCamelCase_: str=2 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: str=True , **UpperCamelCase_: Any , ): super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = max_relative_position __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = classifier_dropout __lowerCamelCase = use_cache
80
0
'''simple docstring''' from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : str = 'new-model' if is_tf_available(): class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Optional[Any] = NewModelConfig @require_tf class lowerCamelCase__( unittest.TestCase): @slow def lowerCAmelCase__ ( self: int ): __lowerCamelCase = """bert-base-cased""" __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModel.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCAmelCase__ ( self: str ): __lowerCamelCase = """bert-base-cased""" __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModelForPreTraining.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCAmelCase__ ( self: int ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModelForCausalLM.from_pretrained(UpperCamelCase_ ) __lowerCamelCase, __lowerCamelCase = TFAutoModelForCausalLM.from_pretrained(UpperCamelCase_ , output_loading_info=UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCAmelCase__ ( self: Dict ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModelWithLMHead.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCAmelCase__ ( self: Optional[Any] ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModelForMaskedLM.from_pretrained(UpperCamelCase_ ) __lowerCamelCase, __lowerCamelCase = TFAutoModelForMaskedLM.from_pretrained(UpperCamelCase_ , output_loading_info=UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCAmelCase__ ( self: List[Any] ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained(UpperCamelCase_ ) __lowerCamelCase, __lowerCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained(UpperCamelCase_ , output_loading_info=UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCAmelCase__ ( self: Dict ): for model_name in ["bert-base-uncased"]: __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModelForSequenceClassification.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCAmelCase__ ( self: int ): for model_name in ["bert-base-uncased"]: __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModelForQuestionAnswering.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow @require_tensorflow_probability def lowerCAmelCase__ ( self: Dict ): for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: __lowerCamelCase = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = TFAutoModelForTableQuestionAnswering.from_pretrained(UpperCamelCase_ ) __lowerCamelCase, __lowerCamelCase = TFAutoModelForTableQuestionAnswering.from_pretrained( UpperCamelCase_ , output_loading_info=UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = TFAutoModelWithLMHead.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=UpperCamelCase_ ) , 1_44_10 ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = TFAutoModelWithLMHead.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=UpperCamelCase_ ) , 1_44_10 ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = TFAutoModel.from_pretrained("""sgugger/funnel-random-tiny""" ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = copy.deepcopy(model.config ) __lowerCamelCase = ["""FunnelBaseModel"""] __lowerCamelCase = TFAutoModel.from_config(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(UpperCamelCase_ ) __lowerCamelCase = TFAutoModel.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase__ ( self: Dict ): try: AutoConfig.register("""new-model""" , UpperCamelCase_ ) __lowerCamelCase = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(UpperCamelCase_ ): auto_class.register(UpperCamelCase_ , UpperCamelCase_ ) auto_class.register(UpperCamelCase_ , UpperCamelCase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(UpperCamelCase_ ): auto_class.register(UpperCamelCase_ , UpperCamelCase_ ) # Now that the config is registered, it can be used as any other config with the auto-API __lowerCamelCase = BertModelTester(self ).get_config() __lowerCamelCase = NewModelConfig(**tiny_config.to_dict() ) __lowerCamelCase = auto_class.from_config(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(UpperCamelCase_ ) __lowerCamelCase = auto_class.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def lowerCAmelCase__ ( self: Tuple ): with self.assertRaisesRegex( UpperCamelCase_ , """bert-base is not a local folder and is not a valid model identifier""" ): __lowerCamelCase = TFAutoModel.from_pretrained("""bert-base""" ) def lowerCAmelCase__ ( self: Optional[int] ): with self.assertRaisesRegex( UpperCamelCase_ , r"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): __lowerCamelCase = TFAutoModel.from_pretrained(UpperCamelCase_ , revision="""aaaaaa""" ) def lowerCAmelCase__ ( self: int ): with self.assertRaisesRegex( UpperCamelCase_ , """hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin""" , ): __lowerCamelCase = TFAutoModel.from_pretrained("""hf-internal-testing/config-no-model""" ) def lowerCAmelCase__ ( self: List[Any] ): with self.assertRaisesRegex(UpperCamelCase_ , """Use `from_pt=True` to load this model""" ): __lowerCamelCase = TFAutoModel.from_pretrained("""hf-internal-testing/tiny-bert-pt-only""" ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = TFAutoModel.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) with RequestCounter() as counter: __lowerCamelCase = TFAutoModel.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint __lowerCamelCase = TFAutoModel.from_pretrained("""ArthurZ/tiny-random-bert-sharded""" ) with RequestCounter() as counter: __lowerCamelCase = TFAutoModel.from_pretrained("""ArthurZ/tiny-random-bert-sharded""" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
701
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__: def __init__( self: Union[str, Any] , UpperCamelCase_: str = None , UpperCamelCase_: uuid.UUID = None , UpperCamelCase_: Dict=None , UpperCamelCase_: Any=None ): if not conversation_id: __lowerCamelCase = uuid.uuida() if past_user_inputs is None: __lowerCamelCase = [] if generated_responses is None: __lowerCamelCase = [] __lowerCamelCase = conversation_id __lowerCamelCase = past_user_inputs __lowerCamelCase = generated_responses __lowerCamelCase = text def __eq__( self: Optional[Any] , UpperCamelCase_: Union[str, Any] ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: str , UpperCamelCase_: bool = False ): if self.new_user_input: if overwrite: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' F'with: "{text}".' ) __lowerCamelCase = text else: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' F'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input' ) else: __lowerCamelCase = text def lowerCAmelCase__ ( self: List[str] ): if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __lowerCamelCase = None def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): self.generated_responses.append(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple ): for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self: Union[str, Any] ): __lowerCamelCase = F'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): __lowerCamelCase = """user""" if is_user else """bot""" output += F'{name} >> {text} \n' return output @add_end_docstrings( __lowerCamelCase , r'\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n ' , ) class lowerCamelCase__( __lowerCamelCase): def __init__( self: List[str] , *UpperCamelCase_: List[Any] , **UpperCamelCase_: str ): super().__init__(*UpperCamelCase_ , **UpperCamelCase_ ) if self.tokenizer.pad_token_id is None: __lowerCamelCase = self.tokenizer.eos_token def lowerCAmelCase__ ( self: str , UpperCamelCase_: int=None , UpperCamelCase_: Any=None , UpperCamelCase_: Union[str, Any]=None , **UpperCamelCase_: int ): __lowerCamelCase = {} __lowerCamelCase = {} __lowerCamelCase = {} if min_length_for_response is not None: __lowerCamelCase = min_length_for_response if minimum_tokens is not None: __lowerCamelCase = minimum_tokens if "max_length" in generate_kwargs: __lowerCamelCase = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __lowerCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCamelCase_ ) return preprocess_params, forward_params, postprocess_params def __call__( self: Any , UpperCamelCase_: Union[Conversation, List[Conversation]] , UpperCamelCase_: Optional[int]=0 , **UpperCamelCase_: Optional[int] ): __lowerCamelCase = super().__call__(UpperCamelCase_ , num_workers=UpperCamelCase_ , **UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and len(UpperCamelCase_ ) == 1: return outputs[0] return outputs def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Conversation , UpperCamelCase_: Optional[Any]=32 ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( F'Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ' """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): __lowerCamelCase = self.tokenizer._build_conversation_input_ids(UpperCamelCase_ ) else: # If the tokenizer cannot handle conversations, we default to only the old version __lowerCamelCase = self._legacy_parse_and_tokenize(UpperCamelCase_ ) if self.framework == "pt": __lowerCamelCase = torch.LongTensor([input_ids] ) elif self.framework == "tf": __lowerCamelCase = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str=10 , **UpperCamelCase_: List[str] ): __lowerCamelCase = generate_kwargs.get("""max_length""" , self.model.config.max_length ) __lowerCamelCase = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(F'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})' ) __lowerCamelCase = max_length - minimum_tokens __lowerCamelCase = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: __lowerCamelCase = model_inputs["""attention_mask"""][:, -trim:] __lowerCamelCase = model_inputs.pop("""conversation""" ) __lowerCamelCase = max_length __lowerCamelCase = self.model.generate(**UpperCamelCase_ , **UpperCamelCase_ ) if self.model.config.is_encoder_decoder: __lowerCamelCase = 1 else: __lowerCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Optional[Any] , UpperCamelCase_: int=True ): __lowerCamelCase = model_outputs["""output_ids"""] __lowerCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_ , ) __lowerCamelCase = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCamelCase_ ) return conversation def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Conversation ): __lowerCamelCase = self.tokenizer.eos_token_id __lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) ) if len(UpperCamelCase_ ) > self.tokenizer.model_max_length: __lowerCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
80
0
from math import sqrt def lowerCamelCase__ ( A__ : int = 1000000 ): '''simple docstring''' __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowercase__ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f"""{solution() = }""")
702
import math def lowerCamelCase__ ( A__ : int ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = 2 __lowerCamelCase = int(math.sqrt(A__ ) ) # Size of every segment __lowerCamelCase = [True] * (end + 1) __lowerCamelCase = [] while start <= end: if temp[start] is True: in_prime.append(A__ ) for i in range(start * start , end + 1 , A__ ): __lowerCamelCase = False start += 1 prime += in_prime __lowerCamelCase = end + 1 __lowerCamelCase = min(2 * end , A__ ) while low <= n: __lowerCamelCase = [True] * (high - low + 1) for each in in_prime: __lowerCamelCase = math.floor(low / each ) * each if t < low: t += each for j in range(A__ , high + 1 , A__ ): __lowerCamelCase = False for j in range(len(A__ ) ): if temp[j] is True: prime.append(j + low ) __lowerCamelCase = high + 1 __lowerCamelCase = min(high + end , A__ ) return prime print(sieve(10**6))
80
0
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) __lowerCamelCase = Vector() def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(lowercase__ ) , """(0,0,0,0,0,1)""" ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = Vector([1, 2, 3, 4] ) self.assertEqual(len(lowercase__ ) , 4 ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = Vector([1, 2] ) __lowerCamelCase = Vector([1, 2, 3, 4, 5] ) __lowerCamelCase = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) __lowerCamelCase = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = Vector([1, 2, 3] ) __lowerCamelCase = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = Vector([1, 2, 3] ) __lowerCamelCase = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = Vector([1, 2, 3] ) __lowerCamelCase = Vector([2, -1, 4] ) # for test of dot product __lowerCamelCase = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , """(3.0,6.0,9.0)""" ) self.assertEqual((a * b) , 0 ) def lowerCAmelCase__ ( self: Optional[Any] ): self.assertEqual(str(zero_vector(10 ) ).count("""0""" ) , 10 ) def lowerCAmelCase__ ( self: int ): self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , """(0,1,0)""" ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = Vector([1, 2, 3] ) __lowerCamelCase = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , lowercase__ , lowercase__ ) ) , """(3,4,7)""" ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = Vector([1, 0, 0, 0, 0, 0] ) __lowerCamelCase = x.copy() self.assertEqual(str(lowercase__ ) , str(lowercase__ ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(lowercase__ ) , """(0,1,0)""" ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("""|1,2,3|\n|2,4,5|\n|6,7,8|\n""" , str(lowercase__ ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCamelCase = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(lowercase__ , lowercase__ ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCamelCase = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(lowercase__ , lowercase__ ) ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) __lowerCamelCase = Vector([1, 2, 3] ) self.assertEqual("""(14,32,50)""" , str(a * x ) ) self.assertEqual("""|2,4,6|\n|8,10,12|\n|14,16,18|\n""" , str(a * 2 ) ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("""|1,2,5|\n|2,4,5|\n|6,7,8|\n""" , str(lowercase__ ) ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCamelCase = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("""|2,4,10|\n|4,8,10|\n|12,14,18|\n""" , str(a + b ) ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __lowerCamelCase = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("""|0,0,-4|\n|0,0,0|\n|0,0,-2|\n""" , str(a - b ) ) def lowerCAmelCase__ ( self: Tuple ): self.assertEqual( """|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n""" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
703
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase_ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') class lowerCamelCase__( __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : int = BartphoTokenizer UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : List[str] = True def lowerCAmelCase__ ( self: Tuple ): super().setUp() __lowerCamelCase = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] __lowerCamelCase = dict(zip(UpperCamelCase_ , range(len(UpperCamelCase_ ) ) ) ) __lowerCamelCase = {"""unk_token""": """<unk>"""} __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F'{token} {vocab_tokens[token]}\n' ) __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase__ ( self: List[str] , **UpperCamelCase_: List[str] ): kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): __lowerCamelCase = """This is a là test""" __lowerCamelCase = """This is a<unk><unk> test""" return input_text, output_text def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) __lowerCamelCase = """This is a là test""" __lowerCamelCase = """▁This ▁is ▁a ▁l à ▁t est""".split() __lowerCamelCase = tokenizer.tokenize(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase_ ) , UpperCamelCase_ )
80
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) def lowerCamelCase__ ( A__ : Any , A__ : Dict=False , A__ : int=False ): '''simple docstring''' __lowerCamelCase = """backbone.""" if is_semantic else """""" __lowerCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'{prefix}blocks.{i}.norm1.weight', f'beit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'{prefix}blocks.{i}.norm1.bias', f'beit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append( (f'{prefix}blocks.{i}.attn.proj.weight', f'beit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append( (f'{prefix}blocks.{i}.attn.proj.bias', f'beit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'{prefix}blocks.{i}.norm2.weight', f'beit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'{prefix}blocks.{i}.norm2.bias', f'beit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'{prefix}blocks.{i}.mlp.fc1.weight', f'beit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'{prefix}blocks.{i}.mlp.fc1.bias', f'beit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'{prefix}blocks.{i}.mlp.fc2.weight', f'beit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'{prefix}blocks.{i}.mlp.fc2.bias', f'beit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ (f'{prefix}cls_token', """beit.embeddings.cls_token"""), (f'{prefix}patch_embed.proj.weight', """beit.embeddings.patch_embeddings.projection.weight"""), (f'{prefix}patch_embed.proj.bias', """beit.embeddings.patch_embeddings.projection.bias"""), (f'{prefix}pos_embed', """beit.embeddings.position_embeddings"""), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("""mask_token""", """beit.embeddings.mask_token"""), ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ] ) else: # layernorm + classification head rename_keys.extend( [ ("""fc_norm.weight""", """beit.pooler.layernorm.weight"""), ("""fc_norm.bias""", """beit.pooler.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def lowerCamelCase__ ( A__ : Any , A__ : str , A__ : Optional[Any]=False , A__ : int=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): __lowerCamelCase = """backbone.""" if is_semantic else """""" # queries, keys and values __lowerCamelCase = state_dict.pop(f'{prefix}blocks.{i}.attn.qkv.weight' ) __lowerCamelCase = state_dict.pop(f'{prefix}blocks.{i}.attn.q_bias' ) __lowerCamelCase = state_dict.pop(f'{prefix}blocks.{i}.attn.v_bias' ) __lowerCamelCase = in_proj_weight[ : config.hidden_size, : ] __lowerCamelCase = q_bias __lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] __lowerCamelCase = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained __lowerCamelCase = state_dict.pop(f'{prefix}blocks.{i}.gamma_1' ) __lowerCamelCase = state_dict.pop(f'{prefix}blocks.{i}.gamma_2' ) __lowerCamelCase = gamma_a __lowerCamelCase = gamma_a def lowerCamelCase__ ( A__ : Tuple , A__ : List[str] , A__ : int ): '''simple docstring''' __lowerCamelCase = dct.pop(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = val def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" __lowerCamelCase = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( A__ : Any , A__ : Optional[int] , A__ : Union[str, Any]=False ): '''simple docstring''' __lowerCamelCase = False if """rvlcdip""" in checkpoint_url else True __lowerCamelCase = BeitConfig(use_absolute_position_embeddings=SCREAMING_SNAKE_CASE__ , use_mask_token=SCREAMING_SNAKE_CASE__ ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: __lowerCamelCase = 1024 __lowerCamelCase = 4096 __lowerCamelCase = 24 __lowerCamelCase = 16 # labels if "rvlcdip" in checkpoint_url: __lowerCamelCase = 16 __lowerCamelCase = """huggingface/label-files""" __lowerCamelCase = """rvlcdip-id2label.json""" __lowerCamelCase = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type="""dataset""" ) , """r""" ) ) __lowerCamelCase = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} __lowerCamelCase = idalabel __lowerCamelCase = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys __lowerCamelCase = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ , map_location="""cpu""" )["""model"""] __lowerCamelCase = create_rename_keys(SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) read_in_q_k_v(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) # load HuggingFace model __lowerCamelCase = BeitForMaskedImageModeling(SCREAMING_SNAKE_CASE__ ) if has_lm_head else BeitForImageClassification(SCREAMING_SNAKE_CASE__ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # Check outputs on an image __lowerCamelCase = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="""pt""" ) __lowerCamelCase = encoding["""pixel_values"""] __lowerCamelCase = model(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = outputs.logits # verify logits __lowerCamelCase = [1, 16] if """rvlcdip""" in checkpoint_url else [1, 196, 8192] assert logits.shape == torch.Size(SCREAMING_SNAKE_CASE__ ), "Shape of logits not as expected" Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if push_to_hub: if has_lm_head: __lowerCamelCase = """dit-base""" if """base""" in checkpoint_url else """dit-large""" else: __lowerCamelCase = """dit-base-finetuned-rvlcdip""" if """dit-b""" in checkpoint_url else """dit-large-finetuned-rvlcdip""" image_processor.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) model.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument( '--checkpoint_url', default='https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth', type=str, help='URL to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument( '--push_to_hub', action='store_true', ) UpperCAmelCase_ = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
704
def lowerCamelCase__ ( A__ : dict ): '''simple docstring''' __lowerCamelCase = set() # To detect a back edge, keep track of vertices currently in the recursion stack __lowerCamelCase = set() return any( node not in visited and depth_first_search(A__ , A__ , A__ , A__ ) for node in graph ) def lowerCamelCase__ ( A__ : dict , A__ : int , A__ : set , A__ : set ): '''simple docstring''' visited.add(A__ ) rec_stk.add(A__ ) for node in graph[vertex]: if node not in visited: if depth_first_search(A__ , A__ , A__ , A__ ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(A__ ) return False if __name__ == "__main__": from doctest import testmod testmod()
80
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ = { 'configuration_blenderbot_small': [ 'BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlenderbotSmallConfig', 'BlenderbotSmallOnnxConfig', ], 'tokenization_blenderbot_small': ['BlenderbotSmallTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['BlenderbotSmallTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ 'BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST', 'BlenderbotSmallForCausalLM', 'BlenderbotSmallForConditionalGeneration', 'BlenderbotSmallModel', 'BlenderbotSmallPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ 'TFBlenderbotSmallForConditionalGeneration', 'TFBlenderbotSmallModel', 'TFBlenderbotSmallPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ 'FlaxBlenderbotSmallForConditionalGeneration', 'FlaxBlenderbotSmallModel', 'FlaxBlenderbotSmallPreTrainedModel', ] if TYPE_CHECKING: from .configuration_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotSmallConfig, BlenderbotSmallOnnxConfig, ) from .tokenization_blenderbot_small import BlenderbotSmallTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_small_fast import BlenderbotSmallTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotSmallForCausalLM, BlenderbotSmallForConditionalGeneration, BlenderbotSmallModel, BlenderbotSmallPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot_small import ( TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel, TFBlenderbotSmallPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, FlaxBlenderbotSmallPreTrainedModel, ) else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
705
from __future__ import annotations def lowerCamelCase__ ( A__ : list[float] , A__ : list[float] ): '''simple docstring''' __lowerCamelCase = sorted(numsa + numsa ) __lowerCamelCase, __lowerCamelCase = divmod(len(A__ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ = [float(x) for x in input('Enter the elements of first array: ').split()] UpperCAmelCase_ = [float(x) for x in input('Enter the elements of second array: ').split()] print(f"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
80
0
from functools import lru_cache @lru_cache def lowerCamelCase__ ( A__ : int ): '''simple docstring''' if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
706
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) self.assertTrue(torch.allclose(gelu_python(UpperCamelCase_ ) , torch_builtin(UpperCamelCase_ ) ) ) self.assertFalse(torch.allclose(gelu_python(UpperCamelCase_ ) , gelu_new(UpperCamelCase_ ) ) ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = get_activation("""gelu_10""" ) __lowerCamelCase = torch_builtin(UpperCamelCase_ ) __lowerCamelCase = geluaa(UpperCamelCase_ ) __lowerCamelCase = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(UpperCamelCase_ ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def lowerCAmelCase__ ( self: str ): get_activation("""gelu""" ) get_activation("""gelu_10""" ) get_activation("""gelu_fast""" ) get_activation("""gelu_new""" ) get_activation("""gelu_python""" ) get_activation("""gelu_pytorch_tanh""" ) get_activation("""linear""" ) get_activation("""mish""" ) get_activation("""quick_gelu""" ) get_activation("""relu""" ) get_activation("""sigmoid""" ) get_activation("""silu""" ) get_activation("""swish""" ) get_activation("""tanh""" ) with self.assertRaises(UpperCamelCase_ ): get_activation("""bogus""" ) with self.assertRaises(UpperCamelCase_ ): get_activation(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = 1 __lowerCamelCase = get_activation("""gelu""" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(UpperCamelCase_ ): __lowerCamelCase = acta.a
80
0
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record UpperCAmelCase_ = '\\n@article{wang2019superglue,\n title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems},\n author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R},\n journal={arXiv preprint arXiv:1905.00537},\n year={2019}\n}\n' UpperCAmelCase_ = '\\nSuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after\nGLUE with a new set of more difficult language understanding tasks, improved\nresources, and a new public leaderboard.\n' UpperCAmelCase_ = '\nCompute SuperGLUE evaluation metric associated to each SuperGLUE dataset.\nArgs:\n predictions: list of predictions to score. Depending on the SuperGlUE subset:\n - for \'record\': list of question-answer dictionaries with the following keys:\n - \'idx\': index of the question as specified by the dataset\n - \'prediction_text\': the predicted answer text\n - for \'multirc\': list of question-answer dictionaries with the following keys:\n - \'idx\': index of the question-answer pair as specified by the dataset\n - \'prediction\': the predicted answer label\n - otherwise: list of predicted labels\n references: list of reference labels. Depending on the SuperGLUE subset:\n - for \'record\': list of question-answers dictionaries with the following keys:\n - \'idx\': index of the question as specified by the dataset\n - \'answers\': list of possible answers\n - otherwise: list of reference labels\nReturns: depending on the SuperGLUE subset:\n - for \'record\':\n - \'exact_match\': Exact match between answer and gold answer\n - \'f1\': F1 score\n - for \'multirc\':\n - \'exact_match\': Exact match between answer and gold answer\n - \'f1_m\': Per-question macro-F1 score\n - \'f1_a\': Average F1 score over all answers\n - for \'axb\':\n \'matthews_correlation\': Matthew Correlation\n - for \'cb\':\n - \'accuracy\': Accuracy\n - \'f1\': F1 score\n - for all others:\n - \'accuracy\': Accuracy\nExamples:\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'copa\') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"]\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'cb\')\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'record\')\n >>> predictions = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'prediction_text\': \'answer\'}]\n >>> references = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'answers\': [\'answer\', \'another_answer\']}]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 1.0, \'f1\': 1.0}\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'multirc\')\n >>> predictions = [{\'idx\': {\'answer\': 0, \'paragraph\': 0, \'question\': 0}, \'prediction\': 0}, {\'idx\': {\'answer\': 1, \'paragraph\': 2, \'question\': 3}, \'prediction\': 1}]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 1.0, \'f1_m\': 1.0, \'f1_a\': 1.0}\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'axb\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'matthews_correlation\': 1.0}\n' def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : List[str] ): '''simple docstring''' return float((preds == labels).mean() ) def lowerCamelCase__ ( A__ : Optional[Any] , A__ : Optional[Any] , A__ : Optional[int]="binary" ): '''simple docstring''' __lowerCamelCase = simple_accuracy(lowercase_ , lowercase_ ) __lowerCamelCase = float(fa_score(y_true=lowercase_ , y_pred=lowercase_ , average=lowercase_ ) ) return { "accuracy": acc, "f1": fa, } def lowerCamelCase__ ( A__ : List[Any] , A__ : List[str] ): '''simple docstring''' __lowerCamelCase = {} for id_pred, label in zip(lowercase_ , lowercase_ ): __lowerCamelCase = f'{id_pred["idx"]["paragraph"]}-{id_pred["idx"]["question"]}' __lowerCamelCase = id_pred["prediction"] if question_id in question_map: question_map[question_id].append((pred, label) ) else: __lowerCamelCase = [(pred, label)] __lowerCamelCase = [], [] for question, preds_labels in question_map.items(): __lowerCamelCase = zip(*lowercase_ ) __lowerCamelCase = fa_score(y_true=lowercase_ , y_pred=lowercase_ , average="""macro""" ) fas.append(lowercase_ ) __lowerCamelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowercase_ ) ) ems.append(lowercase_ ) __lowerCamelCase = float(sum(lowercase_ ) / len(lowercase_ ) ) __lowerCamelCase = sum(lowercase_ ) / len(lowercase_ ) __lowerCamelCase = float(fa_score(y_true=lowercase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class lowerCamelCase__( datasets.Metric): def lowerCAmelCase__ ( self: Dict ): if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def lowerCAmelCase__ ( self: str ): if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Dict , UpperCamelCase_: Dict ): if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__a , __a )} elif self.config_name == "cb": return acc_and_fa(__a , __a , fa_avg="""macro""" ) elif self.config_name == "record": __lowerCamelCase = [ { "qas": [ {"id": ref["idx"]["query"], "answers": [{"text": ans} for ans in ref["answers"]]} for ref in references ] } ] __lowerCamelCase = {pred["idx"]["query"]: pred["prediction_text"] for pred in predictions} return evaluate_record(__a , __a )[0] elif self.config_name == "multirc": return evaluate_multirc(__a , __a ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__a , __a )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
707
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class lowerCamelCase__( __lowerCamelCase): @slow @require_torch def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = EncoderDecoderModel.from_encoder_decoder_pretrained("""prajjwal1/bert-tiny""" , """prajjwal1/bert-tiny""" ) __lowerCamelCase = BertTokenizer.from_pretrained("""bert-base-uncased""" ) __lowerCamelCase = bertabert.config.encoder.vocab_size __lowerCamelCase = tokenizer.sep_token_id __lowerCamelCase = tokenizer.cls_token_id __lowerCamelCase = 1_28 __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""train[:1%]""" ) __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""validation[:1%]""" ) __lowerCamelCase = train_dataset.select(range(32 ) ) __lowerCamelCase = val_dataset.select(range(16 ) ) __lowerCamelCase = 4 def _map_to_encoder_decoder_inputs(UpperCamelCase_: List[Any] ): # Tokenizer will automatically set [BOS] <text> [EOS] __lowerCamelCase = tokenizer(batch["""article"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=5_12 ) __lowerCamelCase = tokenizer(batch["""highlights"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=1_28 ) __lowerCamelCase = inputs.input_ids __lowerCamelCase = inputs.attention_mask __lowerCamelCase = outputs.input_ids __lowerCamelCase = outputs.input_ids.copy() __lowerCamelCase = [ [-1_00 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["""labels"""] ] __lowerCamelCase = outputs.attention_mask assert all(len(UpperCamelCase_ ) == 5_12 for x in inputs.input_ids ) assert all(len(UpperCamelCase_ ) == 1_28 for x in outputs.input_ids ) return batch def _compute_metrics(UpperCamelCase_: int ): __lowerCamelCase = pred.label_ids __lowerCamelCase = pred.predictions # all unnecessary tokens are removed __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = sum([int(pred_str[i] == label_str[i] ) for i in range(len(UpperCamelCase_ ) )] ) / len(UpperCamelCase_ ) return {"accuracy": accuracy} # map train dataset __lowerCamelCase = train_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) train_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) # same for validation dataset __lowerCamelCase = val_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) val_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) __lowerCamelCase = self.get_auto_remove_tmp_dir() __lowerCamelCase = SeqaSeqTrainingArguments( output_dir=UpperCamelCase_ , per_device_train_batch_size=UpperCamelCase_ , per_device_eval_batch_size=UpperCamelCase_ , predict_with_generate=UpperCamelCase_ , evaluation_strategy="""steps""" , do_train=UpperCamelCase_ , do_eval=UpperCamelCase_ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer __lowerCamelCase = SeqaSeqTrainer( model=UpperCamelCase_ , args=UpperCamelCase_ , compute_metrics=_compute_metrics , train_dataset=UpperCamelCase_ , eval_dataset=UpperCamelCase_ , tokenizer=UpperCamelCase_ , ) # start training trainer.train()
80
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class lowerCamelCase__( UpperCAmelCase_ , UpperCAmelCase_): UpperCAmelCase__ : str = "bit" UpperCAmelCase__ : Any = ["preactivation", "bottleneck"] UpperCAmelCase__ : Optional[int] = ["SAME", "VALID"] def __init__( self: Dict , UpperCamelCase_: List[str]=3 , UpperCamelCase_: Optional[Any]=64 , UpperCamelCase_: Dict=[2_56, 5_12, 10_24, 20_48] , UpperCamelCase_: Any=[3, 4, 6, 3] , UpperCamelCase_: Union[str, Any]="preactivation" , UpperCamelCase_: Optional[int]="relu" , UpperCamelCase_: Optional[Any]=None , UpperCamelCase_: str=32 , UpperCamelCase_: Dict=0.0 , UpperCamelCase_: Any=False , UpperCamelCase_: List[str]=32 , UpperCamelCase_: Optional[int]=1 , UpperCamelCase_: Optional[Any]=None , UpperCamelCase_: Optional[Any]=None , **UpperCamelCase_: int , ): super().__init__(**_snake_case ) if layer_type not in self.layer_types: raise ValueError(F'layer_type={layer_type} is not one of {",".join(self.layer_types )}' ) if global_padding is not None: if global_padding.upper() in self.supported_padding: __lowerCamelCase = global_padding.upper() else: raise ValueError(F'Padding strategy {global_padding} not supported' ) __lowerCamelCase = num_channels __lowerCamelCase = embedding_size __lowerCamelCase = hidden_sizes __lowerCamelCase = depths __lowerCamelCase = layer_type __lowerCamelCase = hidden_act __lowerCamelCase = global_padding __lowerCamelCase = num_groups __lowerCamelCase = drop_path_rate __lowerCamelCase = embedding_dynamic_padding __lowerCamelCase = output_stride __lowerCamelCase = width_factor __lowerCamelCase = ["""stem"""] + [F'stage{idx}' for idx in range(1 , len(_snake_case ) + 1 )] __lowerCamelCase, __lowerCamelCase = get_aligned_output_features_output_indices( out_features=_snake_case , out_indices=_snake_case , stage_names=self.stage_names )
708
class lowerCamelCase__: # Public class to implement a graph def __init__( self: Dict , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): __lowerCamelCase = row __lowerCamelCase = col __lowerCamelCase = graph def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): # Checking all 8 elements surrounding nth element __lowerCamelCase = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __lowerCamelCase = [-1, 0, 1, -1, 1, -1, 0, 1] __lowerCamelCase = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): # And finally, count all islands. __lowerCamelCase = [[False for j in range(self.COL )] for i in range(self.ROW )] __lowerCamelCase = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) count += 1 return count
80
0
from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor UpperCAmelCase_ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def lowerCamelCase__ ( A__ : Optional[int] ): '''simple docstring''' if isinstance(A__ , torch.Tensor ): return image elif isinstance(A__ , PIL.Image.Image ): __lowerCamelCase = [image] __lowerCamelCase = [trans(img.convert("""RGB""" ) ) for img in image] __lowerCamelCase = torch.stack(A__ ) return image class lowerCamelCase__( _lowerCAmelCase): def __init__( self: Dict , UpperCamelCase_: str , UpperCamelCase_: Optional[Any] ): super().__init__() # make sure scheduler can always be converted to DDIM __lowerCamelCase = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=_lowerCAmelCase , scheduler=_lowerCAmelCase ) def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: List[Any] ): if strength < 0 or strength > 1: raise ValueError(F'The value of strength should in [0.0, 1.0] but is {strength}' ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: Any , UpperCamelCase_: List[str] , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = min(int(num_inference_steps * strength ) , _lowerCAmelCase ) __lowerCamelCase = max(num_inference_steps - init_timestep , 0 ) __lowerCamelCase = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCAmelCase__ ( self: int , UpperCamelCase_: str , UpperCamelCase_: Optional[Any] , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: List[Any]=None ): if not isinstance(_lowerCAmelCase , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F'`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_lowerCAmelCase )}' ) __lowerCamelCase = image.to(device=_lowerCAmelCase , dtype=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) and len(_lowerCAmelCase ) != batch_size: raise ValueError( F'You have passed a list of generators of length {len(_lowerCAmelCase )}, but requested an effective batch' F' size of {batch_size}. Make sure the batch size matches the length of the generators.' ) __lowerCamelCase = init_latents.shape __lowerCamelCase = randn_tensor(_lowerCAmelCase , generator=_lowerCAmelCase , device=_lowerCAmelCase , dtype=_lowerCAmelCase ) # get latents print("""add noise to latents at timestep""" , _lowerCAmelCase ) __lowerCamelCase = self.scheduler.add_noise(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowerCamelCase = init_latents return latents @torch.no_grad() def __call__( self: Any , UpperCamelCase_: Union[torch.FloatTensor, PIL.Image.Image] = None , UpperCamelCase_: float = 0.8 , UpperCamelCase_: int = 1 , UpperCamelCase_: Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCamelCase_: float = 0.0 , UpperCamelCase_: int = 50 , UpperCamelCase_: Optional[bool] = None , UpperCamelCase_: Optional[str] = "pil" , UpperCamelCase_: bool = True , ): self.check_inputs(_lowerCAmelCase ) # 2. Preprocess image __lowerCamelCase = preprocess(_lowerCAmelCase ) # 3. set timesteps self.scheduler.set_timesteps(_lowerCAmelCase , device=self.device ) __lowerCamelCase, __lowerCamelCase = self.get_timesteps(_lowerCAmelCase , _lowerCAmelCase , self.device ) __lowerCamelCase = timesteps[:1].repeat(_lowerCAmelCase ) # 4. Prepare latent variables __lowerCamelCase = self.prepare_latents(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , self.unet.dtype , self.device , _lowerCAmelCase ) __lowerCamelCase = latents # 5. Denoising loop for t in self.progress_bar(_lowerCAmelCase ): # 1. predict noise model_output __lowerCamelCase = self.unet(_lowerCAmelCase , _lowerCAmelCase ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 __lowerCamelCase = self.scheduler.step( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , eta=_lowerCAmelCase , use_clipped_model_output=_lowerCAmelCase , generator=_lowerCAmelCase , ).prev_sample __lowerCamelCase = (image / 2 + 0.5).clamp(0 , 1 ) __lowerCamelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __lowerCamelCase = self.numpy_to_pil(_lowerCAmelCase ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=_lowerCAmelCase )
709
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = DPTConfig() if "large" in checkpoint_url: __lowerCamelCase = 1024 __lowerCamelCase = 4096 __lowerCamelCase = 24 __lowerCamelCase = 16 __lowerCamelCase = [5, 11, 17, 23] __lowerCamelCase = [256, 512, 1024, 1024] __lowerCamelCase = (1, 384, 384) if "ade" in checkpoint_url: __lowerCamelCase = True __lowerCamelCase = 150 __lowerCamelCase = """huggingface/label-files""" __lowerCamelCase = """ade20k-id2label.json""" __lowerCamelCase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type="""dataset""" ) ) , """r""" ) ) __lowerCamelCase = {int(A__ ): v for k, v in idalabel.items()} __lowerCamelCase = idalabel __lowerCamelCase = {v: k for k, v in idalabel.items()} __lowerCamelCase = [1, 150, 480, 480] return config, expected_shape def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' __lowerCamelCase = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.encoder""" ) if "pretrained.model" in name: __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.embeddings""" ) if "patch_embed" in name: __lowerCamelCase = name.replace("""patch_embed""" , """patch_embeddings""" ) if "pos_embed" in name: __lowerCamelCase = name.replace("""pos_embed""" , """position_embeddings""" ) if "attn.proj" in name: __lowerCamelCase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "proj" in name and "project" not in name: __lowerCamelCase = name.replace("""proj""" , """projection""" ) if "blocks" in name: __lowerCamelCase = name.replace("""blocks""" , """layer""" ) if "mlp.fc1" in name: __lowerCamelCase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: __lowerCamelCase = name.replace("""mlp.fc2""" , """output.dense""" ) if "norm1" in name: __lowerCamelCase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: __lowerCamelCase = name.replace("""norm2""" , """layernorm_after""" ) if "scratch.output_conv" in name: __lowerCamelCase = name.replace("""scratch.output_conv""" , """head""" ) if "scratch" in name: __lowerCamelCase = name.replace("""scratch""" , """neck""" ) if "layer1_rn" in name: __lowerCamelCase = name.replace("""layer1_rn""" , """convs.0""" ) if "layer2_rn" in name: __lowerCamelCase = name.replace("""layer2_rn""" , """convs.1""" ) if "layer3_rn" in name: __lowerCamelCase = name.replace("""layer3_rn""" , """convs.2""" ) if "layer4_rn" in name: __lowerCamelCase = name.replace("""layer4_rn""" , """convs.3""" ) if "refinenet" in name: __lowerCamelCase = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __lowerCamelCase = name.replace(f'refinenet{layer_idx}' , f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: __lowerCamelCase = name.replace("""out_conv""" , """projection""" ) if "resConfUnit1" in name: __lowerCamelCase = name.replace("""resConfUnit1""" , """residual_layer1""" ) if "resConfUnit2" in name: __lowerCamelCase = name.replace("""resConfUnit2""" , """residual_layer2""" ) if "conv1" in name: __lowerCamelCase = name.replace("""conv1""" , """convolution1""" ) if "conv2" in name: __lowerCamelCase = name.replace("""conv2""" , """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.0.project.0""" , """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.0.project.0""" , """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.0.project.0""" , """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.0.project.0""" , """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.3""" , """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.4""" , """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.3""" , """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.4""" , """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.3""" , """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.3""" , """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.4""" , """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: __lowerCamelCase = name.replace("""pretrained""" , """dpt""" ) if "bn" in name: __lowerCamelCase = name.replace("""bn""" , """batch_norm""" ) if "head" in name: __lowerCamelCase = name.replace("""head""" , """head.head""" ) if "encoder.norm" in name: __lowerCamelCase = name.replace("""encoder.norm""" , """layernorm""" ) if "auxlayer" in name: __lowerCamelCase = name.replace("""auxlayer""" , """auxiliary_head.head""" ) return name def lowerCamelCase__ ( A__ : Tuple , A__ : Any ): '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __lowerCamelCase = in_proj_weight[: config.hidden_size, :] __lowerCamelCase = in_proj_bias[: config.hidden_size] __lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] __lowerCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" __lowerCamelCase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( A__ : Optional[int] , A__ : Union[str, Any] , A__ : List[str] , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = get_dpt_config(A__ ) # load original state_dict from URL __lowerCamelCase = torch.hub.load_state_dict_from_url(A__ , map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(A__ ) # rename keys for key in state_dict.copy().keys(): __lowerCamelCase = state_dict.pop(A__ ) __lowerCamelCase = val # read in qkv matrices read_in_q_k_v(A__ , A__ ) # load HuggingFace model __lowerCamelCase = DPTForSemanticSegmentation(A__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(A__ ) model.load_state_dict(A__ ) model.eval() # Check outputs on an image __lowerCamelCase = 480 if """ade""" in checkpoint_url else 384 __lowerCamelCase = DPTImageProcessor(size=A__ ) __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(A__ , return_tensors="""pt""" ) # forward pass __lowerCamelCase = model(**A__ ).logits if """ade""" in checkpoint_url else model(**A__ ).predicted_depth # Assert logits __lowerCamelCase = torch.tensor([[6.3_199, 6.3_629, 6.4_148], [6.3_850, 6.3_615, 6.4_166], [6.3_519, 6.3_176, 6.3_575]] ) if "ade" in checkpoint_url: __lowerCamelCase = torch.tensor([[4.0_480, 4.2_420, 4.4_360], [4.3_124, 4.5_693, 4.8_261], [4.5_768, 4.8_965, 5.2_163]] ) assert outputs.shape == torch.Size(A__ ) assert ( torch.allclose(outputs[0, 0, :3, :3] , A__ , atol=1E-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3] , A__ ) ) Path(A__ ).mkdir(exist_ok=A__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(A__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=A__ , ) image_processor.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=A__ , ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) UpperCAmelCase_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
80
0
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = HfArgumentParser(snake_case_ ) __lowerCamelCase = parser.parse_args_into_dataclasses()[0] __lowerCamelCase = TensorFlowBenchmark(args=snake_case_ ) try: __lowerCamelCase = parser.parse_args_into_dataclasses()[0] except ValueError as e: __lowerCamelCase = """Arg --no_{0} is no longer used, please use --no-{0} instead.""" __lowerCamelCase = """ """.join(str(snake_case_ ).split(""" """ )[:-1] ) __lowerCamelCase = """""" __lowerCamelCase = eval(str(snake_case_ ).split(""" """ )[-1] ) __lowerCamelCase = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(snake_case_ ) if len(snake_case_ ) > 0: __lowerCamelCase = full_error_msg + begin_error_msg + str(snake_case_ ) raise ValueError(snake_case_ ) benchmark.run() if __name__ == "__main__": main()
710
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
80
0
import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = inspect.getfile(accelerate.test_utils ) __lowerCamelCase = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """external_deps""", """test_metrics.py"""] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 __lowerCamelCase = test_metrics @require_cpu def lowerCAmelCase__ ( self: Tuple ): debug_launcher(self.test_metrics.main , num_processes=1 ) @require_cpu def lowerCAmelCase__ ( self: Optional[Any] ): debug_launcher(self.test_metrics.main ) @require_single_gpu def lowerCAmelCase__ ( self: Optional[Any] ): self.test_metrics.main() @require_multi_gpu def lowerCAmelCase__ ( self: Union[str, Any] ): print(F'Found {torch.cuda.device_count()} devices.' ) __lowerCamelCase = ['''torchrun''', F'--nproc_per_node={torch.cuda.device_count()}', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(UpperCamelCase__ , env=os.environ.copy() )
711
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/config.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/config.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/config.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/config.json', 'bert-base-multilingual-uncased': 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json', 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/config.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/config.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-base-cased-finetuned-mrpc': 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json', 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json', 'bert-base-german-dbmdz-uncased': 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json', 'cl-tohoku/bert-base-japanese': 'https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json', 'cl-tohoku/bert-base-japanese-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json' ), 'wietsedv/bert-base-dutch-cased': 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json', # See all BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = 'bert' def __init__( self: List[str] , UpperCamelCase_: str=3_05_22 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Tuple=12 , UpperCamelCase_: int=12 , UpperCamelCase_: int=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[Any]=0.1 , UpperCamelCase_: Optional[int]=5_12 , UpperCamelCase_: List[Any]=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: List[str]=1E-12 , UpperCamelCase_: Dict=0 , UpperCamelCase_: List[Any]="absolute" , UpperCamelCase_: Tuple=True , UpperCamelCase_: Tuple=None , **UpperCamelCase_: Optional[Any] , ): super().__init__(pad_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = position_embedding_type __lowerCamelCase = use_cache __lowerCamelCase = classifier_dropout class lowerCamelCase__( __lowerCamelCase): @property def lowerCAmelCase__ ( self: Any ): if self.task == "multiple-choice": __lowerCamelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowerCamelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
80
0
import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class lowerCamelCase__( lowerCAmelCase__): def __init__( self: Tuple , UpperCamelCase_: Dict=0.01 , UpperCamelCase_: List[str]=10_00 ): __lowerCamelCase = p_stop __lowerCamelCase = max_length def __iter__( self: Any ): __lowerCamelCase = 0 __lowerCamelCase = False while not stop and count < self.max_length: yield count count += 1 __lowerCamelCase = random.random() < self.p_stop class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Optional[int]=False , UpperCamelCase_: Union[str, Any]=True ): __lowerCamelCase = [ BatchSamplerShard(_lowerCamelCase , 2 , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) for i in range(2 ) ] __lowerCamelCase = [list(_lowerCamelCase ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(_lowerCamelCase ) for shard in batch_sampler_shards] , [len(_lowerCamelCase ) for e in expected] ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = BatchSampler(range(24 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) __lowerCamelCase = BatchSampler(range(24 ) , batch_size=3 , drop_last=_lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. __lowerCamelCase = BatchSampler(range(21 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) __lowerCamelCase = BatchSampler(range(21 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. __lowerCamelCase = BatchSampler(range(22 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) __lowerCamelCase = BatchSampler(range(22 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. __lowerCamelCase = BatchSampler(range(20 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) __lowerCamelCase = BatchSampler(range(20 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) # Check the shards when the dataset is very small. __lowerCamelCase = BatchSampler(range(2 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) __lowerCamelCase = BatchSampler(range(2 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [[], []] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = BatchSampler(range(24 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(24 ) , batch_size=4 , drop_last=_lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. __lowerCamelCase = BatchSampler(range(22 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(22 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. __lowerCamelCase = BatchSampler(range(21 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(21 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase ) # Check the shards when the dataset is very small. __lowerCamelCase = BatchSampler(range(2 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(2 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [[], []] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = BatchSampler(range(24 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(24 ) , batch_size=3 , drop_last=_lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. __lowerCamelCase = BatchSampler(range(21 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(21 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. __lowerCamelCase = BatchSampler(range(22 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(22 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. __lowerCamelCase = BatchSampler(range(20 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(20 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) # Check the shards when the dataset is very small. __lowerCamelCase = BatchSampler(range(2 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [[[0, 1]], []] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(2 ) , batch_size=3 , drop_last=_lowerCamelCase ) __lowerCamelCase = [[], []] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , even_batches=_lowerCamelCase ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = BatchSampler(range(24 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(24 ) , batch_size=4 , drop_last=_lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. __lowerCamelCase = BatchSampler(range(22 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(22 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. __lowerCamelCase = BatchSampler(range(21 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(21 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) # Check the shards when the dataset is very small. __lowerCamelCase = BatchSampler(range(2 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [[[0, 1]], []] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) __lowerCamelCase = BatchSampler(range(2 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = [[], []] self.check_batch_sampler_shards(_lowerCamelCase , _lowerCamelCase , split_batches=_lowerCamelCase , even_batches=_lowerCamelCase ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] __lowerCamelCase = [BatchSamplerShard(_lowerCamelCase , 2 , _lowerCamelCase , even_batches=_lowerCamelCase ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 10, 11]] ) def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: List[Any] , UpperCamelCase_: Optional[int] , UpperCamelCase_: Optional[int] , UpperCamelCase_: Tuple=False , UpperCamelCase_: Optional[int]=2 , UpperCamelCase_: Tuple=False ): random.seed(_lowerCamelCase ) __lowerCamelCase = list(_lowerCamelCase ) __lowerCamelCase = [ IterableDatasetShard( _lowerCamelCase , batch_size=_lowerCamelCase , drop_last=_lowerCamelCase , num_processes=_lowerCamelCase , process_index=_lowerCamelCase , split_batches=_lowerCamelCase , ) for i in range(_lowerCamelCase ) ] __lowerCamelCase = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(_lowerCamelCase ) iterable_dataset_lists.append(list(_lowerCamelCase ) ) __lowerCamelCase = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size __lowerCamelCase = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(_lowerCamelCase ) , len(_lowerCamelCase ) ) self.assertTrue(len(_lowerCamelCase ) % shard_batch_size == 0 ) __lowerCamelCase = [] for idx in range(0 , len(_lowerCamelCase ) , _lowerCamelCase ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(_lowerCamelCase ) < len(_lowerCamelCase ): reference += reference self.assertListEqual(_lowerCamelCase , reference[: len(_lowerCamelCase )] ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = 42 __lowerCamelCase = RandomIterableDataset() self.check_iterable_dataset_shards(_lowerCamelCase , _lowerCamelCase , batch_size=4 , drop_last=_lowerCamelCase , split_batches=_lowerCamelCase ) self.check_iterable_dataset_shards(_lowerCamelCase , _lowerCamelCase , batch_size=4 , drop_last=_lowerCamelCase , split_batches=_lowerCamelCase ) self.check_iterable_dataset_shards(_lowerCamelCase , _lowerCamelCase , batch_size=4 , drop_last=_lowerCamelCase , split_batches=_lowerCamelCase ) self.check_iterable_dataset_shards(_lowerCamelCase , _lowerCamelCase , batch_size=4 , drop_last=_lowerCamelCase , split_batches=_lowerCamelCase ) # Edge case with a very small dataset __lowerCamelCase = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(_lowerCamelCase , _lowerCamelCase , batch_size=4 , drop_last=_lowerCamelCase , split_batches=_lowerCamelCase ) self.check_iterable_dataset_shards(_lowerCamelCase , _lowerCamelCase , batch_size=4 , drop_last=_lowerCamelCase , split_batches=_lowerCamelCase ) self.check_iterable_dataset_shards(_lowerCamelCase , _lowerCamelCase , batch_size=4 , drop_last=_lowerCamelCase , split_batches=_lowerCamelCase ) self.check_iterable_dataset_shards(_lowerCamelCase , _lowerCamelCase , batch_size=4 , drop_last=_lowerCamelCase , split_batches=_lowerCamelCase ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = BatchSampler(range(16 ) , batch_size=4 , drop_last=_lowerCamelCase ) __lowerCamelCase = SkipBatchSampler(_lowerCamelCase , 2 ) self.assertListEqual(list(_lowerCamelCase ) , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = SkipDataLoader(list(range(16 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = DataLoader(list(range(16 ) ) , batch_size=4 ) __lowerCamelCase = skip_first_batches(_lowerCamelCase , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = DataLoaderShard(list(range(16 ) ) , batch_size=4 ) for idx, _ in enumerate(_lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(_lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def lowerCAmelCase__ ( self: str ): Accelerator() __lowerCamelCase = DataLoaderDispatcher(range(16 ) , batch_size=4 ) for idx, _ in enumerate(_lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(_lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
712
from __future__ import annotations from math import ceil, floor, sqrt def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0] __lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target __lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target __lowerCamelCase = 0 # an estimate of b, using the quadratic formula __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the triangle number corresponding to b_floor __lowerCamelCase = 42 # the triangle number corresponding to b_ceil __lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): __lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 __lowerCamelCase = floor(A__ ) __lowerCamelCase = ceil(A__ ) __lowerCamelCase = triangle_numbers[b_floor] __lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_first_guess * triangle_a __lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_second_guess * triangle_a __lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(f"""{solution() = }""")
80
0
'''simple docstring''' import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class lowerCamelCase__( _A , unittest.TestCase): # FIXME: add fast tests pass @nightly @require_onnxruntime @require_torch_gpu class lowerCamelCase__( unittest.TestCase): @property def lowerCAmelCase__ ( self: List[str] ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = ort.SessionOptions() __lowerCamelCase = False return options def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) __lowerCamelCase = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , safety_checker=UpperCamelCase_ , feature_extractor=UpperCamelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) __lowerCamelCase = "A red cat sitting on a park bench" __lowerCamelCase = np.random.RandomState(0 ) __lowerCamelCase = pipe( prompt=UpperCamelCase_ , image=UpperCamelCase_ , mask_image=UpperCamelCase_ , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCamelCase_ , output_type="""np""" , ) __lowerCamelCase = output.images __lowerCamelCase = images[0, 2_55:2_58, 2_55:2_58, -1] assert images.shape == (1, 5_12, 5_12, 3) __lowerCamelCase = np.array([0.2514, 0.3007, 0.3517, 0.1790, 0.2382, 0.3167, 0.1944, 0.2273, 0.2464] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) __lowerCamelCase = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-inpainting""" , subfolder="""scheduler""" , revision="""onnx""" ) __lowerCamelCase = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , scheduler=UpperCamelCase_ , safety_checker=UpperCamelCase_ , feature_extractor=UpperCamelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) __lowerCamelCase = "A red cat sitting on a park bench" __lowerCamelCase = np.random.RandomState(0 ) __lowerCamelCase = pipe( prompt=UpperCamelCase_ , image=UpperCamelCase_ , mask_image=UpperCamelCase_ , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCamelCase_ , output_type="""np""" , ) __lowerCamelCase = output.images __lowerCamelCase = images[0, 2_55:2_58, 2_55:2_58, -1] assert images.shape == (1, 5_12, 5_12, 3) __lowerCamelCase = np.array([0.0086, 0.0077, 0.0083, 0.0093, 0.0107, 0.0139, 0.0094, 0.0097, 0.0125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
713
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[str] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet, attn in zip(self.resnets , self.attentions ): __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: str , UpperCamelCase_: Any , UpperCamelCase_: Optional[int] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet in self.resnets: __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: Tuple , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: List[Any]=True ): for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Optional[Any]=True ): for resnet in self.resnets: # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: int ): # there is always at least one resnet __lowerCamelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] __lowerCamelCase = [] for _ in range(self.num_layers ): __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions def __call__( self: int , UpperCamelCase_: Any , UpperCamelCase_: int , UpperCamelCase_: Dict , UpperCamelCase_: Optional[int]=True ): __lowerCamelCase = self.resnets[0](UpperCamelCase_ , UpperCamelCase_ ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) return hidden_states
80
0
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def lowerCamelCase__ ( A__ : Any , A__ : Union[str, Any] , **A__ : Optional[Any] ): '''simple docstring''' __lowerCamelCase = AutoConfig.from_pretrained(a__ , **a__ ) __lowerCamelCase = AutoModelForSeqaSeqLM.from_config(a__ ) model.save_pretrained(a__ ) AutoTokenizer.from_pretrained(a__ ).save_pretrained(a__ ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
714
import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging UpperCAmelCase_ = ['bart.large', 'bart.large.mnli', 'bart.large.cnn', 'bart_xsum/model.pt'] UpperCAmelCase_ = {'bart.large': BartModel, 'bart.large.mnli': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('0.9.0'): raise Exception('requires fairseq >= 0.9.0') logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = ' Hello world! cécé herlolip' UpperCAmelCase_ = [ ('model.classification_heads.mnli.dense.weight', 'classification_head.dense.weight'), ('model.classification_heads.mnli.dense.bias', 'classification_head.dense.bias'), ('model.classification_heads.mnli.out_proj.weight', 'classification_head.out_proj.weight'), ('model.classification_heads.mnli.out_proj.bias', 'classification_head.out_proj.bias'), ] def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """_float_tensor""", ] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Tuple , A__ : Any , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase = dct.pop(A__ ) __lowerCamelCase = val def lowerCamelCase__ ( A__ : Tuple ): '''simple docstring''' __lowerCamelCase = torch.load(A__ , map_location="""cpu""" ) __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , """bart.large.cnn""" ).eval() hub_interface.model.load_state_dict(sd["""model"""] ) return hub_interface def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = emb.weight.shape __lowerCamelCase = nn.Linear(A__ , A__ , bias=A__ ) __lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Optional[int] , A__ : Dict=None ): '''simple docstring''' if not os.path.exists(A__ ): __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , A__ ).eval() else: __lowerCamelCase = load_xsum_checkpoint(A__ ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: __lowerCamelCase = checkpoint_path.replace(""".""" , """-""" ) __lowerCamelCase = BartConfig.from_pretrained(A__ ) __lowerCamelCase = bart.encode(A__ ).unsqueeze(0 ) __lowerCamelCase = BartTokenizer.from_pretrained(A__ ).encode(A__ , return_tensors="""pt""" ).unsqueeze(0 ) if not torch.eq(A__ , A__ ).all(): raise ValueError( f'converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}' ) if checkpoint_path == "bart.large.mnli": __lowerCamelCase = bart.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""model.decoder.embed_tokens.weight"""] for src, dest in mnli_rename_keys: rename_key(A__ , A__ , A__ ) __lowerCamelCase = BartForSequenceClassification(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = bart.predict("""mnli""" , A__ , return_logits=A__ ) __lowerCamelCase = model(A__ )[0] # logits else: # no classification heads to worry about __lowerCamelCase = bart.model.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""decoder.embed_tokens.weight"""] __lowerCamelCase = bart.extract_features(A__ ) if hf_checkpoint_name == "facebook/bart-large": __lowerCamelCase = BartModel(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = model(A__ ).model[0] else: __lowerCamelCase = BartForConditionalGeneration(A__ ).eval() # an existing summarization ckpt model.model.load_state_dict(A__ ) if hasattr(A__ , """lm_head""" ): __lowerCamelCase = make_linear_from_emb(model.model.shared ) __lowerCamelCase = model.model(A__ )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f'`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}' ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError("""Some values in `fairseq_output` are different from `new_model_outputs`""" ) Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default=None, type=str, help='Which huggingface architecture to use: bart-large-xsum' ) UpperCAmelCase_ = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
80
0
import random from typing import Any def lowerCamelCase__ ( A__ : list ): '''simple docstring''' for _ in range(len(UpperCamelCase__ ) ): __lowerCamelCase = random.randint(0 , len(UpperCamelCase__ ) - 1 ) __lowerCamelCase = random.randint(0 , len(UpperCamelCase__ ) - 1 ) __lowerCamelCase, __lowerCamelCase = data[b], data[a] return data if __name__ == "__main__": UpperCAmelCase_ = [0, 1, 2, 3, 4, 5, 6, 7] UpperCAmelCase_ = ['python', 'says', 'hello', '!'] print('Fisher-Yates Shuffle:') print('List', integers, strings) print('FY Shuffle', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
715
import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class lowerCamelCase__: def __init__( self: Tuple , UpperCamelCase_: Any , UpperCamelCase_: List[Any]=14 , UpperCamelCase_: int=7 , UpperCamelCase_: Union[str, Any]=True , UpperCamelCase_: Dict=True , UpperCamelCase_: Optional[Any]=True , UpperCamelCase_: Tuple=True , UpperCamelCase_: List[str]=True , UpperCamelCase_: int=99 , UpperCamelCase_: str=32 , UpperCamelCase_: List[Any]=5 , UpperCamelCase_: Optional[int]=4 , UpperCamelCase_: List[Any]=37 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[str]=5_12 , UpperCamelCase_: Dict=16 , UpperCamelCase_: List[str]=2 , UpperCamelCase_: Optional[Any]=0.02 , UpperCamelCase_: List[str]=3 , UpperCamelCase_: Tuple=4 , UpperCamelCase_: Tuple=None , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_input_mask __lowerCamelCase = use_labels __lowerCamelCase = use_mc_token_ids __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = self.vocab_size - 1 def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None if self.use_mc_token_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = self.get_config() __lowerCamelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def lowerCAmelCase__ ( self: Dict ): return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: str , UpperCamelCase_: Dict , UpperCamelCase_: Tuple , UpperCamelCase_: Any , UpperCamelCase_: List[str] , *UpperCamelCase_: Optional[Any] ): __lowerCamelCase = CTRLModel(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , head_mask=UpperCamelCase_ ) model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ ) __lowerCamelCase = model(UpperCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Dict , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: List[Any] , *UpperCamelCase_: Tuple ): __lowerCamelCase = CTRLLMHeadModel(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ) = config_and_inputs __lowerCamelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask} return config, inputs_dict def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , *UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = self.num_labels __lowerCamelCase = CTRLForSequenceClassification(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = model(UpperCamelCase_ , token_type_ids=UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : Any = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () UpperCAmelCase__ : Optional[Any] = (CTRLLMHeadModel,) if is_torch_available() else () UpperCAmelCase__ : int = ( { 'feature-extraction': CTRLModel, 'text-classification': CTRLForSequenceClassification, 'text-generation': CTRLLMHeadModel, 'zero-shot': CTRLForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ : List[str] = True UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : Optional[Any] = False def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Any , UpperCamelCase_: List[str] , UpperCamelCase_: Tuple , UpperCamelCase_: Tuple , UpperCamelCase_: List[str] ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = CTRLModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=UpperCamelCase_ , n_embd=37 ) def lowerCAmelCase__ ( self: Optional[int] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__ ( self: Optional[Any] ): self.config_tester.run_common_tests() def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*UpperCamelCase_ ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*UpperCamelCase_ ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCAmelCase__ ( self: List[Any] ): pass @slow def lowerCAmelCase__ ( self: Optional[Any] ): for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = CTRLModel.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) @unittest.skip("""The model doesn't support left padding""" ) # and it's not used enough to be worth fixing :) def lowerCAmelCase__ ( self: Optional[Any] ): pass @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: List[str] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = CTRLLMHeadModel.from_pretrained("""ctrl""" ) model.to(UpperCamelCase_ ) __lowerCamelCase = torch.tensor( [[1_18_59, 0, 16_11, 8]] , dtype=torch.long , device=UpperCamelCase_ ) # Legal the president is __lowerCamelCase = [ 1_18_59, 0, 16_11, 8, 5, 1_50, 2_64_49, 2, 19, 3_48, 4_69, 3, 25_95, 48, 2_07_40, 24_65_33, 24_65_33, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a __lowerCamelCase = model.generate(UpperCamelCase_ , do_sample=UpperCamelCase_ ) self.assertListEqual(output_ids[0].tolist() , UpperCamelCase_ )
80
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['''NllbTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['''NllbTokenizerFast'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
716
def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0 for i in range(n + 1 )] __lowerCamelCase = 1 __lowerCamelCase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , A__ ): __lowerCamelCase = 1 __lowerCamelCase = 0 for i in range(A__ ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(f"""{solution() = }""")
80
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCAmelCase_ = { 'configuration_pix2struct': [ 'PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Pix2StructConfig', 'Pix2StructTextConfig', 'Pix2StructVisionConfig', ], 'processing_pix2struct': ['Pix2StructProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['Pix2StructImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = [ 'PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Pix2StructPreTrainedModel', 'Pix2StructForConditionalGeneration', 'Pix2StructVisionModel', 'Pix2StructTextModel', ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
717
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase): UpperCAmelCase__ : Dict = 1 @register_to_config def __init__( self: List[str] , UpperCamelCase_: int = 10_00 , UpperCamelCase_: Optional[Union[np.ndarray, List[float]]] = None ): # set `betas`, `alphas`, `timesteps` self.set_timesteps(UpperCamelCase_ ) # standard deviation of the initial noise distribution __lowerCamelCase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. __lowerCamelCase = 4 # running values __lowerCamelCase = [] def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: int , UpperCamelCase_: Union[str, torch.device] = None ): __lowerCamelCase = num_inference_steps __lowerCamelCase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] __lowerCamelCase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: __lowerCamelCase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: __lowerCamelCase = torch.sin(steps * math.pi / 2 ) ** 2 __lowerCamelCase = (1.0 - self.betas**2) ** 0.5 __lowerCamelCase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] __lowerCamelCase = timesteps.to(UpperCamelCase_ ) __lowerCamelCase = [] def lowerCAmelCase__ ( self: int , UpperCamelCase_: torch.FloatTensor , UpperCamelCase_: int , UpperCamelCase_: torch.FloatTensor , UpperCamelCase_: bool = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) __lowerCamelCase = (self.timesteps == timestep).nonzero().item() __lowerCamelCase = timestep_index + 1 __lowerCamelCase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(UpperCamelCase_ ) if len(self.ets ) == 1: __lowerCamelCase = self.ets[-1] elif len(self.ets ) == 2: __lowerCamelCase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: __lowerCamelCase = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: __lowerCamelCase = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) __lowerCamelCase = self._get_prev_sample(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: torch.FloatTensor , *UpperCamelCase_: Dict , **UpperCamelCase_: Union[str, Any] ): return sample def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Any , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Any ): __lowerCamelCase = self.alphas[timestep_index] __lowerCamelCase = self.betas[timestep_index] __lowerCamelCase = self.alphas[prev_timestep_index] __lowerCamelCase = self.betas[prev_timestep_index] __lowerCamelCase = (sample - sigma * ets) / max(UpperCamelCase_ , 1E-8 ) __lowerCamelCase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self: List[Any] ): return self.config.num_train_timesteps
80
0
from __future__ import annotations def lowerCamelCase__ ( A__ : Optional[Any] , A__ : Union[str, Any] , A__ : Optional[int] , A__ : Optional[Any] ): # noqa: E741 '''simple docstring''' while r - l > 1: __lowerCamelCase = (l + r) // 2 if v[m] >= key: __lowerCamelCase = m else: __lowerCamelCase = m # noqa: E741 return r def lowerCamelCase__ ( A__ : int ): '''simple docstring''' if len(_lowercase ) == 0: return 0 __lowerCamelCase = [0] * len(_lowercase ) __lowerCamelCase = 1 __lowerCamelCase = v[0] for i in range(1 , len(_lowercase ) ): if v[i] < tail[0]: __lowerCamelCase = v[i] elif v[i] > tail[length - 1]: __lowerCamelCase = v[i] length += 1 else: __lowerCamelCase = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
718
import os from collections.abc import Iterator def lowerCamelCase__ ( A__ : str = "." ): '''simple docstring''' for dir_path, dir_names, filenames in os.walk(A__ ): __lowerCamelCase = [d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(A__ )[1] in (".py", ".ipynb"): yield os.path.join(A__ , A__ ).lstrip("""./""" ) def lowerCamelCase__ ( A__ : Optional[int] ): '''simple docstring''' return f'{i * " "}*' if i else "\n##" def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(A__ ) or old_parts[i] != new_part) and new_part: print(f'{md_prefix(A__ )} {new_part.replace("_" , " " ).title()}' ) return new_path def lowerCamelCase__ ( A__ : str = "." ): '''simple docstring''' __lowerCamelCase = """""" for filepath in sorted(good_file_paths(A__ ) ): __lowerCamelCase, __lowerCamelCase = os.path.split(A__ ) if filepath != old_path: __lowerCamelCase = print_path(A__ , A__ ) __lowerCamelCase = (filepath.count(os.sep ) + 1) if filepath else 0 __lowerCamelCase = f'{filepath}/{filename}'.replace(""" """ , """%20""" ) __lowerCamelCase = os.path.splitext(filename.replace("""_""" , """ """ ).title() )[0] print(f'{md_prefix(A__ )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md('.')
80
0
import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand UpperCAmelCase_ = ( '4S 3H 2C 7S 5H', '9D 8H 2C 6S 7H', '2D 6D 9D TH 7D', 'TC 8C 2S JH 6C', 'JH 8S TH AH QH', 'TS KS 5S 9S AC', 'KD 6S 9D TH AD', 'KS 8D 4D 9S 4S', # pair '8C 4S KH JS 4D', # pair 'QH 8H KD JH 8S', # pair 'KC 4H KS 2H 8D', # pair 'KD 4S KC 3H 8S', # pair 'AH 8S AS KC JH', # pair '3H 4C 4H 3S 2H', # 2 pairs '5S 5D 2C KH KH', # 2 pairs '3C KH 5D 5S KH', # 2 pairs 'AS 3C KH AD KH', # 2 pairs '7C 7S 3S 7H 5S', # 3 of a kind '7C 7S KH 2H 7H', # 3 of a kind 'AC KH QH AH AS', # 3 of a kind '2H 4D 3C AS 5S', # straight (low ace) '3C 5C 4C 2C 6H', # straight '6S 8S 7S 5H 9H', # straight 'JS QS 9H TS KH', # straight 'QC KH TS JS AH', # straight (high ace) '8C 9C 5C 3C TC', # flush '3S 8S 9S 5S KS', # flush '4C 5C 9C 8C KC', # flush 'JH 8H AH KH QH', # flush '3D 2H 3H 2C 2D', # full house '2H 2C 3S 3H 3D', # full house 'KH KC 3S 3H 3D', # full house 'JC 6H JS JD JH', # 4 of a kind 'JC 7H JS JD JH', # 4 of a kind 'JC KH JS JD JH', # 4 of a kind '2S AS 4S 5S 3S', # straight flush (low ace) '2D 6D 3D 4D 5D', # straight flush '5C 6C 3C 7C 4C', # straight flush 'JH 9H TH KH QH', # straight flush 'JH AH TH KH QH', # royal flush (high ace straight flush) ) UpperCAmelCase_ = ( ('2H 3H 4H 5H 6H', 'KS AS TS QS JS', 'Loss'), ('2H 3H 4H 5H 6H', 'AS AD AC AH JD', 'Win'), ('AS AH 2H AD AC', 'JS JD JC JH 3D', 'Win'), ('2S AH 2H AS AC', 'JS JD JC JH AD', 'Loss'), ('2S AH 2H AS AC', '2H 3H 5H 6H 7H', 'Win'), ('AS 3S 4S 8S 2S', '2H 3H 5H 6H 7H', 'Win'), ('2H 3H 5H 6H 7H', '2S 3H 4H 5S 6C', 'Win'), ('2S 3H 4H 5S 6C', '3D 4C 5H 6H 2S', 'Tie'), ('2S 3H 4H 5S 6C', 'AH AC 5H 6H AS', 'Win'), ('2S 2H 4H 5S 4C', 'AH AC 5H 6H AS', 'Loss'), ('2S 2H 4H 5S 4C', 'AH AC 5H 6H 7S', 'Win'), ('6S AD 7H 4S AS', 'AH AC 5H 6H 7S', 'Loss'), ('2S AH 4H 5S KC', 'AH AC 5H 6H 7S', 'Loss'), ('2S 3H 6H 7S 9C', '7H 3C TH 6H 9S', 'Loss'), ('4S 5H 6H TS AC', '3S 5H 6H TS AC', 'Win'), ('2S AH 4H 5S 6C', 'AD 4C 5H 6H 2C', 'Tie'), ('AS AH 3H AD AC', 'AS AH 2H AD AC', 'Win'), ('AH AC 5H 5C QS', 'AH AC 5H 5C KS', 'Loss'), ('AH AC 5H 5C QS', 'KH KC 5H 5C QS', 'Win'), ('7C 7S KH 2H 7H', '3C 3S AH 2H 3H', 'Win'), ('3C 3S AH 2H 3H', '7C 7S KH 2H 7H', 'Loss'), ('6H 5H 4H 3H 2H', '5H 4H 3H 2H AH', 'Win'), ('5H 4H 3H 2H AH', '5H 4H 3H 2H AH', 'Tie'), ('5H 4H 3H 2H AH', '6H 5H 4H 3H 2H', 'Loss'), ('AH AD KS KC AC', 'AH KD KH AC KC', 'Win'), ('2H 4D 3C AS 5S', '2H 4D 3C 6S 5S', 'Loss'), ('2H 3S 3C 3H 2S', '3S 3C 2S 2H 2D', 'Win'), ('4D 6D 5D 2D JH', '3S 8S 3H TC KH', 'Loss'), ('4S 6C 8S 3S 7S', 'AD KS 2D 7D 7C', 'Loss'), ('6S 4C 7H 8C 3H', '5H JC AH 9D 9C', 'Loss'), ('9D 9H JH TC QH', '3C 2S JS 5C 7H', 'Win'), ('2H TC 8S AD 9S', '4H TS 7H 2C 5C', 'Win'), ('9D 3S 2C 7S 7C', 'JC TD 3C TC 9H', 'Loss'), ) UpperCAmelCase_ = ( ('2H 3H 4H 5H 6H', True), ('AS AH 2H AD AC', False), ('2H 3H 5H 6H 7H', True), ('KS AS TS QS JS', True), ('8H 9H QS JS TH', False), ('AS 3S 4S 8S 2S', True), ) UpperCAmelCase_ = ( ('2H 3H 4H 5H 6H', True), ('AS AH 2H AD AC', False), ('2H 3H 5H 6H 7H', False), ('KS AS TS QS JS', True), ('8H 9H QS JS TH', True), ) UpperCAmelCase_ = ( ('2H 4D 3C AS 5S', True, [5, 4, 3, 2, 14]), ('2H 5D 3C AS 5S', False, [14, 5, 5, 3, 2]), ('JH QD KC AS TS', False, [14, 13, 12, 11, 10]), ('9D 3S 2C 7S 7C', False, [9, 7, 7, 3, 2]), ) UpperCAmelCase_ = ( ('JH AH TH KH QH', 0), ('JH 9H TH KH QH', 0), ('JC KH JS JD JH', 7), ('KH KC 3S 3H 3D', 6), ('8C 9C 5C 3C TC', 0), ('JS QS 9H TS KH', 0), ('7C 7S KH 2H 7H', 3), ('3C KH 5D 5S KH', 2), ('QH 8H KD JH 8S', 1), ('2D 6D 9D TH 7D', 0), ) UpperCAmelCase_ = ( ('JH AH TH KH QH', 23), ('JH 9H TH KH QH', 22), ('JC KH JS JD JH', 21), ('KH KC 3S 3H 3D', 20), ('8C 9C 5C 3C TC', 19), ('JS QS 9H TS KH', 18), ('7C 7S KH 2H 7H', 17), ('3C KH 5D 5S KH', 16), ('QH 8H KD JH 8S', 15), ('2D 6D 9D TH 7D', 14), ) def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = randrange(len(_lowerCamelCase ) ), randrange(len(_lowerCamelCase ) ) __lowerCamelCase = ["""Loss""", """Tie""", """Win"""][(play >= oppo) + (play > oppo)] __lowerCamelCase = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def lowerCamelCase__ ( A__ : int = 100 ): '''simple docstring''' return (generate_random_hand() for _ in range(_lowerCamelCase )) @pytest.mark.parametrize("""hand, expected""" , _lowerCamelCase ) def lowerCamelCase__ ( A__ : Optional[int] , A__ : Tuple ): '''simple docstring''' assert PokerHand(_lowerCamelCase )._is_flush() == expected @pytest.mark.parametrize("""hand, expected""" , _lowerCamelCase ) def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Optional[int] ): '''simple docstring''' assert PokerHand(_lowerCamelCase )._is_straight() == expected @pytest.mark.parametrize("""hand, expected, card_values""" , _lowerCamelCase ) def lowerCamelCase__ ( A__ : Dict , A__ : Union[str, Any] , A__ : Optional[Any] ): '''simple docstring''' __lowerCamelCase = PokerHand(_lowerCamelCase ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize("""hand, expected""" , _lowerCamelCase ) def lowerCamelCase__ ( A__ : int , A__ : Union[str, Any] ): '''simple docstring''' assert PokerHand(_lowerCamelCase )._is_same_kind() == expected @pytest.mark.parametrize("""hand, expected""" , _lowerCamelCase ) def lowerCamelCase__ ( A__ : int , A__ : Union[str, Any] ): '''simple docstring''' assert PokerHand(_lowerCamelCase )._hand_type == expected @pytest.mark.parametrize("""hand, other, expected""" , _lowerCamelCase ) def lowerCamelCase__ ( A__ : Dict , A__ : Optional[int] , A__ : Dict ): '''simple docstring''' assert PokerHand(_lowerCamelCase ).compare_with(PokerHand(_lowerCamelCase ) ) == expected @pytest.mark.parametrize("""hand, other, expected""" , generate_random_hands() ) def lowerCamelCase__ ( A__ : Optional[Any] , A__ : Optional[int] , A__ : Tuple ): '''simple docstring''' assert PokerHand(_lowerCamelCase ).compare_with(PokerHand(_lowerCamelCase ) ) == expected def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = [PokerHand(_lowerCamelCase ) for hand in SORTED_HANDS] __lowerCamelCase = poker_hands.copy() shuffle(_lowerCamelCase ) __lowerCamelCase = chain(sorted(_lowerCamelCase ) ) for index, hand in enumerate(_lowerCamelCase ): assert hand == poker_hands[index] def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = [PokerHand("""2D AC 3H 4H 5S""" ), PokerHand("""2S 3H 4H 5S 6C""" )] pokerhands.sort(reverse=_lowerCamelCase ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = PokerHand("""2C 4S AS 3D 5C""" ) __lowerCamelCase = True __lowerCamelCase = [5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = 0 __lowerCamelCase = os.path.abspath(os.path.dirname(_lowerCamelCase ) ) __lowerCamelCase = os.path.join(_lowerCamelCase , """poker_hands.txt""" ) with open(_lowerCamelCase ) as file_hand: for line in file_hand: __lowerCamelCase = line[:14].strip() __lowerCamelCase = line[15:].strip() __lowerCamelCase = PokerHand(_lowerCamelCase ), PokerHand(_lowerCamelCase ) __lowerCamelCase = player.compare_with(_lowerCamelCase ) if output == "Win": answer += 1 assert answer == 376
719
from __future__ import annotations def lowerCamelCase__ ( A__ : list ): '''simple docstring''' if not nums: raise ValueError("""List is empty""" ) return sum(A__ ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
80
0
def lowerCamelCase__ ( A__ : int ): '''simple docstring''' if upper_limit < 0: raise ValueError("""Limit for the Catalan sequence must be ≥ 0""" ) __lowerCamelCase = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 __lowerCamelCase = 1 if upper_limit > 0: __lowerCamelCase = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(lowercase_ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print('\n********* Catalan Numbers Using Dynamic Programming ************\n') print('\n*** Enter -1 at any time to quit ***') print('\nEnter the upper limit (≥ 0) for the Catalan number sequence: ', end='') try: while True: UpperCAmelCase_ = int(input().strip()) if N < 0: print('\n********* Goodbye!! ************') break else: print(f"""The Catalan numbers from 0 through {N} are:""") print(catalan_numbers(N)) print('Try another upper limit for the sequence: ', end='') except (NameError, ValueError): print('\n********* Invalid input, goodbye! ************\n') import doctest doctest.testmod()
720
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__( __lowerCamelCase , __lowerCamelCase): UpperCAmelCase__ : Any = 'maskformer-swin' UpperCAmelCase__ : List[Any] = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Any , UpperCamelCase_: Any=2_24 , UpperCamelCase_: List[str]=4 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: Optional[int]=96 , UpperCamelCase_: List[str]=[2, 2, 6, 2] , UpperCamelCase_: Optional[Any]=[3, 6, 12, 24] , UpperCamelCase_: str=7 , UpperCamelCase_: int=4.0 , UpperCamelCase_: Optional[int]=True , UpperCamelCase_: Union[str, Any]=0.0 , UpperCamelCase_: Optional[int]=0.0 , UpperCamelCase_: Any=0.1 , UpperCamelCase_: Union[str, Any]="gelu" , UpperCamelCase_: int=False , UpperCamelCase_: Optional[int]=0.02 , UpperCamelCase_: Optional[Any]=1E-5 , UpperCamelCase_: Optional[int]=None , UpperCamelCase_: List[Any]=None , **UpperCamelCase_: Union[str, Any] , ): super().__init__(**UpperCamelCase_ ) __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = embed_dim __lowerCamelCase = depths __lowerCamelCase = len(UpperCamelCase_ ) __lowerCamelCase = num_heads __lowerCamelCase = window_size __lowerCamelCase = mlp_ratio __lowerCamelCase = qkv_bias __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_absolute_embeddings __lowerCamelCase = layer_norm_eps __lowerCamelCase = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCamelCase = int(embed_dim * 2 ** (len(UpperCamelCase_ ) - 1) ) __lowerCamelCase = ["""stem"""] + [F'stage{idx}' for idx in range(1 , len(UpperCamelCase_ ) + 1 )] __lowerCamelCase, __lowerCamelCase = get_aligned_output_features_output_indices( out_features=UpperCamelCase_ , out_indices=UpperCamelCase_ , stage_names=self.stage_names )
80
0
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def lowerCamelCase__ ( A__ : List[Any] , A__ : List[Any] , A__ : str ): '''simple docstring''' __lowerCamelCase = TaConfig.from_json_file(lowerCAmelCase__ ) print(f'Building PyTorch model from configuration: {config}' ) __lowerCamelCase = TaForConditionalGeneration(lowerCAmelCase__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) UpperCAmelCase_ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
721
from __future__ import annotations def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): __lowerCamelCase, __lowerCamelCase = array[indexa], array[indexa] def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if length > 1: __lowerCamelCase = int(length / 2 ) for i in range(A__ , low + middle ): comp_and_swap(A__ , A__ , i + middle , A__ ) bitonic_merge(A__ , A__ , A__ , A__ ) bitonic_merge(A__ , low + middle , A__ , A__ ) def lowerCamelCase__ ( A__ : list[int] , A__ : int , A__ : int , A__ : int ): '''simple docstring''' if length > 1: __lowerCamelCase = int(length / 2 ) bitonic_sort(A__ , A__ , A__ , 1 ) bitonic_sort(A__ , low + middle , A__ , 0 ) bitonic_merge(A__ , A__ , A__ , A__ ) if __name__ == "__main__": UpperCAmelCase_ = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase_ = [int(item.strip()) for item in user_input.split(',')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('\nSorted array in ascending order is: ', end='') print(*unsorted, sep=', ') bitonic_merge(unsorted, 0, len(unsorted), 0) print('Sorted array in descending order is: ', end='') print(*unsorted, sep=', ')
80
0
import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase_ = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowerCamelCase__( __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : str = XLNetTokenizer UpperCAmelCase__ : List[Any] = XLNetTokenizerFast UpperCAmelCase__ : int = True UpperCAmelCase__ : str = True def lowerCAmelCase__ ( self: Dict ): super().setUp() # We have a SentencePiece fixture for testing __lowerCamelCase = XLNetTokenizer(UpperCamelCase__ , keep_accents=UpperCamelCase__ ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = """<s>""" __lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCamelCase__ ) , UpperCamelCase__ ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<unk>""" ) self.assertEqual(vocab_keys[1] , """<s>""" ) self.assertEqual(vocab_keys[-1] , """<eod>""" ) self.assertEqual(len(UpperCamelCase__ ) , 10_06 ) def lowerCAmelCase__ ( self: int ): self.assertEqual(self.get_tokenizer().vocab_size , 10_00 ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = XLNetTokenizer(UpperCamelCase__ , keep_accents=UpperCamelCase__ ) __lowerCamelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(UpperCamelCase__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , [2_85, 46, 10, 1_70, 3_82] ) __lowerCamelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = XLNetTokenizer(UpperCamelCase__ , do_lower_case=UpperCamelCase__ ) __lowerCamelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + """""", """i""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """se""", """.""", ] , ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""▁he""", """ll""", """o"""] ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = XLNetTokenizer(UpperCamelCase__ , do_lower_case=UpperCamelCase__ ) __lowerCamelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """se""", """.""", ] , ) @slow def lowerCAmelCase__ ( self: int ): __lowerCamelCase = XLNetTokenizer.from_pretrained("""xlnet-base-cased""" ) __lowerCamelCase = tokenizer.encode("""sequence builders""" , add_special_tokens=UpperCamelCase__ ) __lowerCamelCase = tokenizer.encode("""multi-sequence build""" , add_special_tokens=UpperCamelCase__ ) __lowerCamelCase = tokenizer.build_inputs_with_special_tokens(UpperCamelCase__ ) __lowerCamelCase = tokenizer.build_inputs_with_special_tokens(UpperCamelCase__ , UpperCamelCase__ ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = {"""input_ids""": [[17, 2_14_42, 2_70, 17, 10, 1_46_45, 3_18, 34, 17, 45_46, 31_45, 7_87, 13, 77_52, 2_20_18, 23, 21, 17, 45_46, 31_45, 7_87, 13, 33_52, 1_44_31, 13, 55_00, 11, 11_76, 5_80, 13, 1_68_19, 47_97, 23, 17, 10, 1_71_35, 6_58, 19, 4_57, 79_32, 13, 1_84, 19, 31_54, 1_71_35, 64_68, 19, 14_04, 1_22_69, 19, 42_29, 53_56, 1_62_64, 46, 19, 17, 2_05_45, 1_03_95, 9, 9, 9, 11, 28, 64_21, 95_31, 2_07_29, 17, 10, 3_53, 1_70_22, 11, 21, 64_21, 95_31, 1_69_49, 17, 10, 1_15_09, 7_53, 11, 33, 95, 24_21, 73_85, 9_56, 1_44_31, 26_26, 25, 8_42, 73_85, 48_36, 21, 14_29, 22_72, 98_55, 31_20, 1_61, 2_47_38, 19, 1_32_03, 6_58, 2_18, 7_87, 21, 4_30, 1_84_82, 8_47, 26_37, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_22, 2_21_78, 27, 10_64, 22, 9_56, 13, 1_11_01, 14_29, 58_54, 2_43_13, 1_89_53, 40, 4_22, 2_43_66, 68, 17_58, 37, 1_04_83, 1_42_57, 31, 2_07, 2_63, 21, 2_03, 37_73, 25, 71, 97_35, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 32, 20_49, 34_42, 17, 1_38_94, 33_80, 23, 95, 18, 1_76_34, 22_88, 9, 4, 3]], """token_type_ids""": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCamelCase__ , model_name="""xlnet-base-cased""" , revision="""c841166438c31ec7ca9a106dee7bb312b73ae511""" , )
700
from ... import PretrainedConfig UpperCAmelCase_ = { 'sijunhe/nezha-cn-base': 'https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json', } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Dict = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP UpperCAmelCase__ : Dict = 'nezha' def __init__( self: Dict , UpperCamelCase_: Any=2_11_28 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Optional[int]=12 , UpperCamelCase_: List[str]=12 , UpperCamelCase_: Optional[int]=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: str=0.1 , UpperCamelCase_: Union[str, Any]=5_12 , UpperCamelCase_: Any=64 , UpperCamelCase_: Dict=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: Optional[Any]=1E-12 , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Any=0 , UpperCamelCase_: str=2 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: str=True , **UpperCamelCase_: Any , ): super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = max_relative_position __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = classifier_dropout __lowerCamelCase = use_cache
80
0
'''simple docstring''' class lowerCamelCase__: def __init__( self: List[Any] ): __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = {} def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: int ): if vertex not in self.adjacency: __lowerCamelCase = {} self.num_vertices += 1 def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Any , UpperCamelCase_: Any , UpperCamelCase_: Optional[int] ): self.add_vertex(UpperCamelCase_ ) self.add_vertex(UpperCamelCase_ ) if head == tail: return __lowerCamelCase = weight __lowerCamelCase = weight def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = self.get_edges() for edge in edges: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = edge edges.remove((tail, head, weight) ) for i in range(len(UpperCamelCase_ ) ): __lowerCamelCase = list(edges[i] ) edges.sort(key=lambda UpperCamelCase_ : e[2] ) for i in range(len(UpperCamelCase_ ) - 1 ): if edges[i][2] >= edges[i + 1][2]: __lowerCamelCase = edges[i][2] + 1 for edge in edges: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = edge __lowerCamelCase = weight __lowerCamelCase = weight def __str__( self: int ): __lowerCamelCase = """""" for tail in self.adjacency: for head in self.adjacency[tail]: __lowerCamelCase = self.adjacency[head][tail] string += F'{head} -> {tail} == {weight}\n' return string.rstrip("""\n""" ) def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = [] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def lowerCAmelCase__ ( self: List[Any] ): return self.adjacency.keys() @staticmethod def lowerCAmelCase__ ( UpperCamelCase_: Any=None , UpperCamelCase_: Optional[int]=None ): __lowerCamelCase = Graph() if vertices is None: __lowerCamelCase = [] if edges is None: __lowerCamelCase = [] for vertex in vertices: g.add_vertex(UpperCamelCase_ ) for edge in edges: g.add_edge(*UpperCamelCase_ ) return g class lowerCamelCase__: def __init__( self: List[Any] ): __lowerCamelCase = {} __lowerCamelCase = {} def __len__( self: Union[str, Any] ): return len(self.parent ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Optional[int] ): if item in self.parent: return self.find(UpperCamelCase_ ) __lowerCamelCase = item __lowerCamelCase = 0 return item def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Optional[int] ): if item not in self.parent: return self.make_set(UpperCamelCase_ ) if item != self.parent[item]: __lowerCamelCase = self.find(self.parent[item] ) return self.parent[item] def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: int , UpperCamelCase_: int ): __lowerCamelCase = self.find(UpperCamelCase_ ) __lowerCamelCase = self.find(UpperCamelCase_ ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: __lowerCamelCase = roota return roota if self.rank[roota] < self.rank[roota]: __lowerCamelCase = roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 __lowerCamelCase = roota return roota return None @staticmethod def lowerCAmelCase__ ( UpperCamelCase_: Optional[Any] ): __lowerCamelCase = graph.num_vertices __lowerCamelCase = Graph.UnionFind() __lowerCamelCase = [] while num_components > 1: __lowerCamelCase = {} for vertex in graph.get_vertices(): __lowerCamelCase = -1 __lowerCamelCase = graph.get_edges() for edge in edges: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = edge edges.remove((tail, head, weight) ) for edge in edges: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = edge __lowerCamelCase = union_find.find(UpperCamelCase_ ) __lowerCamelCase = union_find.find(UpperCamelCase_ ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: __lowerCamelCase = [head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: __lowerCamelCase = [head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = cheap_edge[vertex] if union_find.find(UpperCamelCase_ ) != union_find.find(UpperCamelCase_ ): union_find.union(UpperCamelCase_ , UpperCamelCase_ ) mst_edges.append(cheap_edge[vertex] ) __lowerCamelCase = num_components - 1 __lowerCamelCase = Graph.build(edges=UpperCamelCase_ ) return mst
701
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__: def __init__( self: Union[str, Any] , UpperCamelCase_: str = None , UpperCamelCase_: uuid.UUID = None , UpperCamelCase_: Dict=None , UpperCamelCase_: Any=None ): if not conversation_id: __lowerCamelCase = uuid.uuida() if past_user_inputs is None: __lowerCamelCase = [] if generated_responses is None: __lowerCamelCase = [] __lowerCamelCase = conversation_id __lowerCamelCase = past_user_inputs __lowerCamelCase = generated_responses __lowerCamelCase = text def __eq__( self: Optional[Any] , UpperCamelCase_: Union[str, Any] ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: str , UpperCamelCase_: bool = False ): if self.new_user_input: if overwrite: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' F'with: "{text}".' ) __lowerCamelCase = text else: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' F'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input' ) else: __lowerCamelCase = text def lowerCAmelCase__ ( self: List[str] ): if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __lowerCamelCase = None def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): self.generated_responses.append(UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple ): for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self: Union[str, Any] ): __lowerCamelCase = F'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): __lowerCamelCase = """user""" if is_user else """bot""" output += F'{name} >> {text} \n' return output @add_end_docstrings( __lowerCamelCase , r'\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n ' , ) class lowerCamelCase__( __lowerCamelCase): def __init__( self: List[str] , *UpperCamelCase_: List[Any] , **UpperCamelCase_: str ): super().__init__(*UpperCamelCase_ , **UpperCamelCase_ ) if self.tokenizer.pad_token_id is None: __lowerCamelCase = self.tokenizer.eos_token def lowerCAmelCase__ ( self: str , UpperCamelCase_: int=None , UpperCamelCase_: Any=None , UpperCamelCase_: Union[str, Any]=None , **UpperCamelCase_: int ): __lowerCamelCase = {} __lowerCamelCase = {} __lowerCamelCase = {} if min_length_for_response is not None: __lowerCamelCase = min_length_for_response if minimum_tokens is not None: __lowerCamelCase = minimum_tokens if "max_length" in generate_kwargs: __lowerCamelCase = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __lowerCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCamelCase_ ) return preprocess_params, forward_params, postprocess_params def __call__( self: Any , UpperCamelCase_: Union[Conversation, List[Conversation]] , UpperCamelCase_: Optional[int]=0 , **UpperCamelCase_: Optional[int] ): __lowerCamelCase = super().__call__(UpperCamelCase_ , num_workers=UpperCamelCase_ , **UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and len(UpperCamelCase_ ) == 1: return outputs[0] return outputs def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Conversation , UpperCamelCase_: Optional[Any]=32 ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( F'Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ' """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): __lowerCamelCase = self.tokenizer._build_conversation_input_ids(UpperCamelCase_ ) else: # If the tokenizer cannot handle conversations, we default to only the old version __lowerCamelCase = self._legacy_parse_and_tokenize(UpperCamelCase_ ) if self.framework == "pt": __lowerCamelCase = torch.LongTensor([input_ids] ) elif self.framework == "tf": __lowerCamelCase = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: str=10 , **UpperCamelCase_: List[str] ): __lowerCamelCase = generate_kwargs.get("""max_length""" , self.model.config.max_length ) __lowerCamelCase = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(F'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})' ) __lowerCamelCase = max_length - minimum_tokens __lowerCamelCase = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: __lowerCamelCase = model_inputs["""attention_mask"""][:, -trim:] __lowerCamelCase = model_inputs.pop("""conversation""" ) __lowerCamelCase = max_length __lowerCamelCase = self.model.generate(**UpperCamelCase_ , **UpperCamelCase_ ) if self.model.config.is_encoder_decoder: __lowerCamelCase = 1 else: __lowerCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Optional[Any] , UpperCamelCase_: int=True ): __lowerCamelCase = model_outputs["""output_ids"""] __lowerCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_ , ) __lowerCamelCase = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCamelCase_ ) return conversation def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Conversation ): __lowerCamelCase = self.tokenizer.eos_token_id __lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) ) if len(UpperCamelCase_ ) > self.tokenizer.model_max_length: __lowerCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
80
0
UpperCAmelCase_ = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} UpperCAmelCase_ = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def lowerCamelCase__ ( A__ : dict[int, list[int]] , A__ : int , A__ : list[bool] ): '''simple docstring''' __lowerCamelCase = True __lowerCamelCase = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) order.append(SCREAMING_SNAKE_CASE_ ) return order def lowerCamelCase__ ( A__ : dict[int, list[int]] , A__ : int , A__ : list[bool] ): '''simple docstring''' __lowerCamelCase = True __lowerCamelCase = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return component def lowerCamelCase__ ( A__ : dict[int, list[int]] ): '''simple docstring''' __lowerCamelCase = len(SCREAMING_SNAKE_CASE_ ) * [False] __lowerCamelCase = {vert: [] for vert in range(len(SCREAMING_SNAKE_CASE_ ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase = [] for i, was_visited in enumerate(SCREAMING_SNAKE_CASE_ ): if not was_visited: order += topology_sort(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __lowerCamelCase = [] __lowerCamelCase = len(SCREAMING_SNAKE_CASE_ ) * [False] for i in range(len(SCREAMING_SNAKE_CASE_ ) ): __lowerCamelCase = order[len(SCREAMING_SNAKE_CASE_ ) - i - 1] if not visited[vert]: __lowerCamelCase = find_components(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) components_list.append(SCREAMING_SNAKE_CASE_ ) return components_list
702
import math def lowerCamelCase__ ( A__ : int ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = 2 __lowerCamelCase = int(math.sqrt(A__ ) ) # Size of every segment __lowerCamelCase = [True] * (end + 1) __lowerCamelCase = [] while start <= end: if temp[start] is True: in_prime.append(A__ ) for i in range(start * start , end + 1 , A__ ): __lowerCamelCase = False start += 1 prime += in_prime __lowerCamelCase = end + 1 __lowerCamelCase = min(2 * end , A__ ) while low <= n: __lowerCamelCase = [True] * (high - low + 1) for each in in_prime: __lowerCamelCase = math.floor(low / each ) * each if t < low: t += each for j in range(A__ , high + 1 , A__ ): __lowerCamelCase = False for j in range(len(A__ ) ): if temp[j] is True: prime.append(j + low ) __lowerCamelCase = high + 1 __lowerCamelCase = min(high + end , A__ ) return prime print(sieve(10**6))
80
0
from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = { 'repo_name': ['test_repo1', 'test_repo2', 'test_repo3'], 'path': ['test_1.py', 'test_2.py', 'unit_test.py'], 'content': ['a ' * 20, 'a ' * 30, 'b ' * 7], } __lowerCamelCase = Dataset.from_dict(SCREAMING_SNAKE_CASE_ ) return dataset class lowerCamelCase__( __lowerCamelCase): def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = get_dataset() __lowerCamelCase = make_duplicate_clusters(UpperCamelCase_ , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = get_dataset() __lowerCamelCase = deduplicate_dataset(UpperCamelCase_ ) self.assertEqual(len(UpperCamelCase_ ) , 2 ) print(UpperCamelCase_ ) self.assertEqual(duplicate_clusters[0][0]["""copies"""] , 2 ) self.assertEqual(duplicate_clusters[0][0]["""is_extreme"""] , UpperCamelCase_ )
703
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase_ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') class lowerCamelCase__( __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : int = BartphoTokenizer UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : List[str] = True def lowerCAmelCase__ ( self: Tuple ): super().setUp() __lowerCamelCase = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] __lowerCamelCase = dict(zip(UpperCamelCase_ , range(len(UpperCamelCase_ ) ) ) ) __lowerCamelCase = {"""unk_token""": """<unk>"""} __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F'{token} {vocab_tokens[token]}\n' ) __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase__ ( self: List[str] , **UpperCamelCase_: List[str] ): kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: str ): __lowerCamelCase = """This is a là test""" __lowerCamelCase = """This is a<unk><unk> test""" return input_text, output_text def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = BartphoTokenizer(UpperCamelCase_ , self.monolingual_vocab_file , **self.special_tokens_map ) __lowerCamelCase = """This is a là test""" __lowerCamelCase = """▁This ▁is ▁a ▁l à ▁t est""".split() __lowerCamelCase = tokenizer.tokenize(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase_ ) , UpperCamelCase_ )
80
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = BlipImageProcessor() __lowerCamelCase = GPTaTokenizer.from_pretrained("""hf-internal-testing/tiny-random-GPT2Model""" ) __lowerCamelCase = BertTokenizerFast.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) __lowerCamelCase = InstructBlipProcessor(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) processor.save_pretrained(self.tmpdirname ) def lowerCAmelCase__ ( self: str , **UpperCamelCase_: Any ): return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase_ ).tokenizer def lowerCAmelCase__ ( self: int , **UpperCamelCase_: Tuple ): return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase_ ).image_processor def lowerCAmelCase__ ( self: Tuple , **UpperCamelCase_: Optional[int] ): return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase_ ).qformer_tokenizer def lowerCAmelCase__ ( self: str ): shutil.rmtree(self.tmpdirname ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta )] __lowerCamelCase = [Image.fromarray(np.moveaxis(UpperCamelCase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) __lowerCamelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) __lowerCamelCase = self.get_image_processor(do_normalize=UpperCamelCase_ , padding_value=1.0 ) __lowerCamelCase = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCamelCase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCamelCase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase_ ) self.assertIsInstance(processor.qformer_tokenizer , UpperCamelCase_ ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_tokenizer() __lowerCamelCase = self.get_qformer_tokenizer() __lowerCamelCase = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowerCamelCase = self.prepare_image_inputs() __lowerCamelCase = image_processor(UpperCamelCase_ , return_tensors="""np""" ) __lowerCamelCase = processor(images=UpperCamelCase_ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_tokenizer() __lowerCamelCase = self.get_qformer_tokenizer() __lowerCamelCase = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowerCamelCase = 'lower newer' __lowerCamelCase = processor(text=UpperCamelCase_ ) __lowerCamelCase = tokenizer(UpperCamelCase_ , return_token_type_ids=UpperCamelCase_ ) __lowerCamelCase = qformer_tokenizer(UpperCamelCase_ , return_token_type_ids=UpperCamelCase_ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor["""qformer_""" + key] ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_tokenizer() __lowerCamelCase = self.get_qformer_tokenizer() __lowerCamelCase = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowerCamelCase = 'lower newer' __lowerCamelCase = self.prepare_image_inputs() __lowerCamelCase = processor(text=UpperCamelCase_ , images=UpperCamelCase_ ) self.assertListEqual( list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """qformer_input_ids""", """qformer_attention_mask""", """pixel_values"""] , ) # test if it raises when no input is passed with pytest.raises(UpperCamelCase_ ): processor() def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_tokenizer() __lowerCamelCase = self.get_qformer_tokenizer() __lowerCamelCase = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowerCamelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowerCamelCase = processor.batch_decode(UpperCamelCase_ ) __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_tokenizer() __lowerCamelCase = self.get_qformer_tokenizer() __lowerCamelCase = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowerCamelCase = 'lower newer' __lowerCamelCase = self.prepare_image_inputs() __lowerCamelCase = processor(text=UpperCamelCase_ , images=UpperCamelCase_ ) self.assertListEqual( list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """qformer_input_ids""", """qformer_attention_mask""", """pixel_values"""] , )
704
def lowerCamelCase__ ( A__ : dict ): '''simple docstring''' __lowerCamelCase = set() # To detect a back edge, keep track of vertices currently in the recursion stack __lowerCamelCase = set() return any( node not in visited and depth_first_search(A__ , A__ , A__ , A__ ) for node in graph ) def lowerCamelCase__ ( A__ : dict , A__ : int , A__ : set , A__ : set ): '''simple docstring''' visited.add(A__ ) rec_stk.add(A__ ) for node in graph[vertex]: if node not in visited: if depth_first_search(A__ , A__ , A__ , A__ ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(A__ ) return False if __name__ == "__main__": from doctest import testmod testmod()
80
0
import heapq def lowerCamelCase__ ( A__ : List[Any] ): __lowerCamelCase = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(__snake_case , [-1 * len(__snake_case ), (key, value)] ) # chosen_vertices = set of chosen vertices __lowerCamelCase = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices __lowerCamelCase = heapq.heappop(__snake_case )[1][0] chosen_vertices.add(__snake_case ) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: __lowerCamelCase = elem[1][1].index(__snake_case ) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(__snake_case ) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(f"""Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}""")
705
from __future__ import annotations def lowerCamelCase__ ( A__ : list[float] , A__ : list[float] ): '''simple docstring''' __lowerCamelCase = sorted(numsa + numsa ) __lowerCamelCase, __lowerCamelCase = divmod(len(A__ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ = [float(x) for x in input('Enter the elements of first array: ').split()] UpperCAmelCase_ = [float(x) for x in input('Enter the elements of second array: ').split()] print(f"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
80
0
import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__( __a): def __init__( self: List[Any] , *UpperCamelCase_: Optional[int] , **UpperCamelCase_: Optional[int] ): warnings.warn( """The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use YolosImageProcessor instead.""" , snake_case__ , ) super().__init__(*snake_case__ , **snake_case__ )
706
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) self.assertTrue(torch.allclose(gelu_python(UpperCamelCase_ ) , torch_builtin(UpperCamelCase_ ) ) ) self.assertFalse(torch.allclose(gelu_python(UpperCamelCase_ ) , gelu_new(UpperCamelCase_ ) ) ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = get_activation("""gelu_10""" ) __lowerCamelCase = torch_builtin(UpperCamelCase_ ) __lowerCamelCase = geluaa(UpperCamelCase_ ) __lowerCamelCase = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(UpperCamelCase_ ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def lowerCAmelCase__ ( self: str ): get_activation("""gelu""" ) get_activation("""gelu_10""" ) get_activation("""gelu_fast""" ) get_activation("""gelu_new""" ) get_activation("""gelu_python""" ) get_activation("""gelu_pytorch_tanh""" ) get_activation("""linear""" ) get_activation("""mish""" ) get_activation("""quick_gelu""" ) get_activation("""relu""" ) get_activation("""sigmoid""" ) get_activation("""silu""" ) get_activation("""swish""" ) get_activation("""tanh""" ) with self.assertRaises(UpperCamelCase_ ): get_activation("""bogus""" ) with self.assertRaises(UpperCamelCase_ ): get_activation(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = get_activation("""gelu""" ) __lowerCamelCase = 1 __lowerCamelCase = get_activation("""gelu""" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(UpperCamelCase_ ): __lowerCamelCase = acta.a
80
0
import itertools import math def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCAmelCase_ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = 2 while True: if is_prime(lowerCAmelCase_ ): yield num num += 1 def lowerCamelCase__ ( A__ : Any = 10001 ): '''simple docstring''' return next(itertools.islice(prime_generator() , nth - 1 , lowerCAmelCase_ ) ) if __name__ == "__main__": print(f"""{solution() = }""")
707
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class lowerCamelCase__( __lowerCamelCase): @slow @require_torch def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = EncoderDecoderModel.from_encoder_decoder_pretrained("""prajjwal1/bert-tiny""" , """prajjwal1/bert-tiny""" ) __lowerCamelCase = BertTokenizer.from_pretrained("""bert-base-uncased""" ) __lowerCamelCase = bertabert.config.encoder.vocab_size __lowerCamelCase = tokenizer.sep_token_id __lowerCamelCase = tokenizer.cls_token_id __lowerCamelCase = 1_28 __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""train[:1%]""" ) __lowerCamelCase = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""validation[:1%]""" ) __lowerCamelCase = train_dataset.select(range(32 ) ) __lowerCamelCase = val_dataset.select(range(16 ) ) __lowerCamelCase = 4 def _map_to_encoder_decoder_inputs(UpperCamelCase_: List[Any] ): # Tokenizer will automatically set [BOS] <text> [EOS] __lowerCamelCase = tokenizer(batch["""article"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=5_12 ) __lowerCamelCase = tokenizer(batch["""highlights"""] , padding="""max_length""" , truncation=UpperCamelCase_ , max_length=1_28 ) __lowerCamelCase = inputs.input_ids __lowerCamelCase = inputs.attention_mask __lowerCamelCase = outputs.input_ids __lowerCamelCase = outputs.input_ids.copy() __lowerCamelCase = [ [-1_00 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["""labels"""] ] __lowerCamelCase = outputs.attention_mask assert all(len(UpperCamelCase_ ) == 5_12 for x in inputs.input_ids ) assert all(len(UpperCamelCase_ ) == 1_28 for x in outputs.input_ids ) return batch def _compute_metrics(UpperCamelCase_: int ): __lowerCamelCase = pred.label_ids __lowerCamelCase = pred.predictions # all unnecessary tokens are removed __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = tokenizer.batch_decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) __lowerCamelCase = sum([int(pred_str[i] == label_str[i] ) for i in range(len(UpperCamelCase_ ) )] ) / len(UpperCamelCase_ ) return {"accuracy": accuracy} # map train dataset __lowerCamelCase = train_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) train_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) # same for validation dataset __lowerCamelCase = val_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCamelCase_ , batch_size=UpperCamelCase_ , remove_columns=["""article""", """highlights"""] , ) val_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) __lowerCamelCase = self.get_auto_remove_tmp_dir() __lowerCamelCase = SeqaSeqTrainingArguments( output_dir=UpperCamelCase_ , per_device_train_batch_size=UpperCamelCase_ , per_device_eval_batch_size=UpperCamelCase_ , predict_with_generate=UpperCamelCase_ , evaluation_strategy="""steps""" , do_train=UpperCamelCase_ , do_eval=UpperCamelCase_ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer __lowerCamelCase = SeqaSeqTrainer( model=UpperCamelCase_ , args=UpperCamelCase_ , compute_metrics=_compute_metrics , train_dataset=UpperCamelCase_ , eval_dataset=UpperCamelCase_ , tokenizer=UpperCamelCase_ , ) # start training trainer.train()
80
0
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/config.json', 'funnel-transformer/small-base': 'https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json', 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/config.json', 'funnel-transformer/medium-base': 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json', 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/config.json', 'funnel-transformer/large-base': 'https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json', 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json', 'funnel-transformer/xlarge-base': 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json', } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : int = 'funnel' UpperCAmelCase__ : Dict = { 'hidden_size': 'd_model', 'num_attention_heads': 'n_head', } def __init__( self: Dict , UpperCamelCase_: List[Any]=3_05_22 , UpperCamelCase_: Optional[int]=[4, 4, 4] , UpperCamelCase_: str=None , UpperCamelCase_: Optional[int]=2 , UpperCamelCase_: Dict=7_68 , UpperCamelCase_: Dict=12 , UpperCamelCase_: int=64 , UpperCamelCase_: Tuple=30_72 , UpperCamelCase_: int="gelu_new" , UpperCamelCase_: Optional[int]=0.1 , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[str]=0.0 , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: str=None , UpperCamelCase_: Tuple=1E-9 , UpperCamelCase_: Tuple="mean" , UpperCamelCase_: Optional[Any]="relative_shift" , UpperCamelCase_: Any=True , UpperCamelCase_: Optional[int]=True , UpperCamelCase_: str=True , **UpperCamelCase_: List[Any] , ): __lowerCamelCase = vocab_size __lowerCamelCase = block_sizes __lowerCamelCase = [1] * len(UpperCamelCase_ ) if block_repeats is None else block_repeats assert len(UpperCamelCase_ ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." __lowerCamelCase = num_decoder_layers __lowerCamelCase = d_model __lowerCamelCase = n_head __lowerCamelCase = d_head __lowerCamelCase = d_inner __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout __lowerCamelCase = attention_dropout __lowerCamelCase = activation_dropout __lowerCamelCase = initializer_range __lowerCamelCase = initializer_std __lowerCamelCase = layer_norm_eps assert pooling_type in [ "mean", "max", ], F'Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported.' __lowerCamelCase = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F'Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported.' __lowerCamelCase = attention_type __lowerCamelCase = separate_cls __lowerCamelCase = truncate_seq __lowerCamelCase = pool_q_only super().__init__(**UpperCamelCase_ ) @property def lowerCAmelCase__ ( self: Any ): return sum(self.block_sizes ) @num_hidden_layers.setter def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: Tuple ): raise NotImplementedError( """This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.""" ) @property def lowerCAmelCase__ ( self: int ): return len(self.block_sizes ) @num_blocks.setter def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: List[Any] ): raise NotImplementedError("""This model does not support the setting of `num_blocks`. Please set `block_sizes`.""" )
708
class lowerCamelCase__: # Public class to implement a graph def __init__( self: Dict , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): __lowerCamelCase = row __lowerCamelCase = col __lowerCamelCase = graph def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: list[list[bool]] ): # Checking all 8 elements surrounding nth element __lowerCamelCase = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __lowerCamelCase = [-1, 0, 1, -1, 1, -1, 0, 1] __lowerCamelCase = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] ): # And finally, count all islands. __lowerCamelCase = [[False for j in range(self.COL )] for i in range(self.ROW )] __lowerCamelCase = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) count += 1 return count
80
0
import fire from utils import calculate_rouge, save_json def lowerCamelCase__ ( A__ : List[Any] , A__ : Union[str, Any] , A__ : List[Any]=None , **A__ : int ): '''simple docstring''' __lowerCamelCase = [x.strip() for x in open(_A ).readlines()] __lowerCamelCase = [x.strip() for x in open(_A ).readlines()][: len(_A )] __lowerCamelCase = calculate_rouge(_A , _A , **_A ) if save_path is not None: save_json(_A , _A , indent=_A ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
709
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = DPTConfig() if "large" in checkpoint_url: __lowerCamelCase = 1024 __lowerCamelCase = 4096 __lowerCamelCase = 24 __lowerCamelCase = 16 __lowerCamelCase = [5, 11, 17, 23] __lowerCamelCase = [256, 512, 1024, 1024] __lowerCamelCase = (1, 384, 384) if "ade" in checkpoint_url: __lowerCamelCase = True __lowerCamelCase = 150 __lowerCamelCase = """huggingface/label-files""" __lowerCamelCase = """ade20k-id2label.json""" __lowerCamelCase = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type="""dataset""" ) ) , """r""" ) ) __lowerCamelCase = {int(A__ ): v for k, v in idalabel.items()} __lowerCamelCase = idalabel __lowerCamelCase = {v: k for k, v in idalabel.items()} __lowerCamelCase = [1, 150, 480, 480] return config, expected_shape def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' __lowerCamelCase = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Dict ): '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.encoder""" ) if "pretrained.model" in name: __lowerCamelCase = name.replace("""pretrained.model""" , """dpt.embeddings""" ) if "patch_embed" in name: __lowerCamelCase = name.replace("""patch_embed""" , """patch_embeddings""" ) if "pos_embed" in name: __lowerCamelCase = name.replace("""pos_embed""" , """position_embeddings""" ) if "attn.proj" in name: __lowerCamelCase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "proj" in name and "project" not in name: __lowerCamelCase = name.replace("""proj""" , """projection""" ) if "blocks" in name: __lowerCamelCase = name.replace("""blocks""" , """layer""" ) if "mlp.fc1" in name: __lowerCamelCase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: __lowerCamelCase = name.replace("""mlp.fc2""" , """output.dense""" ) if "norm1" in name: __lowerCamelCase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: __lowerCamelCase = name.replace("""norm2""" , """layernorm_after""" ) if "scratch.output_conv" in name: __lowerCamelCase = name.replace("""scratch.output_conv""" , """head""" ) if "scratch" in name: __lowerCamelCase = name.replace("""scratch""" , """neck""" ) if "layer1_rn" in name: __lowerCamelCase = name.replace("""layer1_rn""" , """convs.0""" ) if "layer2_rn" in name: __lowerCamelCase = name.replace("""layer2_rn""" , """convs.1""" ) if "layer3_rn" in name: __lowerCamelCase = name.replace("""layer3_rn""" , """convs.2""" ) if "layer4_rn" in name: __lowerCamelCase = name.replace("""layer4_rn""" , """convs.3""" ) if "refinenet" in name: __lowerCamelCase = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __lowerCamelCase = name.replace(f'refinenet{layer_idx}' , f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: __lowerCamelCase = name.replace("""out_conv""" , """projection""" ) if "resConfUnit1" in name: __lowerCamelCase = name.replace("""resConfUnit1""" , """residual_layer1""" ) if "resConfUnit2" in name: __lowerCamelCase = name.replace("""resConfUnit2""" , """residual_layer2""" ) if "conv1" in name: __lowerCamelCase = name.replace("""conv1""" , """convolution1""" ) if "conv2" in name: __lowerCamelCase = name.replace("""conv2""" , """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.0.project.0""" , """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.0.project.0""" , """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.0.project.0""" , """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.0.project.0""" , """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.3""" , """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess1.4""" , """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.3""" , """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess2.4""" , """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess3.3""" , """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.3""" , """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: __lowerCamelCase = name.replace("""pretrained.act_postprocess4.4""" , """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: __lowerCamelCase = name.replace("""pretrained""" , """dpt""" ) if "bn" in name: __lowerCamelCase = name.replace("""bn""" , """batch_norm""" ) if "head" in name: __lowerCamelCase = name.replace("""head""" , """head.head""" ) if "encoder.norm" in name: __lowerCamelCase = name.replace("""encoder.norm""" , """layernorm""" ) if "auxlayer" in name: __lowerCamelCase = name.replace("""auxlayer""" , """auxiliary_head.head""" ) return name def lowerCamelCase__ ( A__ : Tuple , A__ : Any ): '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) __lowerCamelCase = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __lowerCamelCase = in_proj_weight[: config.hidden_size, :] __lowerCamelCase = in_proj_bias[: config.hidden_size] __lowerCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowerCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __lowerCamelCase = in_proj_weight[ -config.hidden_size :, : ] __lowerCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" __lowerCamelCase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( A__ : Optional[int] , A__ : Union[str, Any] , A__ : List[str] , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = get_dpt_config(A__ ) # load original state_dict from URL __lowerCamelCase = torch.hub.load_state_dict_from_url(A__ , map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(A__ ) # rename keys for key in state_dict.copy().keys(): __lowerCamelCase = state_dict.pop(A__ ) __lowerCamelCase = val # read in qkv matrices read_in_q_k_v(A__ , A__ ) # load HuggingFace model __lowerCamelCase = DPTForSemanticSegmentation(A__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(A__ ) model.load_state_dict(A__ ) model.eval() # Check outputs on an image __lowerCamelCase = 480 if """ade""" in checkpoint_url else 384 __lowerCamelCase = DPTImageProcessor(size=A__ ) __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(A__ , return_tensors="""pt""" ) # forward pass __lowerCamelCase = model(**A__ ).logits if """ade""" in checkpoint_url else model(**A__ ).predicted_depth # Assert logits __lowerCamelCase = torch.tensor([[6.3_199, 6.3_629, 6.4_148], [6.3_850, 6.3_615, 6.4_166], [6.3_519, 6.3_176, 6.3_575]] ) if "ade" in checkpoint_url: __lowerCamelCase = torch.tensor([[4.0_480, 4.2_420, 4.4_360], [4.3_124, 4.5_693, 4.8_261], [4.5_768, 4.8_965, 5.2_163]] ) assert outputs.shape == torch.Size(A__ ) assert ( torch.allclose(outputs[0, 0, :3, :3] , A__ , atol=1E-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3] , A__ ) ) Path(A__ ).mkdir(exist_ok=A__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(A__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=A__ , ) image_processor.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=A__ , ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) UpperCAmelCase_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
80
0
import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict UpperCAmelCase_ = namedtuple( '_TestCommandArgs', [ 'dataset', 'name', 'cache_dir', 'data_dir', 'all_configs', 'save_infos', 'ignore_verifications', 'force_redownload', 'clear_cache', ], defaults=[None, None, None, False, False, False, False, False], ) def lowerCamelCase__ ( A__ : Dict , A__ : str ): '''simple docstring''' return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def lowerCamelCase__ ( A__ : Optional[Any] ): '''simple docstring''' __lowerCamelCase = _TestCommandArgs(dataset=A_ , all_configs=A_ , save_infos=A_ ) __lowerCamelCase = TestCommand(*A_ ) test_command.run() __lowerCamelCase = os.path.join(A_ , """README.md""" ) assert os.path.exists(A_ ) __lowerCamelCase = DatasetInfosDict.from_directory(A_ ) __lowerCamelCase = DatasetInfosDict( { """default""": DatasetInfo( features=Features( { """tokens""": Sequence(Value("""string""" ) ), """ner_tags""": Sequence( ClassLabel(names=["""O""", """B-PER""", """I-PER""", """B-ORG""", """I-ORG""", """B-LOC""", """I-LOC"""] ) ), """langs""": Sequence(Value("""string""" ) ), """spans""": Sequence(Value("""string""" ) ), } ) , splits=[ { """name""": """train""", """num_bytes""": 2351563, """num_examples""": 10000, }, { """name""": """validation""", """num_bytes""": 238418, """num_examples""": 1000, }, ] , download_size=3940680 , dataset_size=2589981 , ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: __lowerCamelCase, __lowerCamelCase = getattr(dataset_infos["""default"""] , A_ ), getattr(expected_dataset_infos["""default"""] , A_ ) if key == "num_bytes": assert is_apercent_close(A_ , A_ ) elif key == "splits": assert list(A_ ) == list(A_ ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes , expected[split].num_bytes ) else: result == expected
710
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
80
0
def lowerCamelCase__ ( A__ : List[str] , A__ : Union[str, Any] = 0 ): '''simple docstring''' __lowerCamelCase = length or len(A__ ) __lowerCamelCase = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: __lowerCamelCase = list_data[i + 1], list_data[i] __lowerCamelCase = True return list_data if not swapped else bubble_sort(A__ , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
711
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/config.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/config.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/config.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/config.json', 'bert-base-multilingual-uncased': 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json', 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/config.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/config.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-base-cased-finetuned-mrpc': 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json', 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json', 'bert-base-german-dbmdz-uncased': 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json', 'cl-tohoku/bert-base-japanese': 'https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json', 'cl-tohoku/bert-base-japanese-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json' ), 'wietsedv/bert-base-dutch-cased': 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json', # See all BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = 'bert' def __init__( self: List[str] , UpperCamelCase_: str=3_05_22 , UpperCamelCase_: Optional[int]=7_68 , UpperCamelCase_: Tuple=12 , UpperCamelCase_: int=12 , UpperCamelCase_: int=30_72 , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[Any]=0.1 , UpperCamelCase_: Optional[int]=5_12 , UpperCamelCase_: List[Any]=2 , UpperCamelCase_: int=0.02 , UpperCamelCase_: List[str]=1E-12 , UpperCamelCase_: Dict=0 , UpperCamelCase_: List[Any]="absolute" , UpperCamelCase_: Tuple=True , UpperCamelCase_: Tuple=None , **UpperCamelCase_: Optional[Any] , ): super().__init__(pad_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = position_embedding_type __lowerCamelCase = use_cache __lowerCamelCase = classifier_dropout class lowerCamelCase__( __lowerCamelCase): @property def lowerCAmelCase__ ( self: Any ): if self.task == "multiple-choice": __lowerCamelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowerCamelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
80
0
from PIL import Image def lowerCamelCase__ ( A__ : List[str] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = image.size __lowerCamelCase = 0 __lowerCamelCase = image.load() for i in range(__snake_case ): for j in range(__snake_case ): __lowerCamelCase = pixels[j, i] mean += pixel mean //= width * height for j in range(__snake_case ): for i in range(__snake_case ): __lowerCamelCase = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": UpperCAmelCase_ = mean_threshold(Image.open('path_to_image').convert('L')) image.save('output_image_path')
712
from __future__ import annotations from math import ceil, floor, sqrt def lowerCamelCase__ ( A__ : int = 2000000 ): '''simple docstring''' __lowerCamelCase = [0] __lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target __lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target __lowerCamelCase = 0 # an estimate of b, using the quadratic formula __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the largest integer less than b_estimate __lowerCamelCase = 42 # the triangle number corresponding to b_floor __lowerCamelCase = 42 # the triangle number corresponding to b_ceil __lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): __lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 __lowerCamelCase = floor(A__ ) __lowerCamelCase = ceil(A__ ) __lowerCamelCase = triangle_numbers[b_floor] __lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_first_guess * triangle_a __lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): __lowerCamelCase = triangle_b_second_guess * triangle_a __lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(f"""{solution() = }""")
80
0
'''simple docstring''' import functools from typing import Any def lowerCamelCase__ ( A__ : Dict , A__ : List[Any] ): '''simple docstring''' if not isinstance(lowerCamelCase_ , lowerCamelCase_ ) or len(lowerCamelCase_ ) == 0: raise ValueError("""the string should be not empty string""" ) if not isinstance(lowerCamelCase_ , lowerCamelCase_ ) or not all( isinstance(lowerCamelCase_ , lowerCamelCase_ ) and len(lowerCamelCase_ ) > 0 for item in words ): raise ValueError("""the words should be a list of non-empty strings""" ) # Build trie __lowerCamelCase = {} __lowerCamelCase = """WORD_KEEPER""" for word in words: __lowerCamelCase = trie for c in word: if c not in trie_node: __lowerCamelCase = {} __lowerCamelCase = trie_node[c] __lowerCamelCase = True __lowerCamelCase = len(lowerCamelCase_ ) # Dynamic programming method @functools.cache def is_breakable(A__ : List[Any] ) -> bool: if index == len_string: return True __lowerCamelCase = trie for i in range(lowerCamelCase_ , lowerCamelCase_ ): __lowerCamelCase = trie_node.get(string[i] , lowerCamelCase_ ) if trie_node is None: return False if trie_node.get(lowerCamelCase_ , lowerCamelCase_ ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
713
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[str] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Dict , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet, attn in zip(self.resnets , self.attentions ): __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=UpperCamelCase_ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_downsample: __lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: str , UpperCamelCase_: Any , UpperCamelCase_: Optional[int] , UpperCamelCase_: int=True ): __lowerCamelCase = () for resnet in self.resnets: __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) output_states += (hidden_states,) if self.add_downsample: __lowerCamelCase = self.downsamplers_a(UpperCamelCase_ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [] __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: Tuple , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: List[Any]=True ): for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = True UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = [] for i in range(self.num_layers ): __lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets if self.add_upsample: __lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Optional[Any]=True ): for resnet in self.resnets: # pop res hidden states __lowerCamelCase = res_hidden_states_tuple[-1] __lowerCamelCase = res_hidden_states_tuple[:-1] __lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) if self.add_upsample: __lowerCamelCase = self.upsamplers_a(UpperCamelCase_ ) return hidden_states class lowerCamelCase__( nn.Module): UpperCAmelCase__ : int UpperCAmelCase__ : float = 0.0 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : int = 1 UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : jnp.dtype = jnp.floataa def lowerCAmelCase__ ( self: int ): # there is always at least one resnet __lowerCamelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] __lowerCamelCase = [] for _ in range(self.num_layers ): __lowerCamelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(UpperCamelCase_ ) __lowerCamelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(UpperCamelCase_ ) __lowerCamelCase = resnets __lowerCamelCase = attentions def __call__( self: int , UpperCamelCase_: Any , UpperCamelCase_: int , UpperCamelCase_: Dict , UpperCamelCase_: Optional[int]=True ): __lowerCamelCase = self.resnets[0](UpperCamelCase_ , UpperCamelCase_ ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): __lowerCamelCase = attn(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) __lowerCamelCase = resnet(UpperCamelCase_ , UpperCamelCase_ , deterministic=UpperCamelCase_ ) return hidden_states
80
0
from random import shuffle import tensorflow as tf from numpy import array def lowerCamelCase__ ( A__ : List[str] , A__ : Tuple ): '''simple docstring''' __lowerCamelCase = int(lowerCamelCase__ ) assert noofclusters < len(lowerCamelCase__ ) # Find out the dimensionality __lowerCamelCase = len(vectors[0] ) # Will help select random centroids from among the available vectors __lowerCamelCase = list(range(len(lowerCamelCase__ ) ) ) shuffle(lowerCamelCase__ ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. __lowerCamelCase = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION __lowerCamelCase = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points __lowerCamelCase = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(lowerCamelCase__ ) ] ##These nodes will assign the centroid Variables the appropriate ##values __lowerCamelCase = tf.placeholder("""float64""" , [dim] ) __lowerCamelCase = [] for centroid in centroids: cent_assigns.append(tf.assign(lowerCamelCase__ , lowerCamelCase__ ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) __lowerCamelCase = [tf.Variable(0 ) for i in range(len(lowerCamelCase__ ) )] ##These nodes will assign an assignment Variable the appropriate ##value __lowerCamelCase = tf.placeholder("""int32""" ) __lowerCamelCase = [] for assignment in assignments: cluster_assigns.append(tf.assign(lowerCamelCase__ , lowerCamelCase__ ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input __lowerCamelCase = tf.placeholder("""float""" , [None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors __lowerCamelCase = tf.reduce_mean(lowerCamelCase__ , 0 ) ##Node for computing Euclidean distances # Placeholders for input __lowerCamelCase = tf.placeholder("""float""" , [dim] ) __lowerCamelCase = tf.placeholder("""float""" , [dim] ) __lowerCamelCase = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(lowerCamelCase__ , lowerCamelCase__ ) , 2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input __lowerCamelCase = tf.placeholder("""float""" , [noofclusters] ) __lowerCamelCase = tf.argmin(lowerCamelCase__ , 0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. __lowerCamelCase = tf.initialize_all_variables() # Initialize all variables sess.run(lowerCamelCase__ ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. __lowerCamelCase = 100 for _ in range(lowerCamelCase__ ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(lowerCamelCase__ ) ): __lowerCamelCase = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. __lowerCamelCase = [ sess.run(lowerCamelCase__ , feed_dict={va: vect, va: sess.run(lowerCamelCase__ )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input __lowerCamelCase = sess.run( lowerCamelCase__ , feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n] , feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(lowerCamelCase__ ): # Collect all the vectors assigned to this cluster __lowerCamelCase = [ vectors[i] for i in range(len(lowerCamelCase__ ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location __lowerCamelCase = sess.run( lowerCamelCase__ , feed_dict={mean_input: array(lowerCamelCase__ )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n] , feed_dict={centroid_value: new_location} ) # Return centroids and assignments __lowerCamelCase = sess.run(lowerCamelCase__ ) __lowerCamelCase = sess.run(lowerCamelCase__ ) return centroids, assignments
714
import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging UpperCAmelCase_ = ['bart.large', 'bart.large.mnli', 'bart.large.cnn', 'bart_xsum/model.pt'] UpperCAmelCase_ = {'bart.large': BartModel, 'bart.large.mnli': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('0.9.0'): raise Exception('requires fairseq >= 0.9.0') logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = ' Hello world! cécé herlolip' UpperCAmelCase_ = [ ('model.classification_heads.mnli.dense.weight', 'classification_head.dense.weight'), ('model.classification_heads.mnli.dense.bias', 'classification_head.dense.bias'), ('model.classification_heads.mnli.out_proj.weight', 'classification_head.out_proj.weight'), ('model.classification_heads.mnli.out_proj.bias', 'classification_head.out_proj.bias'), ] def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """_float_tensor""", ] for k in ignore_keys: state_dict.pop(A__ , A__ ) def lowerCamelCase__ ( A__ : Tuple , A__ : Any , A__ : Union[str, Any] ): '''simple docstring''' __lowerCamelCase = dct.pop(A__ ) __lowerCamelCase = val def lowerCamelCase__ ( A__ : Tuple ): '''simple docstring''' __lowerCamelCase = torch.load(A__ , map_location="""cpu""" ) __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , """bart.large.cnn""" ).eval() hub_interface.model.load_state_dict(sd["""model"""] ) return hub_interface def lowerCamelCase__ ( A__ : List[Any] ): '''simple docstring''' __lowerCamelCase, __lowerCamelCase = emb.weight.shape __lowerCamelCase = nn.Linear(A__ , A__ , bias=A__ ) __lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Optional[int] , A__ : Dict=None ): '''simple docstring''' if not os.path.exists(A__ ): __lowerCamelCase = torch.hub.load("""pytorch/fairseq""" , A__ ).eval() else: __lowerCamelCase = load_xsum_checkpoint(A__ ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: __lowerCamelCase = checkpoint_path.replace(""".""" , """-""" ) __lowerCamelCase = BartConfig.from_pretrained(A__ ) __lowerCamelCase = bart.encode(A__ ).unsqueeze(0 ) __lowerCamelCase = BartTokenizer.from_pretrained(A__ ).encode(A__ , return_tensors="""pt""" ).unsqueeze(0 ) if not torch.eq(A__ , A__ ).all(): raise ValueError( f'converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}' ) if checkpoint_path == "bart.large.mnli": __lowerCamelCase = bart.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""model.decoder.embed_tokens.weight"""] for src, dest in mnli_rename_keys: rename_key(A__ , A__ , A__ ) __lowerCamelCase = BartForSequenceClassification(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = bart.predict("""mnli""" , A__ , return_logits=A__ ) __lowerCamelCase = model(A__ )[0] # logits else: # no classification heads to worry about __lowerCamelCase = bart.model.state_dict() remove_ignore_keys_(A__ ) __lowerCamelCase = state_dict["""decoder.embed_tokens.weight"""] __lowerCamelCase = bart.extract_features(A__ ) if hf_checkpoint_name == "facebook/bart-large": __lowerCamelCase = BartModel(A__ ).eval() model.load_state_dict(A__ ) __lowerCamelCase = model(A__ ).model[0] else: __lowerCamelCase = BartForConditionalGeneration(A__ ).eval() # an existing summarization ckpt model.model.load_state_dict(A__ ) if hasattr(A__ , """lm_head""" ): __lowerCamelCase = make_linear_from_emb(model.model.shared ) __lowerCamelCase = model.model(A__ )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f'`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}' ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError("""Some values in `fairseq_output` are different from `new_model_outputs`""" ) Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default=None, type=str, help='Which huggingface architecture to use: bart-large-xsum' ) UpperCAmelCase_ = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
80
0