code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = """M-CLIP""" def __init__( self : Any , lowerCamelCase_ : Optional[Any]=1024 , lowerCamelCase_ : Optional[Any]=768 , **lowerCamelCase_ : List[str] ): """simple docstring""" UpperCamelCase = transformerDimSize UpperCamelCase = imageDimSize super().__init__(**lowerCamelCase_ ) class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = MCLIPConfig def __init__( self : Dict , lowerCamelCase_ : Tuple , *lowerCamelCase_ : Optional[int] , **lowerCamelCase_ : str ): """simple docstring""" super().__init__(lowerCamelCase_ , *lowerCamelCase_ , **lowerCamelCase_ ) UpperCamelCase = XLMRobertaModel(lowerCamelCase_ ) UpperCamelCase = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : int , lowerCamelCase_ : int ): """simple docstring""" UpperCamelCase = self.transformer(input_ids=lowerCamelCase_ , attention_mask=lowerCamelCase_ )[0] UpperCamelCase = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(lowerCamelCase_ ), embs
343
import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand _SCREAMING_SNAKE_CASE = ( """4S 3H 2C 7S 5H""", """9D 8H 2C 6S 7H""", """2D 6D 9D TH 7D""", """TC 8C 2S JH 6C""", """JH 8S TH AH QH""", """TS KS 5S 9S AC""", """KD 6S 9D TH AD""", """KS 8D 4D 9S 4S""", # pair """8C 4S KH JS 4D""", # pair """QH 8H KD JH 8S""", # pair """KC 4H KS 2H 8D""", # pair """KD 4S KC 3H 8S""", # pair """AH 8S AS KC JH""", # pair """3H 4C 4H 3S 2H""", # 2 pairs """5S 5D 2C KH KH""", # 2 pairs """3C KH 5D 5S KH""", # 2 pairs """AS 3C KH AD KH""", # 2 pairs """7C 7S 3S 7H 5S""", # 3 of a kind """7C 7S KH 2H 7H""", # 3 of a kind """AC KH QH AH AS""", # 3 of a kind """2H 4D 3C AS 5S""", # straight (low ace) """3C 5C 4C 2C 6H""", # straight """6S 8S 7S 5H 9H""", # straight """JS QS 9H TS KH""", # straight """QC KH TS JS AH""", # straight (high ace) """8C 9C 5C 3C TC""", # flush """3S 8S 9S 5S KS""", # flush """4C 5C 9C 8C KC""", # flush """JH 8H AH KH QH""", # flush """3D 2H 3H 2C 2D""", # full house """2H 2C 3S 3H 3D""", # full house """KH KC 3S 3H 3D""", # full house """JC 6H JS JD JH""", # 4 of a kind """JC 7H JS JD JH""", # 4 of a kind """JC KH JS JD JH""", # 4 of a kind """2S AS 4S 5S 3S""", # straight flush (low ace) """2D 6D 3D 4D 5D""", # straight flush """5C 6C 3C 7C 4C""", # straight flush """JH 9H TH KH QH""", # straight flush """JH AH TH KH QH""", # royal flush (high ace straight flush) ) _SCREAMING_SNAKE_CASE = ( ("""2H 3H 4H 5H 6H""", """KS AS TS QS JS""", """Loss"""), ("""2H 3H 4H 5H 6H""", """AS AD AC AH JD""", """Win"""), ("""AS AH 2H AD AC""", """JS JD JC JH 3D""", """Win"""), ("""2S AH 2H AS AC""", """JS JD JC JH AD""", """Loss"""), ("""2S AH 2H AS AC""", """2H 3H 5H 6H 7H""", """Win"""), ("""AS 3S 4S 8S 2S""", """2H 3H 5H 6H 7H""", """Win"""), ("""2H 3H 5H 6H 7H""", """2S 3H 4H 5S 6C""", """Win"""), ("""2S 3H 4H 5S 6C""", """3D 4C 5H 6H 2S""", """Tie"""), ("""2S 3H 4H 5S 6C""", """AH AC 5H 6H AS""", """Win"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H AS""", """Loss"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H 7S""", """Win"""), ("""6S AD 7H 4S AS""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S AH 4H 5S KC""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S 3H 6H 7S 9C""", """7H 3C TH 6H 9S""", """Loss"""), ("""4S 5H 6H TS AC""", """3S 5H 6H TS AC""", """Win"""), ("""2S AH 4H 5S 6C""", """AD 4C 5H 6H 2C""", """Tie"""), ("""AS AH 3H AD AC""", """AS AH 2H AD AC""", """Win"""), ("""AH AC 5H 5C QS""", """AH AC 5H 5C KS""", """Loss"""), ("""AH AC 5H 5C QS""", """KH KC 5H 5C QS""", """Win"""), ("""7C 7S KH 2H 7H""", """3C 3S AH 2H 3H""", """Win"""), ("""3C 3S AH 2H 3H""", """7C 7S KH 2H 7H""", """Loss"""), ("""6H 5H 4H 3H 2H""", """5H 4H 3H 2H AH""", """Win"""), ("""5H 4H 3H 2H AH""", """5H 4H 3H 2H AH""", """Tie"""), ("""5H 4H 3H 2H AH""", """6H 5H 4H 3H 2H""", """Loss"""), ("""AH AD KS KC AC""", """AH KD KH AC KC""", """Win"""), ("""2H 4D 3C AS 5S""", """2H 4D 3C 6S 5S""", """Loss"""), ("""2H 3S 3C 3H 2S""", """3S 3C 2S 2H 2D""", """Win"""), ("""4D 6D 5D 2D JH""", """3S 8S 3H TC KH""", """Loss"""), ("""4S 6C 8S 3S 7S""", """AD KS 2D 7D 7C""", """Loss"""), ("""6S 4C 7H 8C 3H""", """5H JC AH 9D 9C""", """Loss"""), ("""9D 9H JH TC QH""", """3C 2S JS 5C 7H""", """Win"""), ("""2H TC 8S AD 9S""", """4H TS 7H 2C 5C""", """Win"""), ("""9D 3S 2C 7S 7C""", """JC TD 3C TC 9H""", """Loss"""), ) _SCREAMING_SNAKE_CASE = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", True), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", False), ("""AS 3S 4S 8S 2S""", True), ) _SCREAMING_SNAKE_CASE = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", False), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", True), ) _SCREAMING_SNAKE_CASE = ( ("""2H 4D 3C AS 5S""", True, [5, 4, 3, 2, 1_4]), ("""2H 5D 3C AS 5S""", False, [1_4, 5, 5, 3, 2]), ("""JH QD KC AS TS""", False, [1_4, 1_3, 1_2, 1_1, 1_0]), ("""9D 3S 2C 7S 7C""", False, [9, 7, 7, 3, 2]), ) _SCREAMING_SNAKE_CASE = ( ("""JH AH TH KH QH""", 0), ("""JH 9H TH KH QH""", 0), ("""JC KH JS JD JH""", 7), ("""KH KC 3S 3H 3D""", 6), ("""8C 9C 5C 3C TC""", 0), ("""JS QS 9H TS KH""", 0), ("""7C 7S KH 2H 7H""", 3), ("""3C KH 5D 5S KH""", 2), ("""QH 8H KD JH 8S""", 1), ("""2D 6D 9D TH 7D""", 0), ) _SCREAMING_SNAKE_CASE = ( ("""JH AH TH KH QH""", 2_3), ("""JH 9H TH KH QH""", 2_2), ("""JC KH JS JD JH""", 2_1), ("""KH KC 3S 3H 3D""", 2_0), ("""8C 9C 5C 3C TC""", 1_9), ("""JS QS 9H TS KH""", 1_8), ("""7C 7S KH 2H 7H""", 1_7), ("""3C KH 5D 5S KH""", 1_6), ("""QH 8H KD JH 8S""", 1_5), ("""2D 6D 9D TH 7D""", 1_4), ) def lowercase( ) -> Dict: '''simple docstring''' UpperCamelCase , UpperCamelCase = randrange(len(UpperCamelCase_ ) ), randrange(len(UpperCamelCase_ ) ) UpperCamelCase = ["""Loss""", """Tie""", """Win"""][(play >= oppo) + (play > oppo)] UpperCamelCase , UpperCamelCase = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def lowercase( UpperCamelCase_ = 100 ) -> List[Any]: '''simple docstring''' return (generate_random_hand() for _ in range(UpperCamelCase_ )) @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Optional[int]: '''simple docstring''' assert PokerHand(UpperCamelCase_ )._is_flush() == expected @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Tuple: '''simple docstring''' assert PokerHand(UpperCamelCase_ )._is_straight() == expected @pytest.mark.parametrize("""hand, expected, card_values""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Dict: '''simple docstring''' UpperCamelCase = PokerHand(UpperCamelCase_ ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Optional[int]: '''simple docstring''' assert PokerHand(UpperCamelCase_ )._is_same_kind() == expected @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Any: '''simple docstring''' assert PokerHand(UpperCamelCase_ )._hand_type == expected @pytest.mark.parametrize("""hand, other, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[Any]: '''simple docstring''' assert PokerHand(UpperCamelCase_ ).compare_with(PokerHand(UpperCamelCase_ ) ) == expected @pytest.mark.parametrize("""hand, other, expected""" , generate_random_hands() ) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> int: '''simple docstring''' assert PokerHand(UpperCamelCase_ ).compare_with(PokerHand(UpperCamelCase_ ) ) == expected def lowercase( ) -> Dict: '''simple docstring''' UpperCamelCase = [PokerHand(UpperCamelCase_ ) for hand in SORTED_HANDS] UpperCamelCase = poker_hands.copy() shuffle(UpperCamelCase_ ) UpperCamelCase = chain(sorted(UpperCamelCase_ ) ) for index, hand in enumerate(UpperCamelCase_ ): assert hand == poker_hands[index] def lowercase( ) -> Union[str, Any]: '''simple docstring''' # Test that five high straights are compared correctly. UpperCamelCase = [PokerHand("""2D AC 3H 4H 5S""" ), PokerHand("""2S 3H 4H 5S 6C""" )] pokerhands.sort(reverse=UpperCamelCase_ ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def lowercase( ) -> str: '''simple docstring''' # Multiple calls to five_high_straight function should still return True # and shouldn't mutate the list in every call other than the first. UpperCamelCase = PokerHand("""2C 4S AS 3D 5C""" ) UpperCamelCase = True UpperCamelCase = [5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def lowercase( ) -> int: '''simple docstring''' # Problem number 54 from Project Euler # Testing from poker_hands.txt file UpperCamelCase = 0 UpperCamelCase = os.path.abspath(os.path.dirname(UpperCamelCase_ ) ) UpperCamelCase = os.path.join(UpperCamelCase_ , """poker_hands.txt""" ) with open(UpperCamelCase_ ) as file_hand: for line in file_hand: UpperCamelCase = line[:14].strip() UpperCamelCase = line[15:].strip() UpperCamelCase , UpperCamelCase = PokerHand(UpperCamelCase_ ), PokerHand(UpperCamelCase_ ) UpperCamelCase = player.compare_with(UpperCamelCase_ ) if output == "Win": answer += 1 assert answer == 376
343
1
'''simple docstring''' from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar _SCREAMING_SNAKE_CASE = TypeVar("T") _SCREAMING_SNAKE_CASE = TypeVar("U") class _lowerCAmelCase ( Generic[T, U] ): """simple docstring""" def __init__( self : Optional[Any] , __snake_case : T | None , __snake_case : U | None )-> Tuple: snake_case = key snake_case = val snake_case = None snake_case = None def __repr__( self : List[Any] )-> Tuple: return ( f'''Node: key: {self.key}, val: {self.val}, ''' f'''has next: {bool(self.next )}, has prev: {bool(self.prev )}''' ) class _lowerCAmelCase ( Generic[T, U] ): """simple docstring""" def __init__( self : List[Any] )-> Dict: snake_case = DoubleLinkedListNode(a_ , a_ ) snake_case = DoubleLinkedListNode(a_ , a_ ) snake_case , snake_case = self.rear, self.head def __repr__( self : Dict )-> Any: snake_case = ["""DoubleLinkedList"""] snake_case = self.head while node.next is not None: rep.append(str(a_ ) ) snake_case = node.next rep.append(str(self.rear ) ) return ",\n ".join(a_ ) def lowerCAmelCase ( self : Optional[Any] , __snake_case : DoubleLinkedListNode[T, U] )-> Optional[Any]: snake_case = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None snake_case = node snake_case = previous snake_case = node snake_case = self.rear def lowerCAmelCase ( self : Dict , __snake_case : DoubleLinkedListNode[T, U] )-> Tuple: if node.prev is None or node.next is None: return None snake_case = node.next snake_case = node.prev snake_case = None snake_case = None return node class _lowerCAmelCase ( Generic[T, U] ): """simple docstring""" snake_case_ = {} def __init__( self : List[str] , __snake_case : int )-> Optional[Any]: snake_case = DoubleLinkedList() snake_case = capacity snake_case = 0 snake_case = 0 snake_case = 0 snake_case = {} def __repr__( self : Optional[int] )-> Any: return ( f'''CacheInfo(hits={self.hits}, misses={self.miss}, ''' f'''capacity={self.capacity}, current size={self.num_keys})''' ) def __contains__( self : Tuple , __snake_case : T )-> List[Any]: return key in self.cache def lowerCAmelCase ( self : Optional[Any] , __snake_case : T )-> Optional[Any]: if key in self.cache: self.hits += 1 snake_case = self.cache[key] snake_case = self.list.remove(self.cache[key] ) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(a_ ) return node.val self.miss += 1 return None def lowerCAmelCase ( self : int , __snake_case : T , __snake_case : U )-> List[str]: if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity snake_case = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(a_ ) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 snake_case = DoubleLinkedListNode(a_ , a_ ) self.list.add(self.cache[key] ) self.num_keys += 1 else: # bump node to the end of the list, update value snake_case = self.list.remove(self.cache[key] ) assert node is not None # node guaranteed to be in list snake_case = value self.list.add(a_ ) @classmethod def lowerCAmelCase ( cls : List[Any] , __snake_case : int = 1_28 )-> str: def cache_decorator_inner(__snake_case : Callable[[T], U] ) -> Callable[..., U]: def cache_decorator_wrapper(*__snake_case : T ) -> U: if func not in cls.decorator_function_to_instance_map: snake_case = LRUCache(a_ ) snake_case = cls.decorator_function_to_instance_map[func].get(args[0] ) if result is None: snake_case = func(*a_ ) cls.decorator_function_to_instance_map[func].put(args[0] , a_ ) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(a_ , """cache_info""" , a_ ) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
358
'''simple docstring''' import requests from bsa import BeautifulSoup def __lowerCamelCase ( __lowerCAmelCase : str = "https://www.worldometers.info/coronavirus" ) -> dict: snake_case = BeautifulSoup(requests.get(__lowerCAmelCase ).text , """html.parser""" ) snake_case = soup.findAll("""h1""" ) snake_case = soup.findAll("""div""" , {"""class""": """maincounter-number"""} ) keys += soup.findAll("""span""" , {"""class""": """panel-title"""} ) values += soup.findAll("""div""" , {"""class""": """number-table-main"""} ) return {key.text.strip(): value.text.strip() for key, value in zip(__lowerCAmelCase , __lowerCAmelCase )} if __name__ == "__main__": print("\033[1m" + "COVID-19 Status of the World" + "\033[0m\n") for key, value in world_covidaa_stats().items(): print(F"""{key}\n{value}\n""")
3
0
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 _snake_case = sys.version_info >= (3, 10) def A ( _lowerCamelCase=None , _lowerCamelCase=None ): '''simple docstring''' return field(default_factory=lambda: default , metadata=_lowerCamelCase ) @dataclass class UpperCAmelCase_ : lowerCamelCase__ = 42 lowerCamelCase__ = 42 lowerCamelCase__ = 42 lowerCamelCase__ = 42 @dataclass class UpperCAmelCase_ : lowerCamelCase__ = 42 lowerCamelCase__ = field(default='toto' , metadata={'help': 'help message'}) @dataclass class UpperCAmelCase_ : lowerCamelCase__ = False lowerCamelCase__ = True lowerCamelCase__ = None class UpperCAmelCase_ ( a): lowerCamelCase__ = 'titi' lowerCamelCase__ = 'toto' class UpperCAmelCase_ ( a): lowerCamelCase__ = 'titi' lowerCamelCase__ = 'toto' lowerCamelCase__ = 42 @dataclass class UpperCAmelCase_ : lowerCamelCase__ = "toto" def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Optional[int] = BasicEnum(self.foo) @dataclass class UpperCAmelCase_ : lowerCamelCase__ = "toto" def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Tuple = MixedTypeEnum(self.foo) @dataclass class UpperCAmelCase_ : lowerCamelCase__ = None lowerCamelCase__ = field(default=a , metadata={'help': 'help message'}) lowerCamelCase__ = None lowerCamelCase__ = list_field(default=[]) lowerCamelCase__ = list_field(default=[]) @dataclass class UpperCAmelCase_ : lowerCamelCase__ = list_field(default=[]) lowerCamelCase__ = list_field(default=[1, 2, 3]) lowerCamelCase__ = list_field(default=['Hallo', 'Bonjour', 'Hello']) lowerCamelCase__ = list_field(default=[0.1, 0.2, 0.3]) @dataclass class UpperCAmelCase_ : lowerCamelCase__ = field() lowerCamelCase__ = field() lowerCamelCase__ = field() def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Optional[Any] = BasicEnum(self.required_enum) @dataclass class UpperCAmelCase_ : lowerCamelCase__ = 42 lowerCamelCase__ = field() lowerCamelCase__ = None lowerCamelCase__ = field(default='toto' , metadata={'help': 'help message'}) lowerCamelCase__ = list_field(default=['Hallo', 'Bonjour', 'Hello']) if is_python_no_less_than_3_10: @dataclass class UpperCAmelCase_ : lowerCamelCase__ = False lowerCamelCase__ = True lowerCamelCase__ = None @dataclass class UpperCAmelCase_ : lowerCamelCase__ = None lowerCamelCase__ = field(default=a , metadata={'help': 'help message'}) lowerCamelCase__ = None lowerCamelCase__ = list_field(default=[]) lowerCamelCase__ = list_field(default=[]) class UpperCAmelCase_ ( unittest.TestCase): def snake_case__ ( self, __a, __a): '''simple docstring''' self.assertEqual(len(a._actions), len(b._actions)) for x, y in zip(a._actions, b._actions): _lowerCAmelCase : int = {k: v for k, v in vars(__a).items() if k != "container"} _lowerCAmelCase : Dict = {k: v for k, v in vars(__a).items() if k != "container"} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get("choices", __a) and yy.get("choices", __a): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx["type"](__a), yy["type"](__a)) del xx["type"], yy["type"] self.assertEqual(__a, __a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : List[str] = HfArgumentParser(__a) _lowerCAmelCase : List[Any] = argparse.ArgumentParser() expected.add_argument("--foo", type=__a, required=__a) expected.add_argument("--bar", type=__a, required=__a) expected.add_argument("--baz", type=__a, required=__a) expected.add_argument("--flag", type=__a, default=__a, const=__a, nargs="?") self.argparsersEqual(__a, __a) _lowerCAmelCase : int = ["--foo", "1", "--baz", "quux", "--bar", "0.5"] ((_lowerCAmelCase) , ) : str = parser.parse_args_into_dataclasses(__a, look_for_args_file=__a) self.assertFalse(example.flag) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Optional[Any] = HfArgumentParser(__a) _lowerCAmelCase : Any = argparse.ArgumentParser() expected.add_argument("--foo", default=42, type=__a) expected.add_argument("--baz", default="toto", type=__a, help="help message") self.argparsersEqual(__a, __a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : List[str] = argparse.ArgumentParser() expected.add_argument("--foo", type=__a, default=__a, const=__a, nargs="?") expected.add_argument("--baz", type=__a, default=__a, const=__a, nargs="?") # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument("--no_baz", action="store_false", default=__a, dest="baz") expected.add_argument("--opt", type=__a, default=__a) _lowerCAmelCase : List[str] = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(__a) for dataclass_type in dataclass_types: _lowerCAmelCase : Any = HfArgumentParser(__a) self.argparsersEqual(__a, __a) _lowerCAmelCase : int = parser.parse_args([]) self.assertEqual(__a, Namespace(foo=__a, baz=__a, opt=__a)) _lowerCAmelCase : int = parser.parse_args(["--foo", "--no_baz"]) self.assertEqual(__a, Namespace(foo=__a, baz=__a, opt=__a)) _lowerCAmelCase : Optional[int] = parser.parse_args(["--foo", "--baz"]) self.assertEqual(__a, Namespace(foo=__a, baz=__a, opt=__a)) _lowerCAmelCase : Dict = parser.parse_args(["--foo", "True", "--baz", "True", "--opt", "True"]) self.assertEqual(__a, Namespace(foo=__a, baz=__a, opt=__a)) _lowerCAmelCase : str = parser.parse_args(["--foo", "False", "--baz", "False", "--opt", "False"]) self.assertEqual(__a, Namespace(foo=__a, baz=__a, opt=__a)) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Optional[int] = HfArgumentParser(__a) _lowerCAmelCase : str = argparse.ArgumentParser() expected.add_argument( "--foo", default="toto", choices=["titi", "toto", 42], type=make_choice_type_function(["titi", "toto", 42]), ) self.argparsersEqual(__a, __a) _lowerCAmelCase : str = parser.parse_args([]) self.assertEqual(args.foo, "toto") _lowerCAmelCase : Any = parser.parse_args_into_dataclasses([])[0] self.assertEqual(enum_ex.foo, MixedTypeEnum.toto) _lowerCAmelCase : Optional[int] = parser.parse_args(["--foo", "titi"]) self.assertEqual(args.foo, "titi") _lowerCAmelCase : List[Any] = parser.parse_args_into_dataclasses(["--foo", "titi"])[0] self.assertEqual(enum_ex.foo, MixedTypeEnum.titi) _lowerCAmelCase : Union[str, Any] = parser.parse_args(["--foo", "42"]) self.assertEqual(args.foo, 42) _lowerCAmelCase : Optional[Any] = parser.parse_args_into_dataclasses(["--foo", "42"])[0] self.assertEqual(enum_ex.foo, MixedTypeEnum.fourtytwo) def snake_case__ ( self): '''simple docstring''' @dataclass class UpperCAmelCase_ : lowerCamelCase__ = "toto" _lowerCAmelCase : Tuple = HfArgumentParser(__a) _lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() expected.add_argument( "--foo", default="toto", choices=("titi", "toto", 42), type=make_choice_type_function(["titi", "toto", 42]), ) self.argparsersEqual(__a, __a) _lowerCAmelCase : Any = parser.parse_args([]) self.assertEqual(args.foo, "toto") _lowerCAmelCase : Optional[Any] = parser.parse_args(["--foo", "titi"]) self.assertEqual(args.foo, "titi") _lowerCAmelCase : Optional[int] = parser.parse_args(["--foo", "42"]) self.assertEqual(args.foo, 42) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : List[str] = HfArgumentParser(__a) _lowerCAmelCase : Any = argparse.ArgumentParser() expected.add_argument("--foo_int", nargs="+", default=[], type=__a) expected.add_argument("--bar_int", nargs="+", default=[1, 2, 3], type=__a) expected.add_argument("--foo_str", nargs="+", default=["Hallo", "Bonjour", "Hello"], type=__a) expected.add_argument("--foo_float", nargs="+", default=[0.1, 0.2, 0.3], type=__a) self.argparsersEqual(__a, __a) _lowerCAmelCase : Optional[Any] = parser.parse_args([]) self.assertEqual( __a, Namespace(foo_int=[], bar_int=[1, 2, 3], foo_str=["Hallo", "Bonjour", "Hello"], foo_float=[0.1, 0.2, 0.3]), ) _lowerCAmelCase : Optional[Any] = parser.parse_args("--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7".split()) self.assertEqual(__a, Namespace(foo_int=[1], bar_int=[2, 3], foo_str=["a", "b", "c"], foo_float=[0.1, 0.7])) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Dict = argparse.ArgumentParser() expected.add_argument("--foo", default=__a, type=__a) expected.add_argument("--bar", default=__a, type=__a, help="help message") expected.add_argument("--baz", default=__a, type=__a) expected.add_argument("--ces", nargs="+", default=[], type=__a) expected.add_argument("--des", nargs="+", default=[], type=__a) _lowerCAmelCase : Optional[Any] = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(__a) for dataclass_type in dataclass_types: _lowerCAmelCase : Tuple = HfArgumentParser(__a) self.argparsersEqual(__a, __a) _lowerCAmelCase : int = parser.parse_args([]) self.assertEqual(__a, Namespace(foo=__a, bar=__a, baz=__a, ces=[], des=[])) _lowerCAmelCase : List[Any] = parser.parse_args("--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3".split()) self.assertEqual(__a, Namespace(foo=12, bar=3.14, baz="42", ces=["a", "b", "c"], des=[1, 2, 3])) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = HfArgumentParser(__a) _lowerCAmelCase : Any = argparse.ArgumentParser() expected.add_argument("--required_list", nargs="+", type=__a, required=__a) expected.add_argument("--required_str", type=__a, required=__a) expected.add_argument( "--required_enum", type=make_choice_type_function(["titi", "toto"]), choices=["titi", "toto"], required=__a, ) self.argparsersEqual(__a, __a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Any = HfArgumentParser(__a) _lowerCAmelCase : Optional[int] = argparse.ArgumentParser() expected.add_argument("--foo", type=__a, required=__a) expected.add_argument( "--required_enum", type=make_choice_type_function(["titi", "toto"]), choices=["titi", "toto"], required=__a, ) expected.add_argument("--opt", type=__a, default=__a) expected.add_argument("--baz", default="toto", type=__a, help="help message") expected.add_argument("--foo_str", nargs="+", default=["Hallo", "Bonjour", "Hello"], type=__a) self.argparsersEqual(__a, __a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Dict = HfArgumentParser(__a) _lowerCAmelCase : Tuple = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } _lowerCAmelCase : List[str] = parser.parse_dict(__a)[0] _lowerCAmelCase : Optional[int] = BasicExample(**__a) self.assertEqual(__a, __a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : str = HfArgumentParser(__a) _lowerCAmelCase : Optional[int] = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, "extra": 42, } self.assertRaises(__a, parser.parse_dict, __a, allow_extra_keys=__a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Tuple = HfArgumentParser(__a) _lowerCAmelCase : Tuple = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } with tempfile.TemporaryDirectory() as tmp_dir: _lowerCAmelCase : Dict = os.path.join(__a, "temp_json") os.mkdir(__a) with open(temp_local_path + ".json", "w+") as f: json.dump(__a, __a) _lowerCAmelCase : Optional[int] = parser.parse_yaml_file(Path(temp_local_path + ".json"))[0] _lowerCAmelCase : str = BasicExample(**__a) self.assertEqual(__a, __a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : str = HfArgumentParser(__a) _lowerCAmelCase : int = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } with tempfile.TemporaryDirectory() as tmp_dir: _lowerCAmelCase : Any = os.path.join(__a, "temp_yaml") os.mkdir(__a) with open(temp_local_path + ".yaml", "w+") as f: yaml.dump(__a, __a) _lowerCAmelCase : str = parser.parse_yaml_file(Path(temp_local_path + ".yaml"))[0] _lowerCAmelCase : Optional[int] = BasicExample(**__a) self.assertEqual(__a, __a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : List[str] = HfArgumentParser(__a) self.assertIsNotNone(__a)
36
"""simple docstring""" import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets lowercase_ = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' lowercase_ = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' lowercase_ = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Any ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, homepage='''https://github.com/krishnap25/mauve''', inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { '''predictions''': datasets.Value('''string''', id='''sequence''' ), '''references''': datasets.Value('''string''', id='''sequence''' ), } ), codebase_urls=['''https://github.com/krishnap25/mauve'''], reference_urls=[ '''https://arxiv.org/abs/2102.01454''', '''https://github.com/krishnap25/mauve''', ], ) def _SCREAMING_SNAKE_CASE ( self : int, _lowerCamelCase : str, _lowerCamelCase : Optional[Any], _lowerCamelCase : Any=None, _lowerCamelCase : Tuple=None, _lowerCamelCase : Optional[Any]=None, _lowerCamelCase : Union[str, Any]=None, _lowerCamelCase : str="auto", _lowerCamelCase : Union[str, Any]=-1, _lowerCamelCase : List[str]=0.9, _lowerCamelCase : int=5, _lowerCamelCase : Tuple=5_00, _lowerCamelCase : Union[str, Any]="gpt2-large", _lowerCamelCase : int=-1, _lowerCamelCase : Union[str, Any]=10_24, _lowerCamelCase : Union[str, Any]=25, _lowerCamelCase : str=5, _lowerCamelCase : Any=True, _lowerCamelCase : Union[str, Any]=25, ): '''simple docstring''' __A = compute_mauve( p_text=_lowerCamelCase, q_text=_lowerCamelCase, p_features=_lowerCamelCase, q_features=_lowerCamelCase, p_tokens=_lowerCamelCase, q_tokens=_lowerCamelCase, num_buckets=_lowerCamelCase, pca_max_data=_lowerCamelCase, kmeans_explained_var=_lowerCamelCase, kmeans_num_redo=_lowerCamelCase, kmeans_max_iter=_lowerCamelCase, featurize_model_name=_lowerCamelCase, device_id=_lowerCamelCase, max_text_length=_lowerCamelCase, divergence_curve_discretization_size=_lowerCamelCase, mauve_scaling_factor=_lowerCamelCase, verbose=_lowerCamelCase, seed=_lowerCamelCase, ) return out
266
0
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A ( UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : Any = UnCLIPImageVariationPipeline __UpperCAmelCase : str = IMAGE_VARIATION_PARAMS - {'height', 'width', 'guidance_scale'} __UpperCAmelCase : Any = IMAGE_VARIATION_BATCH_PARAMS __UpperCAmelCase : str = [ 'generator', 'return_dict', 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] __UpperCAmelCase : Optional[int] = False @property def lowercase_ (self : Tuple ) -> Any: """simple docstring""" return 3_2 @property def lowercase_ (self : Dict ) -> List[Any]: """simple docstring""" return 3_2 @property def lowercase_ (self : str ) -> Optional[Any]: """simple docstring""" return self.time_input_dim @property def lowercase_ (self : Any ) -> Optional[Any]: """simple docstring""" return self.time_input_dim * 4 @property def lowercase_ (self : Optional[Any] ) -> List[Any]: """simple docstring""" return 1_0_0 @property def lowercase_ (self : Optional[int] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) return tokenizer @property def lowercase_ (self : Tuple ) -> Dict: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(__UpperCAmelCase ) @property def lowercase_ (self : Optional[int] ) -> Any: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase__ = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=3_2 , intermediate_size=3_7 , patch_size=1 , ) return CLIPVisionModelWithProjection(__UpperCAmelCase ) @property def lowercase_ (self : List[Any] ) -> Dict: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase__ = { "clip_embeddings_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "cross_attention_dim": self.cross_attention_dim, } UpperCAmelCase__ = UnCLIPTextProjModel(**__UpperCAmelCase ) return model @property def lowercase_ (self : Any ) -> Optional[int]: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase__ = { "sample_size": 3_2, # RGB in channels "in_channels": 3, # Out channels is double in channels because predicts mean and variance "out_channels": 6, "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": "identity", } UpperCAmelCase__ = UNetaDConditionModel(**__UpperCAmelCase ) return model @property def lowercase_ (self : Tuple ) -> List[str]: """simple docstring""" return { "sample_size": 6_4, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def lowercase_ (self : str ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase__ = UNetaDModel(**self.dummy_super_res_kwargs ) return model @property def lowercase_ (self : Tuple ) -> Optional[Any]: """simple docstring""" torch.manual_seed(1 ) UpperCAmelCase__ = UNetaDModel(**self.dummy_super_res_kwargs ) return model def lowercase_ (self : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = self.dummy_decoder UpperCAmelCase__ = self.dummy_text_proj UpperCAmelCase__ = self.dummy_text_encoder UpperCAmelCase__ = self.dummy_tokenizer UpperCAmelCase__ = self.dummy_super_res_first UpperCAmelCase__ = self.dummy_super_res_last UpperCAmelCase__ = UnCLIPScheduler( variance_type="learned_range" , prediction_type="epsilon" , num_train_timesteps=1_0_0_0 , ) UpperCAmelCase__ = UnCLIPScheduler( variance_type="fixed_small_log" , prediction_type="epsilon" , num_train_timesteps=1_0_0_0 , ) UpperCAmelCase__ = CLIPImageProcessor(crop_size=3_2 , size=3_2 ) UpperCAmelCase__ = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : int=0 , __UpperCAmelCase : Optional[Any]=True ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) if str(__UpperCAmelCase ).startswith("mps" ): UpperCAmelCase__ = torch.manual_seed(__UpperCAmelCase ) else: UpperCAmelCase__ = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) if pil_image: UpperCAmelCase__ = input_image * 0.5 + 0.5 UpperCAmelCase__ = input_image.clamp(0 , 1 ) UpperCAmelCase__ = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() UpperCAmelCase__ = DiffusionPipeline.numpy_to_pil(__UpperCAmelCase )[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def lowercase_ (self : Dict ) -> str: """simple docstring""" UpperCAmelCase__ = "cpu" UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = self.pipeline_class(**__UpperCAmelCase ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase , pil_image=__UpperCAmelCase ) UpperCAmelCase__ = pipe(**__UpperCAmelCase ) UpperCAmelCase__ = output.images UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase , pil_image=__UpperCAmelCase ) UpperCAmelCase__ = pipe( **__UpperCAmelCase , return_dict=__UpperCAmelCase , )[0] UpperCAmelCase__ = image[0, -3:, -3:, -1] UpperCAmelCase__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) UpperCAmelCase__ = np.array( [ 0.9997, 0.0002, 0.9997, 0.9997, 0.9969, 0.0023, 0.9997, 0.9969, 0.9970, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowercase_ (self : int ) -> Any: """simple docstring""" UpperCAmelCase__ = "cpu" UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = self.pipeline_class(**__UpperCAmelCase ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase , pil_image=__UpperCAmelCase ) UpperCAmelCase__ = pipe(**__UpperCAmelCase ) UpperCAmelCase__ = output.images UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase , pil_image=__UpperCAmelCase ) UpperCAmelCase__ = pipe( **__UpperCAmelCase , return_dict=__UpperCAmelCase , )[0] UpperCAmelCase__ = image[0, -3:, -3:, -1] UpperCAmelCase__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) UpperCAmelCase__ = np.array([0.9997, 0.0003, 0.9997, 0.9997, 0.9970, 0.0024, 0.9997, 0.9971, 0.9971] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowercase_ (self : Any ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = "cpu" UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = self.pipeline_class(**__UpperCAmelCase ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase , pil_image=__UpperCAmelCase ) UpperCAmelCase__ = [ pipeline_inputs["image"], pipeline_inputs["image"], ] UpperCAmelCase__ = pipe(**__UpperCAmelCase ) UpperCAmelCase__ = output.images UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase , pil_image=__UpperCAmelCase ) UpperCAmelCase__ = [ tuple_pipeline_inputs["image"], tuple_pipeline_inputs["image"], ] UpperCAmelCase__ = pipe( **__UpperCAmelCase , return_dict=__UpperCAmelCase , )[0] UpperCAmelCase__ = image[0, -3:, -3:, -1] UpperCAmelCase__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 6_4, 6_4, 3) UpperCAmelCase__ = np.array( [ 0.9997, 0.9989, 0.0008, 0.0021, 0.9960, 0.0018, 0.0014, 0.0002, 0.9933, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = torch.device("cpu" ) class A : __UpperCAmelCase : Optional[int] = 1 UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = self.pipeline_class(**__UpperCAmelCase ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = torch.Generator(device=__UpperCAmelCase ).manual_seed(0 ) UpperCAmelCase__ = pipe.decoder.dtype UpperCAmelCase__ = 1 UpperCAmelCase__ = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) UpperCAmelCase__ = pipe.prepare_latents( __UpperCAmelCase , dtype=__UpperCAmelCase , device=__UpperCAmelCase , generator=__UpperCAmelCase , latents=__UpperCAmelCase , scheduler=DummyScheduler() ) UpperCAmelCase__ = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) UpperCAmelCase__ = pipe.prepare_latents( __UpperCAmelCase , dtype=__UpperCAmelCase , device=__UpperCAmelCase , generator=__UpperCAmelCase , latents=__UpperCAmelCase , scheduler=DummyScheduler() ) UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase , pil_image=__UpperCAmelCase ) UpperCAmelCase__ = pipe( **__UpperCAmelCase , decoder_latents=__UpperCAmelCase , super_res_latents=__UpperCAmelCase ).images UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase , pil_image=__UpperCAmelCase ) # Don't pass image, instead pass embedding UpperCAmelCase__ = pipeline_inputs.pop("image" ) UpperCAmelCase__ = pipe.image_encoder(__UpperCAmelCase ).image_embeds UpperCAmelCase__ = pipe( **__UpperCAmelCase , decoder_latents=__UpperCAmelCase , super_res_latents=__UpperCAmelCase , image_embeddings=__UpperCAmelCase , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a ).max() < 1E-4 @skip_mps def lowercase_ (self : Optional[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = torch_device == "cpu" # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor UpperCAmelCase__ = 1E-2 self._test_attention_slicing_forward_pass( test_max_difference=__UpperCAmelCase , expected_max_diff=__UpperCAmelCase ) @skip_mps def lowercase_ (self : Dict ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = torch_device == "cpu" UpperCAmelCase__ = True UpperCAmelCase__ = [ "decoder_num_inference_steps", "super_res_num_inference_steps", ] self._test_inference_batch_single_identical( test_max_difference=__UpperCAmelCase , relax_max_difference=__UpperCAmelCase , additional_params_copy_to_batched_inputs=__UpperCAmelCase , ) def lowercase_ (self : List[Any] ) -> Any: """simple docstring""" UpperCAmelCase__ = [ "decoder_num_inference_steps", "super_res_num_inference_steps", ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes UpperCAmelCase__ = [2, 3] self._test_inference_batch_consistent( batch_sizes=__UpperCAmelCase , additional_params_copy_to_batched_inputs=__UpperCAmelCase , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=__UpperCAmelCase ) @skip_mps def lowercase_ (self : str ) -> int: """simple docstring""" return super().test_dict_tuple_outputs_equivalent() @skip_mps def lowercase_ (self : Dict ) -> Tuple: """simple docstring""" return super().test_save_load_local() @skip_mps def lowercase_ (self : Optional[Any] ) -> Optional[Any]: """simple docstring""" return super().test_save_load_optional_components() @slow @require_torch_gpu class A ( unittest.TestCase ): def lowercase_ (self : int ) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ (self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png" ) UpperCAmelCase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/unclip/karlo_v1_alpha_cat_variation_fp16.npy" ) UpperCAmelCase__ = UnCLIPImageVariationPipeline.from_pretrained( "kakaobrain/karlo-v1-alpha-image-variations" , torch_dtype=torch.floataa ) UpperCAmelCase__ = pipeline.to(__UpperCAmelCase ) pipeline.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase__ = pipeline( __UpperCAmelCase , generator=__UpperCAmelCase , output_type="np" , ) UpperCAmelCase__ = output.images[0] assert image.shape == (2_5_6, 2_5_6, 3) assert_mean_pixel_difference(__UpperCAmelCase , __UpperCAmelCase , 1_5 )
143
from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class A ( yaml.SafeLoader ): def lowercase_ (self : Tuple , __UpperCAmelCase : str ) -> Tuple: """simple docstring""" UpperCAmelCase__ = [self.constructed_objects[key_node] for key_node, _ in node.value] UpperCAmelCase__ = [tuple(__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else key for key in keys] UpperCAmelCase__ = Counter(__UpperCAmelCase ) UpperCAmelCase__ = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(f"""Got duplicate yaml keys: {duplicate_keys}""" ) def lowercase_ (self : Dict , __UpperCAmelCase : Dict , __UpperCAmelCase : Any=False ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = super().construct_mapping(__UpperCAmelCase , deep=__UpperCAmelCase ) self._check_no_duplicates_on_constructed_node(__UpperCAmelCase ) return mapping def lowerCAmelCase_ ( __A ) -> Tuple[Optional[str], str]: '''simple docstring''' UpperCAmelCase__ = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: UpperCAmelCase__ = full_content[1:].index("---" ) + 1 UpperCAmelCase__ = "\n".join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(__A ) class A ( UpperCAmelCase_ ): # class attributes __UpperCAmelCase : Optional[Any] = {'train_eval_index'} # train-eval-index in the YAML metadata @classmethod def lowercase_ (cls : List[str] , __UpperCAmelCase : Path ) -> "DatasetMetadata": """simple docstring""" with open(__UpperCAmelCase , encoding="utf-8" ) as readme_file: UpperCAmelCase__ , UpperCAmelCase__ = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(__UpperCAmelCase ) else: return cls() def lowercase_ (self : int , __UpperCAmelCase : Path ) -> str: """simple docstring""" if path.exists(): with open(__UpperCAmelCase , encoding="utf-8" ) as readme_file: UpperCAmelCase__ = readme_file.read() else: UpperCAmelCase__ = None UpperCAmelCase__ = self._to_readme(__UpperCAmelCase ) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as readme_file: readme_file.write(__UpperCAmelCase ) def lowercase_ (self : str , __UpperCAmelCase : Optional[str] = None ) -> str: """simple docstring""" if readme_content is not None: UpperCAmelCase__ , UpperCAmelCase__ = _split_yaml_from_readme(__UpperCAmelCase ) UpperCAmelCase__ = "---\n" + self.to_yaml_string() + "---\n" + content else: UpperCAmelCase__ = "---\n" + self.to_yaml_string() + "---\n" return full_content @classmethod def lowercase_ (cls : Optional[int] , __UpperCAmelCase : str ) -> "DatasetMetadata": """simple docstring""" UpperCAmelCase__ = yaml.load(__UpperCAmelCase , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields UpperCAmelCase__ = { (key.replace("-" , "_" ) if key.replace("-" , "_" ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**__UpperCAmelCase ) def lowercase_ (self : List[str] ) -> str: """simple docstring""" return yaml.safe_dump( { (key.replace("_" , "-" ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=__UpperCAmelCase , allow_unicode=__UpperCAmelCase , encoding="utf-8" , ).decode("utf-8" ) UpperCamelCase__ = { 'image-classification': [], 'translation': [], 'image-segmentation': [], 'fill-mask': [], 'automatic-speech-recognition': [], 'token-classification': [], 'sentence-similarity': [], 'audio-classification': [], 'question-answering': [], 'summarization': [], 'zero-shot-classification': [], 'table-to-text': [], 'feature-extraction': [], 'other': [], 'multiple-choice': [], 'text-classification': [], 'text-to-image': [], 'text2text-generation': [], 'zero-shot-image-classification': [], 'tabular-classification': [], 'tabular-regression': [], 'image-to-image': [], 'tabular-to-text': [], 'unconditional-image-generation': [], 'text-retrieval': [], 'text-to-speech': [], 'object-detection': [], 'audio-to-audio': [], 'text-generation': [], 'conversational': [], 'table-question-answering': [], 'visual-question-answering': [], 'image-to-text': [], 'reinforcement-learning': [], 'voice-activity-detection': [], 'time-series-forecasting': [], 'document-question-answering': [], } if __name__ == "__main__": from argparse import ArgumentParser UpperCamelCase__ = ArgumentParser(usage='Validate the yaml metadata block of a README.md file.') ap.add_argument('readme_filepath') UpperCamelCase__ = ap.parse_args() UpperCamelCase__ = Path(args.readme_filepath) UpperCamelCase__ = DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
143
1
"""simple docstring""" import inspect from typing import Callable, List, Optional, Union import torch from transformers import ( CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, WhisperForConditionalGeneration, WhisperProcessor, ) from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.utils import logging _UpperCamelCase: Tuple = logging.get_logger(__name__) # pylint: disable=invalid-name class a__ ( SCREAMING_SNAKE_CASE__ ): def __init__( self : Tuple, lowerCAmelCase : WhisperForConditionalGeneration, lowerCAmelCase : WhisperProcessor, lowerCAmelCase : AutoencoderKL, lowerCAmelCase : CLIPTextModel, lowerCAmelCase : CLIPTokenizer, lowerCAmelCase : UNetaDConditionModel, lowerCAmelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler], lowerCAmelCase : StableDiffusionSafetyChecker, lowerCAmelCase : CLIPImageProcessor, ) -> Any: super().__init__() if safety_checker is None: logger.warning( f'''You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure''' ' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered' ' results in services or applications open to the public. Both the diffusers team and Hugging Face' ' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling' ' it only for use-cases that involve analyzing network behavior or auditing its results. For more' ' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .' ) self.register_modules( speech_model=lowerCAmelCase, speech_processor=lowerCAmelCase, vae=lowerCAmelCase, text_encoder=lowerCAmelCase, tokenizer=lowerCAmelCase, unet=lowerCAmelCase, scheduler=lowerCAmelCase, feature_extractor=lowerCAmelCase, ) def lowercase ( self : List[Any], lowerCAmelCase : Optional[Union[str, int]] = "auto" ) -> Any: if slice_size == "auto": lowercase : Tuple = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCAmelCase ) def lowercase ( self : Any ) -> str: self.enable_attention_slicing(lowerCAmelCase ) @torch.no_grad() def __call__( self : int, lowerCAmelCase : Any, lowerCAmelCase : str=16000, lowerCAmelCase : int = 512, lowerCAmelCase : int = 512, lowerCAmelCase : int = 50, lowerCAmelCase : float = 7.5, lowerCAmelCase : Optional[Union[str, List[str]]] = None, lowerCAmelCase : Optional[int] = 1, lowerCAmelCase : float = 0.0, lowerCAmelCase : Optional[torch.Generator] = None, lowerCAmelCase : Optional[torch.FloatTensor] = None, lowerCAmelCase : Optional[str] = "pil", lowerCAmelCase : bool = True, lowerCAmelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None, lowerCAmelCase : int = 1, **lowerCAmelCase : Union[str, Any], ) -> int: lowercase : Tuple = self.speech_processor.feature_extractor( lowerCAmelCase, return_tensors='pt', sampling_rate=lowerCAmelCase ).input_features.to(self.device ) lowercase : Any = self.speech_model.generate(lowerCAmelCase, max_length=480000 ) lowercase : Optional[Any] = self.speech_processor.tokenizer.batch_decode(lowerCAmelCase, skip_special_tokens=lowerCAmelCase, normalize=lowerCAmelCase )[ 0 ] if isinstance(lowerCAmelCase, lowerCAmelCase ): lowercase : Optional[int] = 1 elif isinstance(lowerCAmelCase, lowerCAmelCase ): lowercase : Optional[Any] = len(lowerCAmelCase ) else: raise ValueError(f'''`prompt` has to be of type `str` or `list` but is {type(lowerCAmelCase )}''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(lowerCAmelCase, lowerCAmelCase ) or callback_steps <= 0) ): raise ValueError( f'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' f''' {type(lowerCAmelCase )}.''' ) # get prompt text embeddings lowercase : int = self.tokenizer( lowerCAmelCase, padding='max_length', max_length=self.tokenizer.model_max_length, return_tensors='pt', ) lowercase : Tuple = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: lowercase : Tuple = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) lowercase : Any = text_input_ids[:, : self.tokenizer.model_max_length] lowercase : Optional[int] = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method lowercase , lowercase , lowercase : str = text_embeddings.shape lowercase : Union[str, Any] = text_embeddings.repeat(1, lowerCAmelCase, 1 ) lowercase : int = text_embeddings.view(bs_embed * num_images_per_prompt, lowerCAmelCase, -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. lowercase : Tuple = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: lowercase : List[str] if negative_prompt is None: lowercase : Dict = [''] * batch_size elif type(lowerCAmelCase ) is not type(lowerCAmelCase ): raise TypeError( f'''`negative_prompt` should be the same type to `prompt`, but got {type(lowerCAmelCase )} !=''' f''' {type(lowerCAmelCase )}.''' ) elif isinstance(lowerCAmelCase, lowerCAmelCase ): lowercase : List[str] = [negative_prompt] elif batch_size != len(lowerCAmelCase ): raise ValueError( f'''`negative_prompt`: {negative_prompt} has batch size {len(lowerCAmelCase )}, but `prompt`:''' f''' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches''' ' the batch size of `prompt`.' ) else: lowercase : Tuple = negative_prompt lowercase : Dict = text_input_ids.shape[-1] lowercase : Optional[Any] = self.tokenizer( lowerCAmelCase, padding='max_length', max_length=lowerCAmelCase, truncation=lowerCAmelCase, return_tensors='pt', ) lowercase : str = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method lowercase : int = uncond_embeddings.shape[1] lowercase : str = uncond_embeddings.repeat(1, lowerCAmelCase, 1 ) lowercase : List[Any] = uncond_embeddings.view(batch_size * num_images_per_prompt, lowerCAmelCase, -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowercase : Union[str, Any] = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. lowercase : Union[str, Any] = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) lowercase : List[str] = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps lowercase : Any = torch.randn(lowerCAmelCase, generator=lowerCAmelCase, device='cpu', dtype=lowerCAmelCase ).to( self.device ) else: lowercase : Optional[Any] = torch.randn(lowerCAmelCase, generator=lowerCAmelCase, device=self.device, dtype=lowerCAmelCase ) else: if latents.shape != latents_shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) lowercase : Union[str, Any] = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(lowerCAmelCase ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand lowercase : Dict = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler lowercase : List[str] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowercase : Dict = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowercase : Any = {} if accepts_eta: lowercase : Union[str, Any] = eta for i, t in enumerate(self.progress_bar(lowerCAmelCase ) ): # expand the latents if we are doing classifier free guidance lowercase : List[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase : int = self.scheduler.scale_model_input(lowerCAmelCase, lowerCAmelCase ) # predict the noise residual lowercase : Tuple = self.unet(lowerCAmelCase, lowerCAmelCase, encoder_hidden_states=lowerCAmelCase ).sample # perform guidance if do_classifier_free_guidance: lowercase , lowercase : Dict = noise_pred.chunk(2 ) lowercase : Union[str, Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 lowercase : str = self.scheduler.step(lowerCAmelCase, lowerCAmelCase, lowerCAmelCase, **lowerCAmelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ) lowercase : Optional[int] = 1 / 0.1_8215 * latents lowercase : str = self.vae.decode(lowerCAmelCase ).sample lowercase : Optional[Any] = (image / 2 + 0.5).clamp(0, 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowercase : Tuple = image.cpu().permute(0, 2, 3, 1 ).float().numpy() if output_type == "pil": lowercase : Tuple = self.numpy_to_pil(lowerCAmelCase ) if not return_dict: return image return StableDiffusionPipelineOutput(images=lowerCAmelCase, nsfw_content_detected=lowerCAmelCase )
255
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class a__ ( unittest.TestCase ): def lowercase ( self : Optional[Any] ) -> Any: lowercase : int = torch.nn.Linear(10, 10 ) lowercase : Optional[int] = torch.optim.SGD(model.parameters(), 0.1 ) lowercase : List[Any] = Accelerator() lowercase : Optional[Any] = accelerator.prepare(lowerCAmelCase ) try: pickle.loads(pickle.dumps(lowerCAmelCase ) ) except Exception as e: self.fail(f'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
255
1
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging _A : int = logging.get_logger(__name__) # TODO Update this _A : List[Any] = { 'facebook/esm-1b': 'https://huggingface.co/facebook/esm-1b/resolve/main/config.json', # See all ESM models at https://huggingface.co/models?filter=esm } class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): _UpperCAmelCase : str = "esm" def __init__( self : int , A : int=None , A : List[str]=None , A : Union[str, Any]=None , A : Optional[int]=7_6_8 , A : Optional[Any]=1_2 , A : Any=1_2 , A : str=3_0_7_2 , A : str=0.1 , A : Dict=0.1 , A : str=1_0_2_6 , A : Union[str, Any]=0.02 , A : Dict=1e-12 , A : str="absolute" , A : str=True , A : Union[str, Any]=None , A : Dict=False , A : List[str]=False , A : List[str]=None , A : List[str]=None , **A : int , ) ->Tuple: super().__init__(pad_token_id=A , mask_token_id=A , **A ) lowerCamelCase__ : List[Any] = vocab_size lowerCamelCase__ : Union[str, Any] = hidden_size lowerCamelCase__ : Union[str, Any] = num_hidden_layers lowerCamelCase__ : Dict = num_attention_heads lowerCamelCase__ : int = intermediate_size lowerCamelCase__ : Dict = hidden_dropout_prob lowerCamelCase__ : List[Any] = attention_probs_dropout_prob lowerCamelCase__ : Any = max_position_embeddings lowerCamelCase__ : Optional[Any] = initializer_range lowerCamelCase__ : Tuple = layer_norm_eps lowerCamelCase__ : List[Any] = position_embedding_type lowerCamelCase__ : Dict = use_cache lowerCamelCase__ : int = emb_layer_norm_before lowerCamelCase__ : Dict = token_dropout lowerCamelCase__ : Optional[int] = is_folding_model if is_folding_model: if esmfold_config is None: logger.info('''No esmfold_config supplied for folding model, using default values.''' ) lowerCamelCase__ : Tuple = EsmFoldConfig() elif isinstance(A , A ): lowerCamelCase__ : int = EsmFoldConfig(**A ) lowerCamelCase__ : Any = esmfold_config if vocab_list is None: logger.warning('''No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!''' ) lowerCamelCase__ : List[str] = get_default_vocab_list() else: lowerCamelCase__ : Any = vocab_list else: lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Union[str, Any] = None if self.esmfold_config is not None and getattr(self.esmfold_config , '''use_esm_attn_map''' , A ): raise ValueError('''The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!''' ) def __lowerCamelCase ( self : Dict ) ->Optional[int]: lowerCamelCase__ : List[Any] = super().to_dict() if isinstance(self.esmfold_config , A ): lowerCamelCase__ : List[str] = self.esmfold_config.to_dict() return output @dataclass class __SCREAMING_SNAKE_CASE : _UpperCAmelCase : str = None _UpperCAmelCase : bool = True _UpperCAmelCase : bool = False _UpperCAmelCase : bool = False _UpperCAmelCase : bool = False _UpperCAmelCase : float = 0 _UpperCAmelCase : bool = True _UpperCAmelCase : bool = False _UpperCAmelCase : int = 1_2_8 _UpperCAmelCase : "TrunkConfig" = None def __lowerCamelCase ( self : Dict ) ->Any: if self.trunk is None: lowerCamelCase__ : List[Any] = TrunkConfig() elif isinstance(self.trunk , A ): lowerCamelCase__ : List[str] = TrunkConfig(**self.trunk ) def __lowerCamelCase ( self : Dict ) ->List[Any]: lowerCamelCase__ : Optional[int] = asdict(self ) lowerCamelCase__ : Optional[Any] = self.trunk.to_dict() return output @dataclass class __SCREAMING_SNAKE_CASE : _UpperCAmelCase : int = 4_8 _UpperCAmelCase : int = 1_0_2_4 _UpperCAmelCase : int = 1_2_8 _UpperCAmelCase : int = 3_2 _UpperCAmelCase : int = 3_2 _UpperCAmelCase : int = 3_2 _UpperCAmelCase : float = 0 _UpperCAmelCase : float = 0 _UpperCAmelCase : bool = False _UpperCAmelCase : int = 4 _UpperCAmelCase : Optional[int] = 1_2_8 _UpperCAmelCase : "StructureModuleConfig" = None def __lowerCamelCase ( self : int ) ->Union[str, Any]: if self.structure_module is None: lowerCamelCase__ : Optional[Any] = StructureModuleConfig() elif isinstance(self.structure_module , A ): lowerCamelCase__ : str = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(F"`max_recycles` should be positive, got {self.max_recycles}." ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( '''`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got''' F" {self.sequence_state_dim} and {self.sequence_state_dim}." ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( '''`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got''' F" {self.pairwise_state_dim} and {self.pairwise_state_dim}." ) lowerCamelCase__ : List[str] = self.sequence_state_dim // self.sequence_head_width lowerCamelCase__ : int = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( '''`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got''' F" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}." ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( '''`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got''' F" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}." ) if self.pairwise_state_dim % 2 != 0: raise ValueError(F"`pairwise_state_dim` should be even, got {self.pairwise_state_dim}." ) if self.dropout >= 0.4: raise ValueError(F"`dropout` should not be greater than 0.4, got {self.dropout}." ) def __lowerCamelCase ( self : List[str] ) ->Dict: lowerCamelCase__ : Optional[Any] = asdict(self ) lowerCamelCase__ : Tuple = self.structure_module.to_dict() return output @dataclass class __SCREAMING_SNAKE_CASE : _UpperCAmelCase : int = 3_8_4 _UpperCAmelCase : int = 1_2_8 _UpperCAmelCase : int = 1_6 _UpperCAmelCase : int = 1_2_8 _UpperCAmelCase : int = 1_2 _UpperCAmelCase : int = 4 _UpperCAmelCase : int = 8 _UpperCAmelCase : float = 0.1 _UpperCAmelCase : int = 8 _UpperCAmelCase : int = 1 _UpperCAmelCase : int = 2 _UpperCAmelCase : int = 7 _UpperCAmelCase : int = 1_0 _UpperCAmelCase : float = 1e-8 _UpperCAmelCase : float = 1e5 def __lowerCamelCase ( self : str ) ->str: return asdict(self ) def _a ( ) -> Optional[int]: """simple docstring""" return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
265
import os from pathlib import Path def _a ( ) -> Tuple: """simple docstring""" from torch.utils.cpp_extension import load lowerCamelCase__ : List[Any] = Path(UpperCAmelCase ).resolve().parent.parent.parent / '''kernels''' / '''deformable_detr''' lowerCamelCase__ : Any = [ root / filename for filename in [ '''vision.cpp''', os.path.join('''cpu''' , '''ms_deform_attn_cpu.cpp''' ), os.path.join('''cuda''' , '''ms_deform_attn_cuda.cu''' ), ] ] load( '''MultiScaleDeformableAttention''' , UpperCAmelCase , with_cuda=UpperCAmelCase , extra_include_paths=[str(UpperCAmelCase )] , extra_cflags=['''-DWITH_CUDA=1'''] , extra_cuda_cflags=[ '''-DCUDA_HAS_FP16=1''', '''-D__CUDA_NO_HALF_OPERATORS__''', '''-D__CUDA_NO_HALF_CONVERSIONS__''', '''-D__CUDA_NO_HALF2_OPERATORS__''', ] , ) import MultiScaleDeformableAttention as MSDA return MSDA
265
1
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _A ( __magic_name__ , unittest.TestCase): SCREAMING_SNAKE_CASE : Tuple = KandinskyVaaPriorPipeline SCREAMING_SNAKE_CASE : Any = ['''prompt'''] SCREAMING_SNAKE_CASE : List[Any] = ['''prompt''', '''negative_prompt'''] SCREAMING_SNAKE_CASE : List[Any] = [ '''num_images_per_prompt''', '''generator''', '''num_inference_steps''', '''latents''', '''negative_prompt''', '''guidance_scale''', '''output_type''', '''return_dict''', ] SCREAMING_SNAKE_CASE : int = False @property def UpperCAmelCase ( self ): """simple docstring""" return 32 @property def UpperCAmelCase ( self ): """simple docstring""" return 32 @property def UpperCAmelCase ( self ): """simple docstring""" return self.time_input_dim @property def UpperCAmelCase ( self ): """simple docstring""" return self.time_input_dim * 4 @property def UpperCAmelCase ( self ): """simple docstring""" return 100 @property def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def UpperCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(lowerCAmelCase__ ) @property def UpperCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : int = { "num_attention_heads": 2, "attention_head_dim": 12, "embedding_dim": self.text_embedder_hidden_size, "num_layers": 1, } SCREAMING_SNAKE_CASE_ : Optional[Any] = PriorTransformer(**lowerCAmelCase__ ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 SCREAMING_SNAKE_CASE_ : str = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def UpperCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : List[str] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) SCREAMING_SNAKE_CASE_ : List[str] = CLIPVisionModelWithProjection(lowerCAmelCase__ ) return model @property def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = CLIPImageProcessor( crop_size=224 , do_center_crop=lowerCAmelCase__ , do_normalize=lowerCAmelCase__ , do_resize=lowerCAmelCase__ , image_mean=[0.48145466, 0.4578275, 0.40821073] , image_std=[0.26862954, 0.26130258, 0.27577711] , resample=3 , size=224 , ) return image_processor def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = self.dummy_prior SCREAMING_SNAKE_CASE_ : List[Any] = self.dummy_image_encoder SCREAMING_SNAKE_CASE_ : Any = self.dummy_text_encoder SCREAMING_SNAKE_CASE_ : int = self.dummy_tokenizer SCREAMING_SNAKE_CASE_ : Tuple = self.dummy_image_processor SCREAMING_SNAKE_CASE_ : List[Any] = UnCLIPScheduler( variance_type='fixed_small_log' , prediction_type='sample' , num_train_timesteps=1000 , clip_sample=lowerCAmelCase__ , clip_sample_range=10.0 , ) SCREAMING_SNAKE_CASE_ : Optional[int] = { "prior": prior, "image_encoder": image_encoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "scheduler": scheduler, "image_processor": image_processor, } return components def UpperCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=0 ): """simple docstring""" if str(lowerCAmelCase__ ).startswith('mps' ): SCREAMING_SNAKE_CASE_ : Any = torch.manual_seed(lowerCAmelCase__ ) else: SCREAMING_SNAKE_CASE_ : Optional[int] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[Any] = { "prompt": "horse", "generator": generator, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = "cpu" SCREAMING_SNAKE_CASE_ : Dict = self.get_dummy_components() SCREAMING_SNAKE_CASE_ : Dict = self.pipeline_class(**lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Any = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[Any] = pipe(**self.get_dummy_inputs(lowerCAmelCase__ ) ) SCREAMING_SNAKE_CASE_ : Dict = output.image_embeds SCREAMING_SNAKE_CASE_ : List[Any] = pipe( **self.get_dummy_inputs(lowerCAmelCase__ ) , return_dict=lowerCAmelCase__ , )[0] SCREAMING_SNAKE_CASE_ : Union[str, Any] = image[0, -10:] SCREAMING_SNAKE_CASE_ : str = image_from_tuple[0, -10:] assert image.shape == (1, 32) SCREAMING_SNAKE_CASE_ : int = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = torch_device == "cpu" SCREAMING_SNAKE_CASE_ : Union[str, Any] = True SCREAMING_SNAKE_CASE_ : Any = False self._test_inference_batch_single_identical( test_max_difference=lowerCAmelCase__ , relax_max_difference=lowerCAmelCase__ , test_mean_pixel_difference=lowerCAmelCase__ , ) @skip_mps def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = torch_device == "cpu" SCREAMING_SNAKE_CASE_ : Optional[Any] = False self._test_attention_slicing_forward_pass( test_max_difference=lowerCAmelCase__ , test_mean_pixel_difference=lowerCAmelCase__ , )
253
'''simple docstring''' import re import tempfile from pathlib import Path import pytest import yaml from datasets.utils.readme import ReadMe # @pytest.fixture # def example_yaml_structure(): __a = yaml.safe_load( '\\nname: ""\nallow_empty: false\nallow_empty_text: true\nsubsections:\n - name: "Dataset Card for X" # First-level markdown heading\n allow_empty: false\n allow_empty_text: true\n subsections:\n - name: "Table of Contents"\n allow_empty: false\n allow_empty_text: false\n subsections: null\n - name: "Dataset Description"\n allow_empty: false\n allow_empty_text: false\n subsections:\n - name: "Dataset Summary"\n allow_empty: false\n allow_empty_text: false\n subsections: null\n - name: "Supported Tasks and Leaderboards"\n allow_empty: true\n allow_empty_text: true\n subsections: null\n - name: Languages\n allow_empty: false\n allow_empty_text: true\n subsections: null\n' ) __a = { 'name': 'root', 'text': '', 'is_empty_text': True, 'subsections': [ { 'name': 'Dataset Card for My Dataset', 'text': '', 'is_empty_text': True, 'subsections': [ {'name': 'Table of Contents', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': []}, { 'name': 'Dataset Description', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Dataset Summary', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [], }, { 'name': 'Supported Tasks and Leaderboards', 'text': '', 'is_empty_text': True, 'subsections': [], }, {'name': 'Languages', 'text': 'Language Text', 'is_empty_text': False, 'subsections': []}, ], }, ], } ], } __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n#### Extra Ignored Subsection\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = { 'name': 'root', 'text': '', 'is_empty_text': True, 'subsections': [ { 'name': 'Dataset Card for My Dataset', 'text': '', 'is_empty_text': True, 'subsections': [ {'name': 'Table of Contents', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': []}, { 'name': 'Dataset Description', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Dataset Summary', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Extra Ignored Subsection', 'text': '', 'is_empty_text': True, 'subsections': [], } ], }, { 'name': 'Supported Tasks and Leaderboards', 'text': '', 'is_empty_text': True, 'subsections': [], }, {'name': 'Languages', 'text': 'Language Text', 'is_empty_text': False, 'subsections': []}, ], }, ], } ], } __a = '\\n---\n---\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = ( 'The following issues were found for the README at `{path}`:\n-\tEmpty YAML markers are present in the README.' ) __a = '\\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = ( 'The following issues were found for the README at `{path}`:\n-\tNo YAML markers are present in the README.' ) __a = '\\n---\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = 'The following issues were found for the README at `{path}`:\n-\tOnly the start of YAML tags present in the README.' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = 'The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Summary` but it is empty.\n-\tExpected some text in section `Dataset Summary` but it is empty (text in subsections are ignored).' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n' __a = 'The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Card for My Dataset` but it is empty.\n-\tSection `Dataset Card for My Dataset` expected the following subsections: `Table of Contents`, `Dataset Description`. Found \'None\'.' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Languages\nLanguage Text\n' __a = 'The following issues were found for the README at `{path}`:\n-\tSection `Dataset Description` is missing subsection: `Supported Tasks and Leaderboards`.' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\n' __a = 'The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Languages` but it is empty.' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = 'The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n# Dataset Card My Dataset\n' __a = 'The following issues were found for the README at `{path}`:\n-\tThe README has several first-level headings: `Dataset Card for My Dataset`, `Dataset Card My Dataset`. Only one heading is expected. Skipping further validation for this README.' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = 'The following issues were found for the README at `{path}`:\n-\tNo first-level heading starting with `Dataset Card for` found in README. Skipping further validation for this README.' __a = '' __a = 'The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.\n-\tNo YAML markers are present in the README.' __a = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' __a = 'The following issues were found while parsing the README at `{path}`:\n-\tMultiple sections with the same heading `Dataset Card for My Dataset` have been found. Please keep only one of these sections.' @pytest.mark.parametrize( "readme_md, expected_dict", [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ], ) def __UpperCAmelCase ( a_: Any, a_: Any ): assert ReadMe.from_string(a_, a_ ).to_dict() == expected_dict @pytest.mark.parametrize( "readme_md, expected_error", [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ], ) def __UpperCAmelCase ( a_: Optional[int], a_: int ): with pytest.raises(a_, match=re.escape(expected_error.format(path="root" ) ) ): _UpperCAmelCase : Union[str, Any] = ReadMe.from_string(a_, a_ ) readme.validate() @pytest.mark.parametrize( "readme_md, expected_error", [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ], ) def __UpperCAmelCase ( a_: str, a_: Optional[Any] ): with pytest.raises(a_, match=re.escape(expected_error.format(path="root" ) ) ): ReadMe.from_string(a_, a_ ) @pytest.mark.parametrize( "readme_md,", [ (README_MULTIPLE_SAME_HEADING_1), ], ) def __UpperCAmelCase ( a_: str ): ReadMe.from_string(a_, a_, suppress_parsing_errors=a_ ) @pytest.mark.parametrize( "readme_md, expected_dict", [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ], ) def __UpperCAmelCase ( a_: str, a_: Union[str, Any] ): with tempfile.TemporaryDirectory() as tmp_dir: _UpperCAmelCase : Optional[Any] = Path(a_ ) / "README.md" with open(a_, "w+" ) as readme_file: readme_file.write(a_ ) _UpperCAmelCase : Tuple = ReadMe.from_readme(a_, a_ ).to_dict() assert out["name"] == path assert out["text"] == "" assert out["is_empty_text"] assert out["subsections"] == expected_dict["subsections"] @pytest.mark.parametrize( "readme_md, expected_error", [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ], ) def __UpperCAmelCase ( a_: List[Any], a_: str ): with tempfile.TemporaryDirectory() as tmp_dir: _UpperCAmelCase : Optional[Any] = Path(a_ ) / "README.md" with open(a_, "w+" ) as readme_file: readme_file.write(a_ ) _UpperCAmelCase : Any = expected_error.format(path=a_ ) with pytest.raises(a_, match=re.escape(a_ ) ): _UpperCAmelCase : str = ReadMe.from_readme(a_, a_ ) readme.validate() @pytest.mark.parametrize( "readme_md, expected_error", [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ], ) def __UpperCAmelCase ( a_: Tuple, a_: Optional[int] ): with tempfile.TemporaryDirectory() as tmp_dir: _UpperCAmelCase : List[str] = Path(a_ ) / "README.md" with open(a_, "w+" ) as readme_file: readme_file.write(a_ ) _UpperCAmelCase : Any = expected_error.format(path=a_ ) with pytest.raises(a_, match=re.escape(a_ ) ): ReadMe.from_readme(a_, a_ ) @pytest.mark.parametrize( "readme_md,", [ (README_MULTIPLE_SAME_HEADING_1), ], ) def __UpperCAmelCase ( a_: Any ): with tempfile.TemporaryDirectory() as tmp_dir: _UpperCAmelCase : List[Any] = Path(a_ ) / "README.md" with open(a_, "w+" ) as readme_file: readme_file.write(a_ ) ReadMe.from_readme(a_, a_, suppress_parsing_errors=a_ )
145
0
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def UpperCAmelCase_ ( __lowercase : str=None ) -> Optional[int]: '''simple docstring''' if subparsers is not None: _UpperCAmelCase = subparsers.add_parser("env" ) else: _UpperCAmelCase = argparse.ArgumentParser("Accelerate env command" ) parser.add_argument( "--config_file" , default=__lowercase , help="The config file to use for the default values in the launching script." ) if subparsers is not None: parser.set_defaults(func=__lowercase ) return parser def UpperCAmelCase_ ( __lowercase : Optional[Any] ) -> Optional[Any]: '''simple docstring''' _UpperCAmelCase = torch.__version__ _UpperCAmelCase = torch.cuda.is_available() _UpperCAmelCase = is_xpu_available() _UpperCAmelCase = is_npu_available() _UpperCAmelCase = "Not found" # Get the default from the config file. if args.config_file is not None or os.path.isfile(__lowercase ): _UpperCAmelCase = load_config_from_file(args.config_file ).to_dict() _UpperCAmelCase = { "`Accelerate` version": version, "Platform": platform.platform(), "Python version": platform.python_version(), "Numpy version": np.__version__, "PyTorch version (GPU?)": f'{pt_version} ({pt_cuda_available})', "PyTorch XPU available": str(__lowercase ), "PyTorch NPU available": str(__lowercase ), "System RAM": f'{psutil.virtual_memory().total / 1024 ** 3:.2f} GB', } if pt_cuda_available: _UpperCAmelCase = torch.cuda.get_device_name() print("\nCopy-and-paste the text below in your GitHub issue\n" ) print("\n".join([f'- {prop}: {val}' for prop, val in info.items()] ) ) print("- `Accelerate` default config:" if args.config_file is None else "- `Accelerate` config passed:" ) _UpperCAmelCase = ( "\n".join([f'\t- {prop}: {val}' for prop, val in accelerate_config.items()] ) if isinstance(__lowercase , __lowercase ) else f'\t{accelerate_config}' ) print(__lowercase ) _UpperCAmelCase = accelerate_config return info def UpperCAmelCase_ ( ) -> int: '''simple docstring''' _UpperCAmelCase = env_command_parser() _UpperCAmelCase = parser.parse_args() env_command(__lowercase ) return 0 if __name__ == "__main__": raise SystemExit(main())
156
'''simple docstring''' from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING __SCREAMING_SNAKE_CASE :List[Any] = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase_ ) class A_ ( lowerCAmelCase_ ): def __init__( self : List[str] , *snake_case_ : Dict , **snake_case_ : Dict ): super().__init__(*snake_case_ , **snake_case_ ) requires_backends(self , "vision" ) self.check_model_type(snake_case_ ) def __call__( self : Optional[Any] , snake_case_ : Union[str, List[str], "Image.Image", List["Image.Image"]] , **snake_case_ : Optional[int] ): return super().__call__(snake_case_ , **snake_case_ ) def lowercase ( self : Union[str, Any] , **snake_case_ : Union[str, Any] ): return {}, {}, {} def lowercase ( self : Dict , snake_case_ : Optional[int] ): _UpperCAmelCase = load_image(snake_case_ ) _UpperCAmelCase = image.size _UpperCAmelCase = self.image_processor(images=snake_case_ , return_tensors=self.framework ) return model_inputs def lowercase ( self : Optional[int] , snake_case_ : List[Any] ): _UpperCAmelCase = self.model(**snake_case_ ) return model_outputs def lowercase ( self : List[str] , snake_case_ : Dict ): _UpperCAmelCase = model_outputs.predicted_depth _UpperCAmelCase = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) , size=self.image_size[::-1] , mode="bicubic" , align_corners=snake_case_ ) _UpperCAmelCase = prediction.squeeze().cpu().numpy() _UpperCAmelCase = (output * 2_5_5 / np.max(snake_case_ )).astype("uint8" ) _UpperCAmelCase = Image.fromarray(snake_case_ ) _UpperCAmelCase = {} _UpperCAmelCase = predicted_depth _UpperCAmelCase = depth return output_dict
156
1
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _snake_case ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' A: Any = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) A: Dict = get_activation('''gelu''' ) self.assertTrue(torch.allclose(gelu_python(SCREAMING_SNAKE_CASE_ ) , torch_builtin(SCREAMING_SNAKE_CASE_ ) ) ) self.assertFalse(torch.allclose(gelu_python(SCREAMING_SNAKE_CASE_ ) , gelu_new(SCREAMING_SNAKE_CASE_ ) ) ) def _snake_case ( self : Any ) -> List[Any]: '''simple docstring''' A: Optional[Any] = torch.tensor([-1_00, -1, -0.1, 0, 0.1, 1.0, 1_00] ) A: List[Any] = get_activation('''gelu''' ) A: Optional[int] = get_activation('''gelu_10''' ) A: Any = torch_builtin(SCREAMING_SNAKE_CASE_ ) A: Optional[int] = geluaa(SCREAMING_SNAKE_CASE_ ) A: Union[str, Any] = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(SCREAMING_SNAKE_CASE_ ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def _snake_case ( self : Optional[Any] ) -> Tuple: '''simple docstring''' get_activation('''gelu''' ) get_activation('''gelu_10''' ) get_activation('''gelu_fast''' ) get_activation('''gelu_new''' ) get_activation('''gelu_python''' ) get_activation('''gelu_pytorch_tanh''' ) get_activation('''linear''' ) get_activation('''mish''' ) get_activation('''quick_gelu''' ) get_activation('''relu''' ) get_activation('''sigmoid''' ) get_activation('''silu''' ) get_activation('''swish''' ) get_activation('''tanh''' ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): get_activation('''bogus''' ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): get_activation(SCREAMING_SNAKE_CASE_ ) def _snake_case ( self : Tuple ) -> int: '''simple docstring''' A: int = get_activation('''gelu''' ) A: Optional[Any] = 1 A: List[Any] = get_activation('''gelu''' ) self.assertEqual(acta.a , 1 ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): A: Tuple = acta.a
319
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE_ : List[str] , **SCREAMING_SNAKE_CASE_ : Union[str, Any] ) -> None: '''simple docstring''' warnings.warn( '''The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use BeitImageProcessor instead.''' , SCREAMING_SNAKE_CASE_ , ) super().__init__(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ )
319
1
import logging import os import quant_trainer import torch from torch.utils.data import DataLoader from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput UpperCamelCase = logging.getLogger(__name__) if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class _lowerCamelCase ( UpperCamelCase ): """simple docstring""" def __init__( self , *_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE )->Dict: '''simple docstring''' super().__init__(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) A_ : Dict = eval_examples A_ : Optional[Any] = post_process_function A_ : Dict = quant_trainer_args A_ : List[Any] = 128 # default number of calibration samples def _snake_case ( self , _SCREAMING_SNAKE_CASE=None )->Dict: '''simple docstring''' if calib_dataset is None and self.calib_dataset is None: raise ValueError('''Trainer: calibration requires an calib_dataset.''' ) A_ : Union[str, Any] = calib_dataset if calib_dataset is not None else self.calib_dataset A_ : Optional[int] = self._remove_unused_columns(_SCREAMING_SNAKE_CASE , description='''Calibration''' ) return DataLoader( _SCREAMING_SNAKE_CASE , batch_size=self.args.eval_batch_size , collate_fn=self.data_collator , drop_last=self.args.dataloader_drop_last , num_workers=self.args.dataloader_num_workers , pin_memory=self.args.dataloader_pin_memory , shuffle=_SCREAMING_SNAKE_CASE , ) def _snake_case ( self , _SCREAMING_SNAKE_CASE=None )->Tuple: '''simple docstring''' A_ : Tuple = self.train_dataset if calib_dataset is None else calib_dataset A_ : Optional[Any] = self.get_calib_dataloader(_SCREAMING_SNAKE_CASE ) A_ : Any = self.model quant_trainer.configure_model(_SCREAMING_SNAKE_CASE , self.quant_trainer_args , calib=_SCREAMING_SNAKE_CASE ) model.eval() quant_trainer.enable_calibration(_SCREAMING_SNAKE_CASE ) logger.info('''***** Running calibration *****''' ) logger.info(F''' Num examples = {self.calib_num}''' ) logger.info(F''' Batch size = {calib_dataloader.batch_size}''' ) for step, inputs in enumerate(_SCREAMING_SNAKE_CASE ): # Prediction step A_ , A_ , A_ : Union[str, Any] = self.prediction_step(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , prediction_loss_only=_SCREAMING_SNAKE_CASE ) if (step + 1) * calib_dataloader.batch_size >= self.calib_num: break quant_trainer.finish_calibration(_SCREAMING_SNAKE_CASE , self.quant_trainer_args ) A_ : Tuple = model def _snake_case ( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = "eval" )->Optional[int]: '''simple docstring''' A_ : Optional[Any] = self.eval_dataset if eval_dataset is None else eval_dataset A_ : List[str] = self.get_eval_dataloader(_SCREAMING_SNAKE_CASE ) A_ : Optional[Any] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. A_ : List[Any] = self.compute_metrics A_ : Optional[Any] = None A_ : int = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: A_ : Any = eval_loop( _SCREAMING_SNAKE_CASE , description='''Evaluation''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_SCREAMING_SNAKE_CASE , ) finally: A_ : str = compute_metrics if self.post_process_function is not None and self.compute_metrics is not None: A_ : Tuple = self.post_process_function(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , output.predictions ) A_ : List[str] = self.compute_metrics(_SCREAMING_SNAKE_CASE ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): A_ : str = metrics.pop(_SCREAMING_SNAKE_CASE ) self.log(_SCREAMING_SNAKE_CASE ) else: A_ : List[str] = {} if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) A_ : str = self.callback_handler.on_evaluate(self.args , self.state , self.control , _SCREAMING_SNAKE_CASE ) return metrics def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = "test" )->Tuple: '''simple docstring''' A_ : Any = self.get_test_dataloader(_SCREAMING_SNAKE_CASE ) # Temporarily disable metric computation, we will do it in the loop here. A_ : Any = self.compute_metrics A_ : Union[str, Any] = None A_ : Optional[Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: A_ : Optional[int] = eval_loop( _SCREAMING_SNAKE_CASE , description='''Prediction''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_SCREAMING_SNAKE_CASE , ) finally: A_ : Union[str, Any] = compute_metrics if self.post_process_function is None or self.compute_metrics is None: return output A_ : Tuple = self.post_process_function(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , output.predictions , '''predict''' ) A_ : str = self.compute_metrics(_SCREAMING_SNAKE_CASE ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): A_ : Tuple = metrics.pop(_SCREAMING_SNAKE_CASE ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=_SCREAMING_SNAKE_CASE ) def _snake_case ( self , _SCREAMING_SNAKE_CASE="./" )->Tuple: '''simple docstring''' A_ : Any = self.eval_dataset A_ : Any = self.get_eval_dataloader(_SCREAMING_SNAKE_CASE ) A_ : int = next(iter(_SCREAMING_SNAKE_CASE ) ) # saving device - to make it consistent A_ : Any = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) # convert to tuple A_ : Tuple = tuple(v.to(_SCREAMING_SNAKE_CASE ) for k, v in batch.items() ) logger.info('''Converting model to be onnx compatible''' ) from pytorch_quantization.nn import TensorQuantizer A_ : Optional[Any] = True A_ : Tuple = self.model.to(_SCREAMING_SNAKE_CASE ) model.eval() model.float() A_ : Optional[Any] = model.module if hasattr(_SCREAMING_SNAKE_CASE , '''module''' ) else model quant_trainer.configure_model(_SCREAMING_SNAKE_CASE , self.quant_trainer_args ) A_ : Optional[int] = os.path.join(_SCREAMING_SNAKE_CASE , '''model.onnx''' ) logger.info(F'''exporting model to {output_model_file}''' ) A_ : int = {0: '''batch_size''', 1: '''seq_len'''} torch.onnx.export( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , export_params=_SCREAMING_SNAKE_CASE , opset_version=13 , do_constant_folding=_SCREAMING_SNAKE_CASE , input_names=['''input_ids''', '''attention_mask''', '''token_type_ids'''] , output_names=['''output_start_logits''', '''output_end_logits'''] , dynamic_axes={ '''input_ids''': axes, '''attention_mask''': axes, '''token_type_ids''': axes, '''output_start_logits''': axes, '''output_end_logits''': axes, } , verbose=_SCREAMING_SNAKE_CASE , ) logger.info('''onnx export finished''' )
65
import math import random def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value UpperCamelCase = 0.02 def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A_ : str = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(SCREAMING_SNAKE_CASE ): # Forward propagation A_ : Optional[Any] = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? A_ : Any = (expected / 100) - layer_a # Error delta A_ : List[str] = layer_1_error * sigmoid_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = int(input("""Expected value: """)) UpperCamelCase = int(input("""Number of propagations: """)) print(forward_propagation(expected, number_propagations))
65
1
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class UpperCamelCase__ ( __UpperCAmelCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = ["""image_processor""", """tokenizer"""] _SCREAMING_SNAKE_CASE = """LayoutLMv2ImageProcessor""" _SCREAMING_SNAKE_CASE = ("""LayoutXLMTokenizer""", """LayoutXLMTokenizerFast""") def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE_ : List[Any]=None , SCREAMING_SNAKE_CASE_ : Tuple=None , **SCREAMING_SNAKE_CASE_ : Union[str, Any] ): if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , UpperCamelCase__ , ) lowerCAmelCase_ : str = kwargs.pop('feature_extractor' ) lowerCAmelCase_ : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(UpperCamelCase__ , UpperCamelCase__ ) def __call__( self : str , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : List[Any] = None , SCREAMING_SNAKE_CASE_ : List[str] = None , SCREAMING_SNAKE_CASE_ : Any = None , SCREAMING_SNAKE_CASE_ : List[str] = None , SCREAMING_SNAKE_CASE_ : Dict = True , SCREAMING_SNAKE_CASE_ : List[str] = False , SCREAMING_SNAKE_CASE_ : List[str] = None , SCREAMING_SNAKE_CASE_ : Optional[int] = None , SCREAMING_SNAKE_CASE_ : Dict = 0 , SCREAMING_SNAKE_CASE_ : Optional[int] = None , SCREAMING_SNAKE_CASE_ : Tuple = None , SCREAMING_SNAKE_CASE_ : Tuple = None , SCREAMING_SNAKE_CASE_ : Optional[Any] = False , SCREAMING_SNAKE_CASE_ : str = False , SCREAMING_SNAKE_CASE_ : str = False , SCREAMING_SNAKE_CASE_ : Any = False , SCREAMING_SNAKE_CASE_ : Dict = True , SCREAMING_SNAKE_CASE_ : int = None , **SCREAMING_SNAKE_CASE_ : Union[str, Any] , ): if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes ' 'if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError('You cannot return overflowing tokens without returning the offsets mapping.' ) # first, apply the image processor lowerCAmelCase_ : Tuple = self.image_processor(images=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCAmelCase_ : Union[str, Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase_ : Optional[Any] = features['words'] lowerCAmelCase_ : Optional[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , padding=UpperCamelCase__ , truncation=UpperCamelCase__ , max_length=UpperCamelCase__ , stride=UpperCamelCase__ , pad_to_multiple_of=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ , return_attention_mask=UpperCamelCase__ , return_overflowing_tokens=UpperCamelCase__ , return_special_tokens_mask=UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , return_length=UpperCamelCase__ , verbose=UpperCamelCase__ , return_tensors=UpperCamelCase__ , **UpperCamelCase__ , ) # add pixel values lowerCAmelCase_ : List[Any] = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase_ : Dict = self.get_overflowing_images(UpperCamelCase__ , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase_ : Union[str, Any] = images return encoded_inputs def SCREAMING_SNAKE_CASE__ ( self : Dict , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : Tuple ): lowerCAmelCase_ : Any = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(UpperCamelCase__ ) != len(UpperCamelCase__ ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' F" {len(UpperCamelCase__ )} and {len(UpperCamelCase__ )}" ) return images_with_overflow def SCREAMING_SNAKE_CASE__ ( self : List[Any] , *SCREAMING_SNAKE_CASE_ : Tuple , **SCREAMING_SNAKE_CASE_ : Tuple ): return self.tokenizer.batch_decode(*UpperCamelCase__ , **UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , *SCREAMING_SNAKE_CASE_ : List[Any] , **SCREAMING_SNAKE_CASE_ : Tuple ): return self.tokenizer.decode(*UpperCamelCase__ , **UpperCamelCase__ ) @property def SCREAMING_SNAKE_CASE__ ( self : int ): return ["input_ids", "bbox", "attention_mask", "image"] @property def SCREAMING_SNAKE_CASE__ ( self : Dict ): warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , UpperCamelCase__ , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , UpperCamelCase__ , ) return self.image_processor
224
import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def __UpperCamelCase ( _A ): lowerCAmelCase_ = [ '''decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(_A , _A ) def __UpperCamelCase ( _A ): lowerCAmelCase_ , lowerCAmelCase_ = emb.weight.shape lowerCAmelCase_ = nn.Linear(_A , _A , bias=_A ) lowerCAmelCase_ = emb.weight.data return lin_layer def __UpperCamelCase ( _A ): lowerCAmelCase_ = torch.load(_A , map_location='''cpu''' ) lowerCAmelCase_ = Namespace(**checkpoint['''cfg''']['''model'''] ) lowerCAmelCase_ = checkpoint['''model'''] remove_ignore_keys_(_A ) lowerCAmelCase_ = state_dict['''decoder.embed_tokens.weight'''].shape[0] lowerCAmelCase_ = {key.replace('''decoder''' , '''model''' ): val for key, val in state_dict.items()} lowerCAmelCase_ = XGLMConfig( vocab_size=_A , max_position_embeddings=args.max_target_positions , num_layers=args.decoder_layers , attention_heads=args.decoder_attention_heads , ffn_dim=args.decoder_ffn_embed_dim , d_model=args.decoder_embed_dim , layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''gelu''' , scale_embedding=not args.no_scale_embedding , tie_word_embeddings=args.share_decoder_input_output_embed , ) lowerCAmelCase_ = XGLMForCausalLM(_A ) lowerCAmelCase_ = model.load_state_dict(_A , strict=_A ) print(_A ) lowerCAmelCase_ = make_linear_from_emb(model.model.embed_tokens ) return model if __name__ == "__main__": _A = argparse.ArgumentParser() # Required parameters parser.add_argument('''fairseq_path''', type=str, help='''path to a model.pt on local filesystem.''') parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') _A = parser.parse_args() _A = convert_fairseq_xglm_checkpoint_from_disk(args.fairseq_path) model.save_pretrained(args.pytorch_dump_folder_path)
278
0
"""simple docstring""" from collections.abc import Generator def _snake_case ( ) -> Generator[int, None, None]: '''simple docstring''' lowerCAmelCase_ :List[Any] = 0, 1 while True: lowerCAmelCase_ :List[str] = b, a + b yield b def _snake_case ( lowercase__ : int = 1_0_0_0 ) -> int: '''simple docstring''' lowerCAmelCase_ :Dict = 1 lowerCAmelCase_ :int = fibonacci_generator() while len(str(next(lowercase__ ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
353
"""simple docstring""" def _snake_case ( lowercase__ : list , lowercase__ : list , lowercase__ : int , lowercase__ : int , lowercase__ : int ) -> int: '''simple docstring''' if index == number_of_items: return 0 lowerCAmelCase_ :Any = 0 lowerCAmelCase_ :str = 0 lowerCAmelCase_ :Dict = knapsack(lowercase__ , lowercase__ , lowercase__ , lowercase__ , index + 1 ) if weights[index] <= max_weight: lowerCAmelCase_ :str = values[index] + knapsack( lowercase__ , lowercase__ , lowercase__ , max_weight - weights[index] , index + 1 ) return max(lowercase__ , lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
1
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = """audio-spectrogram-transformer""" def __init__( self : List[Any] , lowerCamelCase_ : str=768 , lowerCamelCase_ : Dict=12 , lowerCamelCase_ : str=12 , lowerCamelCase_ : int=3072 , lowerCamelCase_ : int="gelu" , lowerCamelCase_ : Any=0.0 , lowerCamelCase_ : Dict=0.0 , lowerCamelCase_ : int=0.0_2 , lowerCamelCase_ : Tuple=1E-12 , lowerCamelCase_ : Dict=16 , lowerCamelCase_ : Tuple=True , lowerCamelCase_ : List[str]=10 , lowerCamelCase_ : Tuple=10 , lowerCamelCase_ : Dict=1024 , lowerCamelCase_ : Any=128 , **lowerCamelCase_ : str , ): """simple docstring""" super().__init__(**lowerCamelCase_ ) UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = intermediate_size UpperCamelCase = hidden_act UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = patch_size UpperCamelCase = qkv_bias UpperCamelCase = frequency_stride UpperCamelCase = time_stride UpperCamelCase = max_length UpperCamelCase = num_mel_bins
343
import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand _SCREAMING_SNAKE_CASE = ( """4S 3H 2C 7S 5H""", """9D 8H 2C 6S 7H""", """2D 6D 9D TH 7D""", """TC 8C 2S JH 6C""", """JH 8S TH AH QH""", """TS KS 5S 9S AC""", """KD 6S 9D TH AD""", """KS 8D 4D 9S 4S""", # pair """8C 4S KH JS 4D""", # pair """QH 8H KD JH 8S""", # pair """KC 4H KS 2H 8D""", # pair """KD 4S KC 3H 8S""", # pair """AH 8S AS KC JH""", # pair """3H 4C 4H 3S 2H""", # 2 pairs """5S 5D 2C KH KH""", # 2 pairs """3C KH 5D 5S KH""", # 2 pairs """AS 3C KH AD KH""", # 2 pairs """7C 7S 3S 7H 5S""", # 3 of a kind """7C 7S KH 2H 7H""", # 3 of a kind """AC KH QH AH AS""", # 3 of a kind """2H 4D 3C AS 5S""", # straight (low ace) """3C 5C 4C 2C 6H""", # straight """6S 8S 7S 5H 9H""", # straight """JS QS 9H TS KH""", # straight """QC KH TS JS AH""", # straight (high ace) """8C 9C 5C 3C TC""", # flush """3S 8S 9S 5S KS""", # flush """4C 5C 9C 8C KC""", # flush """JH 8H AH KH QH""", # flush """3D 2H 3H 2C 2D""", # full house """2H 2C 3S 3H 3D""", # full house """KH KC 3S 3H 3D""", # full house """JC 6H JS JD JH""", # 4 of a kind """JC 7H JS JD JH""", # 4 of a kind """JC KH JS JD JH""", # 4 of a kind """2S AS 4S 5S 3S""", # straight flush (low ace) """2D 6D 3D 4D 5D""", # straight flush """5C 6C 3C 7C 4C""", # straight flush """JH 9H TH KH QH""", # straight flush """JH AH TH KH QH""", # royal flush (high ace straight flush) ) _SCREAMING_SNAKE_CASE = ( ("""2H 3H 4H 5H 6H""", """KS AS TS QS JS""", """Loss"""), ("""2H 3H 4H 5H 6H""", """AS AD AC AH JD""", """Win"""), ("""AS AH 2H AD AC""", """JS JD JC JH 3D""", """Win"""), ("""2S AH 2H AS AC""", """JS JD JC JH AD""", """Loss"""), ("""2S AH 2H AS AC""", """2H 3H 5H 6H 7H""", """Win"""), ("""AS 3S 4S 8S 2S""", """2H 3H 5H 6H 7H""", """Win"""), ("""2H 3H 5H 6H 7H""", """2S 3H 4H 5S 6C""", """Win"""), ("""2S 3H 4H 5S 6C""", """3D 4C 5H 6H 2S""", """Tie"""), ("""2S 3H 4H 5S 6C""", """AH AC 5H 6H AS""", """Win"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H AS""", """Loss"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H 7S""", """Win"""), ("""6S AD 7H 4S AS""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S AH 4H 5S KC""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S 3H 6H 7S 9C""", """7H 3C TH 6H 9S""", """Loss"""), ("""4S 5H 6H TS AC""", """3S 5H 6H TS AC""", """Win"""), ("""2S AH 4H 5S 6C""", """AD 4C 5H 6H 2C""", """Tie"""), ("""AS AH 3H AD AC""", """AS AH 2H AD AC""", """Win"""), ("""AH AC 5H 5C QS""", """AH AC 5H 5C KS""", """Loss"""), ("""AH AC 5H 5C QS""", """KH KC 5H 5C QS""", """Win"""), ("""7C 7S KH 2H 7H""", """3C 3S AH 2H 3H""", """Win"""), ("""3C 3S AH 2H 3H""", """7C 7S KH 2H 7H""", """Loss"""), ("""6H 5H 4H 3H 2H""", """5H 4H 3H 2H AH""", """Win"""), ("""5H 4H 3H 2H AH""", """5H 4H 3H 2H AH""", """Tie"""), ("""5H 4H 3H 2H AH""", """6H 5H 4H 3H 2H""", """Loss"""), ("""AH AD KS KC AC""", """AH KD KH AC KC""", """Win"""), ("""2H 4D 3C AS 5S""", """2H 4D 3C 6S 5S""", """Loss"""), ("""2H 3S 3C 3H 2S""", """3S 3C 2S 2H 2D""", """Win"""), ("""4D 6D 5D 2D JH""", """3S 8S 3H TC KH""", """Loss"""), ("""4S 6C 8S 3S 7S""", """AD KS 2D 7D 7C""", """Loss"""), ("""6S 4C 7H 8C 3H""", """5H JC AH 9D 9C""", """Loss"""), ("""9D 9H JH TC QH""", """3C 2S JS 5C 7H""", """Win"""), ("""2H TC 8S AD 9S""", """4H TS 7H 2C 5C""", """Win"""), ("""9D 3S 2C 7S 7C""", """JC TD 3C TC 9H""", """Loss"""), ) _SCREAMING_SNAKE_CASE = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", True), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", False), ("""AS 3S 4S 8S 2S""", True), ) _SCREAMING_SNAKE_CASE = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", False), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", True), ) _SCREAMING_SNAKE_CASE = ( ("""2H 4D 3C AS 5S""", True, [5, 4, 3, 2, 1_4]), ("""2H 5D 3C AS 5S""", False, [1_4, 5, 5, 3, 2]), ("""JH QD KC AS TS""", False, [1_4, 1_3, 1_2, 1_1, 1_0]), ("""9D 3S 2C 7S 7C""", False, [9, 7, 7, 3, 2]), ) _SCREAMING_SNAKE_CASE = ( ("""JH AH TH KH QH""", 0), ("""JH 9H TH KH QH""", 0), ("""JC KH JS JD JH""", 7), ("""KH KC 3S 3H 3D""", 6), ("""8C 9C 5C 3C TC""", 0), ("""JS QS 9H TS KH""", 0), ("""7C 7S KH 2H 7H""", 3), ("""3C KH 5D 5S KH""", 2), ("""QH 8H KD JH 8S""", 1), ("""2D 6D 9D TH 7D""", 0), ) _SCREAMING_SNAKE_CASE = ( ("""JH AH TH KH QH""", 2_3), ("""JH 9H TH KH QH""", 2_2), ("""JC KH JS JD JH""", 2_1), ("""KH KC 3S 3H 3D""", 2_0), ("""8C 9C 5C 3C TC""", 1_9), ("""JS QS 9H TS KH""", 1_8), ("""7C 7S KH 2H 7H""", 1_7), ("""3C KH 5D 5S KH""", 1_6), ("""QH 8H KD JH 8S""", 1_5), ("""2D 6D 9D TH 7D""", 1_4), ) def lowercase( ) -> Dict: '''simple docstring''' UpperCamelCase , UpperCamelCase = randrange(len(UpperCamelCase_ ) ), randrange(len(UpperCamelCase_ ) ) UpperCamelCase = ["""Loss""", """Tie""", """Win"""][(play >= oppo) + (play > oppo)] UpperCamelCase , UpperCamelCase = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def lowercase( UpperCamelCase_ = 100 ) -> List[Any]: '''simple docstring''' return (generate_random_hand() for _ in range(UpperCamelCase_ )) @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Optional[int]: '''simple docstring''' assert PokerHand(UpperCamelCase_ )._is_flush() == expected @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Tuple: '''simple docstring''' assert PokerHand(UpperCamelCase_ )._is_straight() == expected @pytest.mark.parametrize("""hand, expected, card_values""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Dict: '''simple docstring''' UpperCamelCase = PokerHand(UpperCamelCase_ ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Optional[int]: '''simple docstring''' assert PokerHand(UpperCamelCase_ )._is_same_kind() == expected @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Any: '''simple docstring''' assert PokerHand(UpperCamelCase_ )._hand_type == expected @pytest.mark.parametrize("""hand, other, expected""" , UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[Any]: '''simple docstring''' assert PokerHand(UpperCamelCase_ ).compare_with(PokerHand(UpperCamelCase_ ) ) == expected @pytest.mark.parametrize("""hand, other, expected""" , generate_random_hands() ) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> int: '''simple docstring''' assert PokerHand(UpperCamelCase_ ).compare_with(PokerHand(UpperCamelCase_ ) ) == expected def lowercase( ) -> Dict: '''simple docstring''' UpperCamelCase = [PokerHand(UpperCamelCase_ ) for hand in SORTED_HANDS] UpperCamelCase = poker_hands.copy() shuffle(UpperCamelCase_ ) UpperCamelCase = chain(sorted(UpperCamelCase_ ) ) for index, hand in enumerate(UpperCamelCase_ ): assert hand == poker_hands[index] def lowercase( ) -> Union[str, Any]: '''simple docstring''' # Test that five high straights are compared correctly. UpperCamelCase = [PokerHand("""2D AC 3H 4H 5S""" ), PokerHand("""2S 3H 4H 5S 6C""" )] pokerhands.sort(reverse=UpperCamelCase_ ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def lowercase( ) -> str: '''simple docstring''' # Multiple calls to five_high_straight function should still return True # and shouldn't mutate the list in every call other than the first. UpperCamelCase = PokerHand("""2C 4S AS 3D 5C""" ) UpperCamelCase = True UpperCamelCase = [5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def lowercase( ) -> int: '''simple docstring''' # Problem number 54 from Project Euler # Testing from poker_hands.txt file UpperCamelCase = 0 UpperCamelCase = os.path.abspath(os.path.dirname(UpperCamelCase_ ) ) UpperCamelCase = os.path.join(UpperCamelCase_ , """poker_hands.txt""" ) with open(UpperCamelCase_ ) as file_hand: for line in file_hand: UpperCamelCase = line[:14].strip() UpperCamelCase = line[15:].strip() UpperCamelCase , UpperCamelCase = PokerHand(UpperCamelCase_ ), PokerHand(UpperCamelCase_ ) UpperCamelCase = player.compare_with(UpperCamelCase_ ) if output == "Win": answer += 1 assert answer == 376
343
1
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __snake_case ( __UpperCamelCase : Union[str, Any] ): """simple docstring""" A_ = args.pruning_method A_ = args.threshold A_ = args.model_name_or_path.rstrip("/" ) A_ = args.target_model_path print(f'''Load fine-pruned model from {model_name_or_path}''' ) A_ = torch.load(os.path.join(lowerCamelCase_ ,"pytorch_model.bin" ) ) A_ = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: A_ = tensor print(f'''Copied layer {name}''' ) elif "classifier" in name or "qa_output" in name: A_ = tensor print(f'''Copied layer {name}''' ) elif "bias" in name: A_ = tensor print(f'''Copied layer {name}''' ) else: if pruning_method == "magnitude": A_ = MagnitudeBinarizer.apply(inputs=lowerCamelCase_ ,threshold=lowerCamelCase_ ) A_ = tensor * mask print(f'''Pruned layer {name}''' ) elif pruning_method == "topK": if "mask_scores" in name: continue A_ = name[:-6] A_ = model[f'''{prefix_}mask_scores'''] A_ = TopKBinarizer.apply(lowerCamelCase_ ,lowerCamelCase_ ) A_ = tensor * mask print(f'''Pruned layer {name}''' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue A_ = name[:-6] A_ = model[f'''{prefix_}mask_scores'''] A_ = ThresholdBinarizer.apply(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) A_ = tensor * mask print(f'''Pruned layer {name}''' ) elif pruning_method == "l0": if "mask_scores" in name: continue A_ = name[:-6] A_ = model[f'''{prefix_}mask_scores'''] A_ = -0.1, 1.1 A_ = torch.sigmoid(lowerCamelCase_ ) A_ = s * (r - l) + l A_ = s_bar.clamp(min=0.0 ,max=1.0 ) A_ = tensor * mask print(f'''Pruned layer {name}''' ) else: raise ValueError("Unknown pruning method" ) if target_model_path is None: A_ = os.path.join( os.path.dirname(lowerCamelCase_ ) ,f'''bertarized_{os.path.basename(lowerCamelCase_ )}''' ) if not os.path.isdir(lowerCamelCase_ ): shutil.copytree(lowerCamelCase_ ,lowerCamelCase_ ) print(f'''\nCreated folder {target_model_path}''' ) torch.save(lowerCamelCase_ ,os.path.join(lowerCamelCase_ ,"pytorch_model.bin" ) ) print("\nPruned model saved! See you later!" ) if __name__ == "__main__": __a :Dict = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) __a :int = parser.parse_args() main(args)
368
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() __a :Optional[Any] = logging.get_logger(__name__) def __snake_case ( __UpperCamelCase : List[str] ,__UpperCamelCase : Any ,__UpperCamelCase : List[str] ,__UpperCamelCase : Optional[Any] ): """simple docstring""" A_ = original_name.split("." )[0] A_ = key.split("." ) A_ = int(key_list[key_list.index(__UpperCamelCase ) - 2] ) A_ = int(key_list[key_list.index(__UpperCamelCase ) - 1] ) A_ = orig_block_num - offset A_ = key.replace(f'''{orig_block_num}.{layer_num}.{original_name}''' ,f'''block.{new_block_num}.{layer_num}.{new_name}''' ) return key def __snake_case ( __UpperCamelCase : Any ): """simple docstring""" A_ = OrderedDict() A_ , A_ = 0, 0 for key, value in state_dict.items(): if key.startswith("network" ): A_ = key.replace("network" ,"poolformer.encoder" ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith("bias" ) and "patch_embed" not in key: patch_emb_offset += 1 A_ = key[: key.find("proj" )] A_ = key.replace(__UpperCamelCase ,f'''patch_embeddings.{total_embed_found}.''' ) A_ = key.replace("proj" ,"projection" ) if key.endswith("bias" ): total_embed_found += 1 if "patch_embeddings" in key: A_ = "poolformer.encoder." + key if "mlp.fc1" in key: A_ = replace_key_with_offset(__UpperCamelCase ,__UpperCamelCase ,"mlp.fc1" ,"output.conv1" ) if "mlp.fc2" in key: A_ = replace_key_with_offset(__UpperCamelCase ,__UpperCamelCase ,"mlp.fc2" ,"output.conv2" ) if "norm1" in key: A_ = replace_key_with_offset(__UpperCamelCase ,__UpperCamelCase ,"norm1" ,"before_norm" ) if "norm2" in key: A_ = replace_key_with_offset(__UpperCamelCase ,__UpperCamelCase ,"norm2" ,"after_norm" ) if "layer_scale_1" in key: A_ = replace_key_with_offset(__UpperCamelCase ,__UpperCamelCase ,"layer_scale_1" ,"layer_scale_1" ) if "layer_scale_2" in key: A_ = replace_key_with_offset(__UpperCamelCase ,__UpperCamelCase ,"layer_scale_2" ,"layer_scale_2" ) if "head" in key: A_ = key.replace("head" ,"classifier" ) A_ = value return new_state_dict def __snake_case ( ): """simple docstring""" A_ = "http://images.cocodataset.org/val2017/000000039769.jpg" A_ = Image.open(requests.get(__UpperCamelCase ,stream=__UpperCamelCase ).raw ) return image @torch.no_grad() def __snake_case ( __UpperCamelCase : Union[str, Any] ,__UpperCamelCase : List[str] ,__UpperCamelCase : List[Any] ): """simple docstring""" A_ = PoolFormerConfig() # set attributes based on model_name A_ = "huggingface/label-files" A_ = model_name[-3:] A_ = 1000 A_ = "imagenet-1k-id2label.json" A_ = (1, 1000) # set config attributes A_ = json.load(open(hf_hub_download(__UpperCamelCase ,__UpperCamelCase ,repo_type="dataset" ) ,"r" ) ) A_ = {int(__UpperCamelCase ): v for k, v in idalabel.items()} A_ = idalabel A_ = {v: k for k, v in idalabel.items()} if size == "s12": A_ = [2, 2, 6, 2] A_ = [64, 128, 320, 512] A_ = 4.0 A_ = 0.9 elif size == "s24": A_ = [4, 4, 12, 4] A_ = [64, 128, 320, 512] A_ = 4.0 A_ = 0.9 elif size == "s36": A_ = [6, 6, 18, 6] A_ = [64, 128, 320, 512] A_ = 4.0 A_ = 1E-6 A_ = 0.9 elif size == "m36": A_ = [6, 6, 18, 6] A_ = [96, 192, 384, 768] A_ = 4.0 A_ = 1E-6 A_ = 0.95 elif size == "m48": A_ = [8, 8, 24, 8] A_ = [96, 192, 384, 768] A_ = 4.0 A_ = 1E-6 A_ = 0.95 else: raise ValueError(f'''Size {size} not supported''' ) # load image processor A_ = PoolFormerImageProcessor(crop_pct=__UpperCamelCase ) # Prepare image A_ = prepare_img() A_ = image_processor(images=__UpperCamelCase ,return_tensors="pt" ).pixel_values logger.info(f'''Converting model {model_name}...''' ) # load original state dict A_ = torch.load(__UpperCamelCase ,map_location=torch.device("cpu" ) ) # rename keys A_ = rename_keys(__UpperCamelCase ) # create HuggingFace model and load state dict A_ = PoolFormerForImageClassification(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) model.eval() # Define image processor A_ = PoolFormerImageProcessor(crop_pct=__UpperCamelCase ) A_ = image_processor(images=prepare_img() ,return_tensors="pt" ).pixel_values # forward pass A_ = model(__UpperCamelCase ) A_ = outputs.logits # define expected logit slices for different models if size == "s12": A_ = torch.tensor([-0.3045, -0.6758, -0.4869] ) elif size == "s24": A_ = torch.tensor([0.4402, -0.1374, -0.8045] ) elif size == "s36": A_ = torch.tensor([-0.6080, -0.5133, -0.5898] ) elif size == "m36": A_ = torch.tensor([0.3952, 0.2263, -1.2668] ) elif size == "m48": A_ = torch.tensor([0.1167, -0.0656, -0.3423] ) else: raise ValueError(f'''Size {size} not supported''' ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] ,__UpperCamelCase ,atol=1E-2 ) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__UpperCamelCase ).mkdir(exist_ok=__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __a :Union[str, Any] = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) __a :int = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
329
0
'''simple docstring''' from __future__ import annotations import math def lowerCAmelCase (__A): """simple docstring""" if num <= 0: _a = F'''{num}: Invalid input, please enter a positive integer.''' raise ValueError(a__) _a = [True] * (num + 1) _a = [] _a = 2 _a = int(math.sqrt(a__)) while start <= end: # If start is a prime if sieve[start] is True: prime.append(a__) # Set multiples of start be False for i in range(start * start , num + 1 , a__): if sieve[i] is True: _a = False start += 1 for j in range(end + 1 , num + 1): if sieve[j] is True: prime.append(a__) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
211
def __lowerCAmelCase ( a__ , a__ , a__ ) -> list: __a = len(a__ ) __a = [[0] * n for i in range(a__ )] for i in range(a__ ): __a = y_points[i] for i in range(2 , a__ ): for j in range(a__ , a__ ): __a = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
6
0
def __UpperCamelCase ( _lowerCAmelCase ) -> int: """simple docstring""" if not numbers: return 0 if not isinstance(_lowerCAmelCase , (list, tuple) ) or not all( isinstance(_lowerCAmelCase , _lowerCAmelCase ) for number in numbers ): raise ValueError("""numbers must be an iterable of integers""" ) A : Dict = numbers[0] for i in range(1 , len(_lowerCAmelCase ) ): # update the maximum and minimum subarray products A : Optional[int] = numbers[i] if number < 0: A , A : List[str] = min_till_now, max_till_now A : Dict = max(_lowerCAmelCase , max_till_now * number ) A : Union[str, Any] = min(_lowerCAmelCase , min_till_now * number ) # update the maximum product found till now A : Any = max(_lowerCAmelCase , _lowerCAmelCase ) return max_prod
115
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available SCREAMING_SNAKE_CASE_:Optional[int] = { """configuration_tapas""": ["""TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TapasConfig"""], """tokenization_tapas""": ["""TapasTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:Optional[int] = [ """TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST""", """TapasForMaskedLM""", """TapasForQuestionAnswering""", """TapasForSequenceClassification""", """TapasModel""", """TapasPreTrainedModel""", """load_tf_weights_in_tapas""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:Optional[Any] = [ """TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFTapasForMaskedLM""", """TFTapasForQuestionAnswering""", """TFTapasForSequenceClassification""", """TFTapasModel""", """TFTapasPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE_:str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
115
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowerCAmelCase = logging.get_logger(__name__) def __lowerCAmelCase ( snake_case__ ): __UpperCamelCase : Tuple = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: __UpperCamelCase : str = 192 __UpperCamelCase : Optional[int] = 768 __UpperCamelCase : Dict = 12 __UpperCamelCase : List[str] = 3 __UpperCamelCase : Optional[Any] = [800, 1_333] __UpperCamelCase : List[Any] = False elif yolos_name == "yolos_s_dWr": __UpperCamelCase : Any = 330 __UpperCamelCase : List[Any] = 14 __UpperCamelCase : Union[str, Any] = 6 __UpperCamelCase : Optional[int] = 1_320 elif "yolos_s" in yolos_name: __UpperCamelCase : Union[str, Any] = 384 __UpperCamelCase : Union[str, Any] = 1_536 __UpperCamelCase : Union[str, Any] = 12 __UpperCamelCase : Optional[Any] = 6 elif "yolos_b" in yolos_name: __UpperCamelCase : Tuple = [800, 1_344] __UpperCamelCase : List[Any] = 91 __UpperCamelCase : int = "huggingface/label-files" __UpperCamelCase : List[Any] = "coco-detection-id2label.json" __UpperCamelCase : Dict = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="dataset" ) , "r" ) ) __UpperCamelCase : List[Any] = {int(snake_case__ ): v for k, v in idalabel.items()} __UpperCamelCase : Any = idalabel __UpperCamelCase : List[Any] = {v: k for k, v in idalabel.items()} return config def __lowerCAmelCase ( snake_case__ , snake_case__ , snake_case__ = False ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __UpperCamelCase : int = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) __UpperCamelCase : List[Any] = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __UpperCamelCase : int = in_proj_weight[: config.hidden_size, :] __UpperCamelCase : Dict = in_proj_bias[: config.hidden_size] __UpperCamelCase : List[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __UpperCamelCase : List[str] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __UpperCamelCase : Tuple = in_proj_weight[-config.hidden_size :, :] __UpperCamelCase : Optional[Any] = in_proj_bias[-config.hidden_size :] def __lowerCAmelCase ( snake_case__ ): if "backbone" in name: __UpperCamelCase : Union[str, Any] = name.replace("backbone" , "vit" ) if "cls_token" in name: __UpperCamelCase : Dict = name.replace("cls_token" , "embeddings.cls_token" ) if "det_token" in name: __UpperCamelCase : Optional[Any] = name.replace("det_token" , "embeddings.detection_tokens" ) if "mid_pos_embed" in name: __UpperCamelCase : Tuple = name.replace("mid_pos_embed" , "encoder.mid_position_embeddings" ) if "pos_embed" in name: __UpperCamelCase : Optional[Any] = name.replace("pos_embed" , "embeddings.position_embeddings" ) if "patch_embed.proj" in name: __UpperCamelCase : Dict = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "blocks" in name: __UpperCamelCase : Optional[Any] = name.replace("blocks" , "encoder.layer" ) if "attn.proj" in name: __UpperCamelCase : Dict = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: __UpperCamelCase : Any = name.replace("attn" , "attention.self" ) if "norm1" in name: __UpperCamelCase : Tuple = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: __UpperCamelCase : List[Any] = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: __UpperCamelCase : Optional[Any] = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: __UpperCamelCase : Union[str, Any] = name.replace("mlp.fc2" , "output.dense" ) if "class_embed" in name: __UpperCamelCase : List[str] = name.replace("class_embed" , "class_labels_classifier" ) if "bbox_embed" in name: __UpperCamelCase : Dict = name.replace("bbox_embed" , "bbox_predictor" ) if "vit.norm" in name: __UpperCamelCase : Dict = name.replace("vit.norm" , "vit.layernorm" ) return name def __lowerCAmelCase ( snake_case__ , snake_case__ ): for key in orig_state_dict.copy().keys(): __UpperCamelCase : Any = orig_state_dict.pop(snake_case__ ) if "qkv" in key: __UpperCamelCase : Optional[Any] = key.split("." ) __UpperCamelCase : List[Any] = int(key_split[2] ) __UpperCamelCase : Tuple = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: __UpperCamelCase : Optional[int] = val[:dim, :] __UpperCamelCase : Dict = val[ dim : dim * 2, : ] __UpperCamelCase : Optional[int] = val[-dim:, :] else: __UpperCamelCase : Optional[int] = val[:dim] __UpperCamelCase : Any = val[dim : dim * 2] __UpperCamelCase : Union[str, Any] = val[-dim:] else: __UpperCamelCase : Any = val return orig_state_dict def __lowerCAmelCase ( ): __UpperCamelCase : int = "http://images.cocodataset.org/val2017/000000039769.jpg" __UpperCamelCase : Any = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = False ): __UpperCamelCase : int = get_yolos_config(snake_case__ ) # load original state_dict __UpperCamelCase : int = torch.load(snake_case__ , map_location="cpu" )["model"] # load 🤗 model __UpperCamelCase : Tuple = YolosForObjectDetection(snake_case__ ) model.eval() __UpperCamelCase : Optional[int] = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by YolosImageProcessor __UpperCamelCase : Tuple = 800 if yolos_name != "yolos_ti" else 512 __UpperCamelCase : str = YolosImageProcessor(format="coco_detection" , size=snake_case__ ) __UpperCamelCase : List[Any] = image_processor(images=prepare_img() , return_tensors="pt" ) __UpperCamelCase : List[Any] = model(**snake_case__ ) __UpperCamelCase , __UpperCamelCase : List[Any] = outputs.logits, outputs.pred_boxes __UpperCamelCase , __UpperCamelCase : Optional[int] = None, None if yolos_name == "yolos_ti": __UpperCamelCase : List[str] = torch.tensor( [[-39.5022, -11.9820, -17.6888], [-29.9574, -9.9769, -17.7691], [-42.3281, -20.7200, -30.6294]] ) __UpperCamelCase : Optional[int] = torch.tensor( [[0.4021, 0.0836, 0.7979], [0.0184, 0.2609, 0.0364], [0.1781, 0.2004, 0.2095]] ) elif yolos_name == "yolos_s_200_pre": __UpperCamelCase : int = torch.tensor( [[-24.0248, -10.3024, -14.8290], [-42.0392, -16.8200, -27.4334], [-27.2743, -11.8154, -18.7148]] ) __UpperCamelCase : Union[str, Any] = torch.tensor( [[0.2559, 0.5455, 0.4706], [0.2989, 0.7279, 0.1875], [0.7732, 0.4017, 0.4462]] ) elif yolos_name == "yolos_s_300_pre": __UpperCamelCase : Union[str, Any] = torch.tensor( [[-36.2220, -14.4385, -23.5457], [-35.6970, -14.7583, -21.3935], [-31.5939, -13.6042, -16.8049]] ) __UpperCamelCase : Any = torch.tensor( [[0.7614, 0.2316, 0.4728], [0.7168, 0.4495, 0.3855], [0.4996, 0.1466, 0.9996]] ) elif yolos_name == "yolos_s_dWr": __UpperCamelCase : Optional[Any] = torch.tensor( [[-42.8668, -24.1049, -41.1690], [-34.7456, -14.1274, -24.9194], [-33.7898, -12.1946, -25.6495]] ) __UpperCamelCase : Dict = torch.tensor( [[0.5587, 0.2773, 0.0605], [0.5004, 0.3014, 0.9994], [0.4999, 0.1548, 0.9994]] ) elif yolos_name == "yolos_base": __UpperCamelCase : Dict = torch.tensor( [[-40.6064, -24.3084, -32.6447], [-55.1990, -30.7719, -35.5877], [-51.4311, -33.3507, -35.6462]] ) __UpperCamelCase : Tuple = torch.tensor( [[0.5555, 0.2794, 0.0655], [0.9049, 0.2664, 0.1894], [0.9183, 0.1984, 0.1635]] ) else: raise ValueError(F"Unknown yolos_name: {yolos_name}" ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1E-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , snake_case__ , atol=1E-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F"Saving model {yolos_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(snake_case__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: __UpperCamelCase : Union[str, Any] = { "yolos_ti": "yolos-tiny", "yolos_s_200_pre": "yolos-small", "yolos_s_300_pre": "yolos-small-300", "yolos_s_dWr": "yolos-small-dwr", "yolos_base": "yolos-base", } print("Pushing to the hub..." ) __UpperCamelCase : Dict = model_mapping[yolos_name] image_processor.push_to_hub(snake_case__ , organization="hustvl" ) model.push_to_hub(snake_case__ , organization="hustvl" ) if __name__ == "__main__": _lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _lowerCAmelCase = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
298
'''simple docstring''' def __lowerCAmelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ): __UpperCamelCase : Dict = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("All input parameters must be positive" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("Relative densities cannot be greater than one" ) else: __UpperCamelCase : str = 1 - (matter_density + radiation_density + dark_energy) __UpperCamelCase : List[Any] = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) __UpperCamelCase : Optional[Any] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _lowerCAmelCase = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1E-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
298
1
"""simple docstring""" import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def lowerCamelCase_ (UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[Any]=() , UpperCamelCase__ : List[Any]=None , UpperCamelCase__ : Union[str, Any]="no" , UpperCamelCase__ : Union[str, Any]="29500" ): _UpperCAmelCase : Optional[int] = False _UpperCAmelCase : Dict = False if any(key.startswith('''KAGGLE''' ) for key in os.environ.keys() ): _UpperCAmelCase : Tuple = True elif "IPython" in sys.modules: _UpperCAmelCase : List[Any] = '''google.colab''' in str(sys.modules['''IPython'''].get_ipython() ) try: _UpperCAmelCase : int = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F'Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.' ) if (in_colab or in_kaggle) and (os.environ.get('''TPU_NAME''' , lowercase__ ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( '''To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ''' '''your training function. Restart your notebook and make sure no cells initializes an ''' '''`Accelerator`.''' ) if num_processes is None: _UpperCAmelCase : Optional[int] = 8 _UpperCAmelCase : Dict = PrepareForLaunch(lowercase__ , distributed_type='''TPU''' ) print(F'Launching a training on {num_processes} TPU cores.' ) xmp.spawn(lowercase__ , args=lowercase__ , nprocs=lowercase__ , start_method='''fork''' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('''Launching training on one GPU.''' ) else: print('''Launching training on one CPU.''' ) function(*lowercase__ ) else: if num_processes is None: raise ValueError( '''You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.''' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( '''To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ''' '''inside your training function. Restart your notebook and make sure no cells initializes an ''' '''`Accelerator`.''' ) if torch.cuda.is_initialized(): raise ValueError( '''To launch a multi-GPU training from your notebook, you need to avoid running any instruction ''' '''using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ''' '''function.''' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=lowercase__ , master_addr='''127.0.01''' , master_port=lowercase__ , mixed_precision=lowercase__ ): _UpperCAmelCase : List[str] = PrepareForLaunch(lowercase__ , distributed_type='''MULTI_GPU''' ) print(F'Launching training on {num_processes} GPUs.' ) try: start_processes(lowercase__ , args=lowercase__ , nprocs=lowercase__ , start_method='''fork''' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( '''CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ''' '''This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ''' '''Please review your imports and test them when running the `notebook_launcher()` to identify ''' '''which one is problematic.''' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): _UpperCAmelCase : Any = '''1''' print('''Launching training on MPS.''' ) elif torch.cuda.is_available(): print('''Launching training on one GPU.''' ) else: print('''Launching training on CPU.''' ) function(*lowercase__ ) def lowerCamelCase_ (UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : int=() , UpperCamelCase__ : Any=2 ): from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=lowercase__ , master_addr='''127.0.01''' , master_port='''29500''' , accelerate_mixed_precision='''no''' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='''yes''' , ): _UpperCAmelCase : Optional[int] = PrepareForLaunch(lowercase__ , debug=lowercase__ ) start_processes(lowercase__ , args=lowercase__ , nprocs=lowercase__ , start_method='''fork''' )
363
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass class _UpperCAmelCase : '''simple docstring''' a__ =None a__ =None a__ =None # sigma(t_i) @classmethod def __lowerCAmelCase ( cls ) -> Tuple: return cls() @dataclass class _UpperCAmelCase ( a ): '''simple docstring''' a__ =42 a__ =42 a__ =42 class _UpperCAmelCase ( a ,a ): '''simple docstring''' @property def __lowerCAmelCase ( self ) -> Optional[int]: return True @register_to_config def __init__( self , A = 0.02 , A = 1_0_0 , A = 1.007 , A = 8_0 , A = 0.05 , A = 5_0 , ) -> Any: pass def __lowerCAmelCase ( self ) -> int: return KarrasVeSchedulerState.create() def __lowerCAmelCase ( self , A , A , A = () ) -> KarrasVeSchedulerState: _UpperCAmelCase : Tuple = jnp.arange(0 , A )[::-1].copy() _UpperCAmelCase : Dict = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in timesteps ] return state.replace( num_inference_steps=A , schedule=jnp.array(A , dtype=jnp.floataa ) , timesteps=A , ) def __lowerCAmelCase ( self , A , A , A , A , ) -> Tuple[jnp.ndarray, float]: if self.config.s_min <= sigma <= self.config.s_max: _UpperCAmelCase : Optional[Any] = min(self.config.s_churn / state.num_inference_steps , 2**0.5 - 1 ) else: _UpperCAmelCase : Tuple = 0 # sample eps ~ N(0, S_noise^2 * I) _UpperCAmelCase : Any = random.split(A , num=1 ) _UpperCAmelCase : Optional[Any] = self.config.s_noise * random.normal(key=A , shape=sample.shape ) _UpperCAmelCase : str = sigma + gamma * sigma _UpperCAmelCase : str = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __lowerCAmelCase ( self , A , A , A , A , A , A = True , ) -> Union[FlaxKarrasVeOutput, Tuple]: _UpperCAmelCase : Dict = sample_hat + sigma_hat * model_output _UpperCAmelCase : int = (sample_hat - pred_original_sample) / sigma_hat _UpperCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=A , derivative=A , state=A ) def __lowerCAmelCase ( self , A , A , A , A , A , A , A , A = True , ) -> Union[FlaxKarrasVeOutput, Tuple]: _UpperCAmelCase : Union[str, Any] = sample_prev + sigma_prev * model_output _UpperCAmelCase : List[Any] = (sample_prev - pred_original_sample) / sigma_prev _UpperCAmelCase : Union[str, Any] = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=A , derivative=A , state=A ) def __lowerCAmelCase ( self , A , A , A , A ) -> Any: raise NotImplementedError()
68
0
"""simple docstring""" from __future__ import annotations import math UpperCAmelCase = """2020.9.26""" UpperCAmelCase = """xcodz-dot, cclaus, dhruvmanila""" def lowercase ( a__ : float , a__ : float , a__ : float , a__ : float , a__ : float ) -> tuple[float, float]: if not all(isinstance(a__ , (float, int) ) for val in locals().values() ): _UpperCamelCase = F'''Input values must either be float or int: {list(locals().values() )}''' raise TypeError(a__ ) _UpperCamelCase = ((x * distance) / (z + distance)) * scale _UpperCamelCase = ((y * distance) / (z + distance)) * scale return projected_x, projected_y def lowercase ( a__ : float , a__ : float , a__ : float , a__ : str , a__ : float ) -> tuple[float, float, float]: if not isinstance(a__ , a__ ): raise TypeError('''Axis must be a str''' ) _UpperCamelCase = locals() del input_variables["axis"] if not all(isinstance(a__ , (float, int) ) for val in input_variables.values() ): _UpperCamelCase = ( '''Input values except axis must either be float or int: ''' F'''{list(input_variables.values() )}''' ) raise TypeError(a__ ) _UpperCamelCase = (angle % 360) / 450 * 180 / math.pi if axis == "z": _UpperCamelCase = x * math.cos(a__ ) - y * math.sin(a__ ) _UpperCamelCase = y * math.cos(a__ ) + x * math.sin(a__ ) _UpperCamelCase = z elif axis == "x": _UpperCamelCase = y * math.cos(a__ ) - z * math.sin(a__ ) _UpperCamelCase = z * math.cos(a__ ) + y * math.sin(a__ ) _UpperCamelCase = x elif axis == "y": _UpperCamelCase = x * math.cos(a__ ) - z * math.sin(a__ ) _UpperCamelCase = z * math.cos(a__ ) + x * math.sin(a__ ) _UpperCamelCase = y else: raise ValueError('''not a valid axis, choose one of \'x\', \'y\', \'z\'''' ) return new_x, new_y, new_z if __name__ == "__main__": import doctest doctest.testmod() print(F'''{convert_to_ad(1.0, 2.0, 3.0, 10.0, 10.0) = }''') print(F'''{rotate(1.0, 2.0, 3.0, 'y', 90.0) = }''')
256
"""simple docstring""" import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, task_specific_params UpperCAmelCase = [ # replace left string with right string to get the relevant state_dict key (identical state dict to bart) ["""memory_attention""", """encoder_attn"""], ["""attention""", """attn"""], ["""/""", """."""], [""".LayerNorm.gamma""", """_layer_norm.weight"""], [""".LayerNorm.beta""", """_layer_norm.bias"""], ["""r.layer_""", """r.layers."""], ["""output_proj""", """out_proj"""], ["""ffn.dense_1.""", """fc2."""], ["""ffn.dense.""", """fc1."""], ["""ffn_layer_norm""", """final_layer_norm"""], ["""kernel""", """weight"""], ["""encoder_layer_norm.""", """encoder.layer_norm."""], ["""decoder_layer_norm.""", """decoder.layer_norm."""], ["""embeddings.weights""", """shared.weight"""], ] def lowercase ( a__ : Optional[Any] ) -> Optional[int]: for pegasus_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(a__ , a__ ) return k def lowercase ( a__ : dict , a__ : dict ) -> PegasusForConditionalGeneration: _UpperCamelCase = DEFAULTS.copy() cfg_kwargs.update(a__ ) _UpperCamelCase = PegasusConfig(**a__ ) _UpperCamelCase = PegasusForConditionalGeneration(a__ ) _UpperCamelCase = torch_model.model.state_dict() _UpperCamelCase = {} for k, v in tf_weights.items(): _UpperCamelCase = rename_state_dict_key(a__ ) if new_k not in sd: raise ValueError(F'''could not find new key {new_k} in state dict. (converted from {k})''' ) if "dense" in k or "proj" in new_k: _UpperCamelCase = v.T _UpperCamelCase = torch.tensor(a__ , dtype=sd[new_k].dtype ) assert v.shape == sd[new_k].shape, F'''{new_k}, {k}, {v.shape}, {sd[new_k].shape}''' # make sure embedding.padding_idx is respected _UpperCamelCase = torch.zeros_like(mapping['''shared.weight'''][cfg.pad_token_id + 1] ) _UpperCamelCase = mapping['''shared.weight'''] _UpperCamelCase = mapping['''shared.weight'''] _UpperCamelCase = {k: torch.zeros_like(a__ ) for k, v in sd.items() if k.endswith('''bias''' ) and k not in mapping} mapping.update(**a__ ) _UpperCamelCase , _UpperCamelCase = torch_model.model.load_state_dict(a__ , strict=a__ ) _UpperCamelCase = [ k for k in missing if k not in ['''encoder.embed_positions.weight''', '''decoder.embed_positions.weight'''] ] assert unexpected_missing == [], F'''no matches found for the following torch keys {unexpected_missing}''' assert extra == [], F'''no matches found for the following tf keys {extra}''' return torch_model def lowercase ( a__ : List[str]="./ckpt/aeslc/model.ckpt-32000" ) -> Dict: _UpperCamelCase = tf.train.list_variables(a__ ) _UpperCamelCase = {} _UpperCamelCase = ['''Adafactor''', '''global_step'''] for name, shape in tqdm(a__ , desc='''converting tf checkpoint to dict''' ): _UpperCamelCase = any(pat in name for pat in ignore_name ) if skip_key: continue _UpperCamelCase = tf.train.load_variable(a__ , a__ ) _UpperCamelCase = array return tf_weights def lowercase ( a__ : str , a__ : str ) -> Optional[Any]: # save tokenizer first _UpperCamelCase = Path(a__ ).parent.name _UpperCamelCase = task_specific_params[F'''summarization_{dataset}''']['''max_position_embeddings'''] _UpperCamelCase = PegasusTokenizer.from_pretrained('''sshleifer/pegasus''' , model_max_length=a__ ) assert tok.model_max_length == desired_max_model_length tok.save_pretrained(a__ ) # convert model _UpperCamelCase = get_tf_weights_as_numpy(a__ ) _UpperCamelCase = task_specific_params[F'''summarization_{dataset}'''] if dataset == "large": _UpperCamelCase = task_specific_params _UpperCamelCase = convert_pegasus(a__ , a__ ) torch_model.save_pretrained(a__ ) _UpperCamelCase = torch_model.state_dict() sd.pop('''model.decoder.embed_positions.weight''' ) sd.pop('''model.encoder.embed_positions.weight''' ) torch.save(a__ , Path(a__ ) / '''pytorch_model.bin''' ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("""tf_ckpt_path""", type=str, help="""passed to tf.train.list_variables""") parser.add_argument("""save_dir""", default=None, type=str, help="""Path to the output PyTorch model.""") UpperCAmelCase = parser.parse_args() if args.save_dir is None: UpperCAmelCase = Path(args.tf_ckpt_path).parent.name UpperCAmelCase = os.path.join("""pegasus""", dataset) convert_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir)
256
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { "weiweishi/roc-bert-base-zh": "https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json", } class A ( _UpperCAmelCase ): """simple docstring""" lowerCamelCase = 'roc_bert' def __init__( self : Union[str, Any],lowercase_ : Dict=3_0_5_2_2,lowercase_ : Any=7_6_8,lowercase_ : Optional[int]=1_2,lowercase_ : Tuple=1_2,lowercase_ : Optional[Any]=3_0_7_2,lowercase_ : Tuple="gelu",lowercase_ : Dict=0.1,lowercase_ : Tuple=0.1,lowercase_ : Tuple=5_1_2,lowercase_ : str=2,lowercase_ : List[str]=0.02,lowercase_ : List[str]=1E-12,lowercase_ : Optional[Any]=True,lowercase_ : List[str]=0,lowercase_ : str="absolute",lowercase_ : int=None,lowercase_ : List[str]=True,lowercase_ : List[Any]=True,lowercase_ : Tuple=7_6_8,lowercase_ : Union[str, Any]=9_1_0,lowercase_ : Optional[Any]=5_1_2,lowercase_ : Tuple=2_4_8_5_8,lowercase_ : Dict=True,**lowercase_ : Optional[Any],)-> Tuple: '''simple docstring''' A__ = vocab_size A__ = max_position_embeddings A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = initializer_range A__ = type_vocab_size A__ = layer_norm_eps A__ = use_cache A__ = enable_pronunciation A__ = enable_shape A__ = pronunciation_embed_dim A__ = pronunciation_vocab_size A__ = shape_embed_dim A__ = shape_vocab_size A__ = concat_input A__ = position_embedding_type A__ = classifier_dropout super().__init__(pad_token_id=lowercase_,**lowercase_ )
282
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
282
1
"""simple docstring""" _a : Union[str, Any] = '0.21.0' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
44
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( a , a , a , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ = StableDiffusionInpaintPipeline lowerCAmelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS lowerCAmelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowerCAmelCase__ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowerCAmelCase__ = frozenset([] ) def UpperCAmelCase__ ( self : str ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=__SCREAMING_SNAKE_CASE , ) __SCREAMING_SNAKE_CASE = PNDMScheduler(skip_prk_steps=__SCREAMING_SNAKE_CASE ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) __SCREAMING_SNAKE_CASE = CLIPTextModel(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) __SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCAmelCase__ ( self : Any , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[Any]=0 ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 32, 32) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ).to(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = image.cpu().permute(0 , 2 , 3 , 1 )[0] __SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(__SCREAMING_SNAKE_CASE ) ).convert("""RGB""" ).resize((64, 64) ) __SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(image + 4 ) ).convert("""RGB""" ).resize((64, 64) ) if str(__SCREAMING_SNAKE_CASE ).startswith("""mps""" ): __SCREAMING_SNAKE_CASE = torch.manual_seed(__SCREAMING_SNAKE_CASE ) else: __SCREAMING_SNAKE_CASE = torch.Generator(device=__SCREAMING_SNAKE_CASE ).manual_seed(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase__ ( self : Union[str, Any] ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE = self.get_dummy_components() __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline(**__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = sd_pipe.to(__SCREAMING_SNAKE_CASE ) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = sd_pipe(**__SCREAMING_SNAKE_CASE ).images __SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCAmelCase__ ( self : Tuple ) -> str: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self : List[Any] ) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self : List[str] ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) __SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench.npy""" ) __SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-2-inpainting""" __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline.from_pretrained(__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE = """Face of a yellow cat, high resolution, sitting on a park bench""" __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __SCREAMING_SNAKE_CASE = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9E-3 def UpperCAmelCase__ ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) __SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench_fp16.npy""" ) __SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-2-inpainting""" __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline.from_pretrained( __SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , safety_checker=__SCREAMING_SNAKE_CASE , ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE = """Face of a yellow cat, high resolution, sitting on a park bench""" __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __SCREAMING_SNAKE_CASE = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def UpperCAmelCase__ ( self : Tuple ) -> Any: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) __SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-2-inpainting""" __SCREAMING_SNAKE_CASE = PNDMScheduler.from_pretrained(__SCREAMING_SNAKE_CASE , subfolder="""scheduler""" ) __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline.from_pretrained( __SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __SCREAMING_SNAKE_CASE = """Face of a yellow cat, high resolution, sitting on a park bench""" __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type="""np""" , ) __SCREAMING_SNAKE_CASE = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
267
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase : Any = logging.get_logger(__name__) _lowerCamelCase : List[Any] = { '''facebook/vit-mae-base''': '''https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json''', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : Optional[int] = "vit_mae" def __init__( self : Any , lowercase : List[Any]=768 , lowercase : int=12 , lowercase : str=12 , lowercase : int=3_072 , lowercase : Optional[int]="gelu" , lowercase : List[str]=0.0 , lowercase : Tuple=0.0 , lowercase : Tuple=0.02 , lowercase : int=1E-12 , lowercase : List[str]=224 , lowercase : int=16 , lowercase : Dict=3 , lowercase : Any=True , lowercase : Optional[int]=16 , lowercase : Tuple=512 , lowercase : Optional[int]=8 , lowercase : Union[str, Any]=2_048 , lowercase : Any=0.75 , lowercase : List[Any]=False , **lowercase : List[Any] , ): '''simple docstring''' super().__init__(**lowercase ) _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = image_size _snake_case = patch_size _snake_case = num_channels _snake_case = qkv_bias _snake_case = decoder_num_attention_heads _snake_case = decoder_hidden_size _snake_case = decoder_num_hidden_layers _snake_case = decoder_intermediate_size _snake_case = mask_ratio _snake_case = norm_pix_loss
130
import random class SCREAMING_SNAKE_CASE__ : '''simple docstring''' @staticmethod def A ( lowercase : str ): '''simple docstring''' _snake_case = [ord(lowercase ) for i in text] _snake_case = [] _snake_case = [] for i in plain: _snake_case = random.randint(1 , 300 ) _snake_case = (i + k) * k cipher.append(lowercase ) key.append(lowercase ) return cipher, key @staticmethod def A ( lowercase : list[int] , lowercase : list[int] ): '''simple docstring''' _snake_case = [] for i in range(len(lowercase ) ): _snake_case = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(lowercase ) ) return "".join(lowercase ) if __name__ == "__main__": _lowerCamelCase , _lowerCamelCase : Optional[int] = Onepad().encrypt('''Hello''') print(c, k) print(Onepad().decrypt(c, k))
130
1
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase_ : Optional[Any] = {'configuration_timm_backbone': ['TimmBackboneConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Optional[int] = ['TimmBackbone'] if TYPE_CHECKING: from .configuration_timm_backbone import TimmBackboneConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timm_backbone import TimmBackbone else: import sys lowerCAmelCase_ : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
63
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _snake_case : Optional[int] = { "configuration_distilbert": [ "DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DistilBertConfig", "DistilBertOnnxConfig", ], "tokenization_distilbert": ["DistilBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ["DistilBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = [ "DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "DistilBertForMaskedLM", "DistilBertForMultipleChoice", "DistilBertForQuestionAnswering", "DistilBertForSequenceClassification", "DistilBertForTokenClassification", "DistilBertModel", "DistilBertPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ "TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDistilBertForMaskedLM", "TFDistilBertForMultipleChoice", "TFDistilBertForQuestionAnswering", "TFDistilBertForSequenceClassification", "TFDistilBertForTokenClassification", "TFDistilBertMainLayer", "TFDistilBertModel", "TFDistilBertPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : str = [ "FlaxDistilBertForMaskedLM", "FlaxDistilBertForMultipleChoice", "FlaxDistilBertForQuestionAnswering", "FlaxDistilBertForSequenceClassification", "FlaxDistilBertForTokenClassification", "FlaxDistilBertModel", "FlaxDistilBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys _snake_case : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
123
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer _A = logging.get_logger(__name__) _A = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _A = { 'vocab_file': { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/vocab.txt', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/vocab.txt', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/vocab.txt', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/vocab.txt', 'bert-base-multilingual-uncased': ( 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/vocab.txt' ), 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/vocab.txt', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/vocab.txt', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/vocab.txt', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/vocab.txt' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/vocab.txt' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt' ), 'bert-base-cased-finetuned-mrpc': ( 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/vocab.txt' ), 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/vocab.txt', 'bert-base-german-dbmdz-uncased': ( 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/vocab.txt' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/vocab.txt' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/vocab.txt' ), 'wietsedv/bert-base-dutch-cased': ( 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/tokenizer.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/tokenizer.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/tokenizer.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/tokenizer.json', 'bert-base-multilingual-uncased': ( 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/tokenizer.json' ), 'bert-base-multilingual-cased': ( 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/tokenizer.json' ), 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/tokenizer.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/tokenizer.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/tokenizer.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/tokenizer.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json' ), 'bert-base-cased-finetuned-mrpc': ( 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/tokenizer.json' ), 'bert-base-german-dbmdz-cased': ( 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/tokenizer.json' ), 'bert-base-german-dbmdz-uncased': ( 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/tokenizer.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/tokenizer.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/tokenizer.json' ), 'wietsedv/bert-base-dutch-cased': ( 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/tokenizer.json' ), }, } _A = { 'bert-base-uncased': 512, 'bert-large-uncased': 512, 'bert-base-cased': 512, 'bert-large-cased': 512, 'bert-base-multilingual-uncased': 512, 'bert-base-multilingual-cased': 512, 'bert-base-chinese': 512, 'bert-base-german-cased': 512, 'bert-large-uncased-whole-word-masking': 512, 'bert-large-cased-whole-word-masking': 512, 'bert-large-uncased-whole-word-masking-finetuned-squad': 512, 'bert-large-cased-whole-word-masking-finetuned-squad': 512, 'bert-base-cased-finetuned-mrpc': 512, 'bert-base-german-dbmdz-cased': 512, 'bert-base-german-dbmdz-uncased': 512, 'TurkuNLP/bert-base-finnish-cased-v1': 512, 'TurkuNLP/bert-base-finnish-uncased-v1': 512, 'wietsedv/bert-base-dutch-cased': 512, } _A = { 'bert-base-uncased': {'do_lower_case': True}, 'bert-large-uncased': {'do_lower_case': True}, 'bert-base-cased': {'do_lower_case': False}, 'bert-large-cased': {'do_lower_case': False}, 'bert-base-multilingual-uncased': {'do_lower_case': True}, 'bert-base-multilingual-cased': {'do_lower_case': False}, 'bert-base-chinese': {'do_lower_case': False}, 'bert-base-german-cased': {'do_lower_case': False}, 'bert-large-uncased-whole-word-masking': {'do_lower_case': True}, 'bert-large-cased-whole-word-masking': {'do_lower_case': False}, 'bert-large-uncased-whole-word-masking-finetuned-squad': {'do_lower_case': True}, 'bert-large-cased-whole-word-masking-finetuned-squad': {'do_lower_case': False}, 'bert-base-cased-finetuned-mrpc': {'do_lower_case': False}, 'bert-base-german-dbmdz-cased': {'do_lower_case': False}, 'bert-base-german-dbmdz-uncased': {'do_lower_case': True}, 'TurkuNLP/bert-base-finnish-cased-v1': {'do_lower_case': False}, 'TurkuNLP/bert-base-finnish-uncased-v1': {'do_lower_case': True}, 'wietsedv/bert-base-dutch-cased': {'do_lower_case': False}, } class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Any = VOCAB_FILES_NAMES UpperCAmelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Optional[int] = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Dict = BertTokenizer def __init__( self , A_=None , A_=None , A_=True , A_="[UNK]" , A_="[SEP]" , A_="[PAD]" , A_="[CLS]" , A_="[MASK]" , A_=True , A_=None , **A_ , ) -> Any: super().__init__( A_ , tokenizer_file=A_ , do_lower_case=A_ , unk_token=A_ , sep_token=A_ , pad_token=A_ , cls_token=A_ , mask_token=A_ , tokenize_chinese_chars=A_ , strip_accents=A_ , **A_ , ) __UpperCamelCase =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , A_ ) != do_lower_case or normalizer_state.get('strip_accents' , A_ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , A_ ) != tokenize_chinese_chars ): __UpperCamelCase =getattr(A_ , normalizer_state.pop('type' ) ) __UpperCamelCase =do_lower_case __UpperCamelCase =strip_accents __UpperCamelCase =tokenize_chinese_chars __UpperCamelCase =normalizer_class(**A_ ) __UpperCamelCase =do_lower_case def _a ( self , A_ , A_=None ) -> List[str]: __UpperCamelCase =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _a ( self , A_ , A_ = None ) -> List[int]: __UpperCamelCase =[self.sep_token_id] __UpperCamelCase =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _a ( self , A_ , A_ = None ) -> Tuple[str]: __UpperCamelCase =self._tokenizer.model.save(A_ , name=A_ ) return tuple(A_ )
117
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from ...utils.dataclasses import ( ComputeEnvironment, DistributedType, DynamoBackend, PrecisionType, SageMakerDistributedType, ) from ..menu import BulletMenu _A = [ 'EAGER', 'AOT_EAGER', 'INDUCTOR', 'NVFUSER', 'AOT_NVFUSER', 'AOT_CUDAGRAPHS', 'OFI', 'FX2TRT', 'ONNXRT', 'IPEX', ] def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[Any]=None , SCREAMING_SNAKE_CASE__ : Optional[Any]=None , SCREAMING_SNAKE_CASE__ : str=None ): __UpperCamelCase =True while ask_again: __UpperCamelCase =input(SCREAMING_SNAKE_CASE__ ) try: if default is not None and len(SCREAMING_SNAKE_CASE__ ) == 0: return default return convert_value(SCREAMING_SNAKE_CASE__ ) if convert_value is not None else result except Exception: if error_message is not None: print(SCREAMING_SNAKE_CASE__ ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : List[str]=[] , SCREAMING_SNAKE_CASE__ : Union[str, Any]=None , SCREAMING_SNAKE_CASE__ : str=0 ): __UpperCamelCase =BulletMenu(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =menu.run(default_choice=SCREAMING_SNAKE_CASE__ ) return convert_value(SCREAMING_SNAKE_CASE__ ) if convert_value is not None else result def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Union[str, Any] ): __UpperCamelCase =int(SCREAMING_SNAKE_CASE__ ) return ComputeEnvironment(['LOCAL_MACHINE', 'AMAZON_SAGEMAKER'][value] ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : List[str] ): __UpperCamelCase =int(SCREAMING_SNAKE_CASE__ ) return DistributedType(['NO', 'MULTI_CPU', 'MULTI_XPU', 'MULTI_GPU', 'MULTI_NPU', 'TPU'][value] ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Optional[int] ): __UpperCamelCase =int(SCREAMING_SNAKE_CASE__ ) return DynamoBackend(DYNAMO_BACKENDS[value] ).value def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Dict ): __UpperCamelCase =int(SCREAMING_SNAKE_CASE__ ) return PrecisionType(['no', 'fp16', 'bf16', 'fp8'][value] ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Optional[Any] ): __UpperCamelCase =int(SCREAMING_SNAKE_CASE__ ) return SageMakerDistributedType(['NO', 'DATA_PARALLEL', 'MODEL_PARALLEL'][value] ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Any ): return {"yes": True, "no": False}[value.lower()] class UpperCAmelCase__ ( argparse.RawDescriptionHelpFormatter ): """simple docstring""" def _a ( self , A_ , A_ , A_ , A_ ) -> Optional[int]: __UpperCamelCase =super()._format_usage(A_ , A_ , A_ , A_ ) __UpperCamelCase =usage.replace('<command> [<args>] ' , '' ) return usage
117
1
def _a ( SCREAMING_SNAKE_CASE : list[int] , SCREAMING_SNAKE_CASE : list[int] ): """simple docstring""" if not len(snake_case_ ) == len(snake_case_ ) == 3: raise ValueError('''Please enter a valid equation.''' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('''Both a & b of two equations can\'t be zero.''' ) # Extract the coefficients UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ : Tuple = equationa UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ : Tuple = equationa # Calculate the determinants of the matrices UpperCamelCase__ : str = aa * ba - aa * ba UpperCamelCase__ : Dict = ca * ba - ca * ba UpperCamelCase__ : Tuple = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('''Infinite solutions. (Consistent system)''' ) else: raise ValueError('''No solution. (Inconsistent system)''' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: UpperCamelCase__ : str = determinant_x / determinant UpperCamelCase__ : int = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
146
'''simple docstring''' import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class _lowercase : '''simple docstring''' def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : List[str]=2 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3 , SCREAMING_SNAKE_CASE__ : Optional[Any]=64 , SCREAMING_SNAKE_CASE__ : int=None ) -> Optional[int]: __lowerCAmelCase = np.random.default_rng(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = length __lowerCAmelCase = rng.normal(size=(length,) ).astype(np.floataa ) __lowerCAmelCase = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self : Union[str, Any] ) -> Optional[Any]: return self.length def __getitem__( self : str , SCREAMING_SNAKE_CASE__ : Dict ) -> Optional[int]: return {"x": self.x[i], "y": self.y[i]} class _lowercase ( torch.nn.Module ): '''simple docstring''' def __init__( self : List[Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0 , SCREAMING_SNAKE_CASE__ : Tuple=0 , SCREAMING_SNAKE_CASE__ : Tuple=False ) -> Any: super().__init__() __lowerCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) __lowerCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) __lowerCAmelCase = True def a ( self : Tuple , SCREAMING_SNAKE_CASE__ : List[Any]=None ) -> str: if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) __lowerCAmelCase = False return x * self.a[0] + self.b[0] class _lowercase ( torch.nn.Module ): '''simple docstring''' def __init__( self : Any , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0 , SCREAMING_SNAKE_CASE__ : Optional[Any]=0 , SCREAMING_SNAKE_CASE__ : Optional[Any]=False ) -> Optional[Any]: super().__init__() __lowerCAmelCase = torch.nn.Parameter(torch.tensor(SCREAMING_SNAKE_CASE__ ).float() ) __lowerCAmelCase = torch.nn.Parameter(torch.tensor(SCREAMING_SNAKE_CASE__ ).float() ) __lowerCAmelCase = True def a ( self : List[str] , SCREAMING_SNAKE_CASE__ : Any=None ) -> int: if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) __lowerCAmelCase = False return x * self.a + self.b def UpperCamelCase_ ( snake_case_ : List[str] , snake_case_ : int = 16 ) -> int: '''simple docstring''' from datasets import load_dataset from transformers import AutoTokenizer __lowerCAmelCase = AutoTokenizer.from_pretrained("""bert-base-cased""" ) __lowerCAmelCase = {"""train""": """tests/test_samples/MRPC/train.csv""", """validation""": """tests/test_samples/MRPC/dev.csv"""} __lowerCAmelCase = load_dataset("""csv""" , data_files=snake_case_ ) __lowerCAmelCase = datasets["""train"""].unique("""label""" ) __lowerCAmelCase = {v: i for i, v in enumerate(snake_case_ )} def tokenize_function(snake_case_ : Optional[int] ): # max_length=None => use the model max length (it's actually the default) __lowerCAmelCase = tokenizer( examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case_ , max_length=snake_case_ , padding="""max_length""" ) if "label" in examples: __lowerCAmelCase = [label_to_id[l] for l in examples["""label"""]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __lowerCAmelCase = datasets.map( snake_case_ , batched=snake_case_ , remove_columns=["""sentence1""", """sentence2""", """label"""] , ) def collate_fn(snake_case_ : List[Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(snake_case_ , padding="""max_length""" , max_length=1_28 , return_tensors="""pt""" ) return tokenizer.pad(snake_case_ , padding="""longest""" , return_tensors="""pt""" ) # Instantiate dataloaders. __lowerCAmelCase = DataLoader(tokenized_datasets["""train"""] , shuffle=snake_case_ , collate_fn=snake_case_ , batch_size=2 ) __lowerCAmelCase = DataLoader(tokenized_datasets["""validation"""] , shuffle=snake_case_ , collate_fn=snake_case_ , batch_size=1 ) return train_dataloader, eval_dataloader
229
0
'''simple docstring''' from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase=1E-12 ) -> List[Any]: lowerCamelCase__ : List[Any] = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(UpperCamelCase , axis=1 ) , a_min=UpperCamelCase ) ).T lowerCamelCase__ : Any = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(UpperCamelCase , axis=1 ) , a_min=UpperCamelCase ) ).T return jnp.matmul(UpperCamelCase , norm_emb_a.T ) class _lowercase ( nn.Module ): a = 42 a = jnp.floataa def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : str = FlaxCLIPVisionModule(self.config.vision_config ) lowerCamelCase__ : List[Any] = nn.Dense(self.config.projection_dim , use_bias=UpperCamelCase__ , dtype=self.dtype ) lowerCamelCase__ : Any = self.param("""concept_embeds""" , jax.nn.initializers.ones , (17, self.config.projection_dim) ) lowerCamelCase__ : int = self.param( """special_care_embeds""" , jax.nn.initializers.ones , (3, self.config.projection_dim) ) lowerCamelCase__ : Optional[Any] = self.param("""concept_embeds_weights""" , jax.nn.initializers.ones , (17,) ) lowerCamelCase__ : Optional[Any] = self.param("""special_care_embeds_weights""" , jax.nn.initializers.ones , (3,) ) def __call__( self: List[Any] , UpperCamelCase__: int ): lowerCamelCase__ : List[Any] = self.vision_model(UpperCamelCase__ )[1] lowerCamelCase__ : str = self.visual_projection(UpperCamelCase__ ) lowerCamelCase__ : Tuple = jax_cosine_distance(UpperCamelCase__ , self.special_care_embeds ) lowerCamelCase__ : List[str] = jax_cosine_distance(UpperCamelCase__ , self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs lowerCamelCase__ : Optional[Any] = 0.0 lowerCamelCase__ : List[Any] = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment lowerCamelCase__ : Optional[Any] = jnp.round(UpperCamelCase__ , 3 ) lowerCamelCase__ : List[str] = jnp.any(special_scores > 0 , axis=1 , keepdims=UpperCamelCase__ ) # Use a lower threshold if an image has any special care concept lowerCamelCase__ : Optional[int] = is_special_care * 0.01 lowerCamelCase__ : Optional[int] = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment lowerCamelCase__ : Dict = jnp.round(UpperCamelCase__ , 3 ) lowerCamelCase__ : Any = jnp.any(concept_scores > 0 , axis=1 ) return has_nsfw_concepts class _lowercase ( _lowercase ): a = CLIPConfig a = """clip_input""" a = FlaxStableDiffusionSafetyCheckerModule def __init__( self: int , UpperCamelCase__: CLIPConfig , UpperCamelCase__: Optional[Tuple] = None , UpperCamelCase__: int = 0 , UpperCamelCase__: jnp.dtype = jnp.floataa , UpperCamelCase__: bool = True , **UpperCamelCase__: List[Any] , ): if input_shape is None: lowerCamelCase__ : int = (1, 224, 224, 3) lowerCamelCase__ : List[str] = self.module_class(config=UpperCamelCase__ , dtype=UpperCamelCase__ , **UpperCamelCase__ ) super().__init__(UpperCamelCase__ , UpperCamelCase__ , input_shape=UpperCamelCase__ , seed=UpperCamelCase__ , dtype=UpperCamelCase__ , _do_init=_do_init ) def lowerCamelCase_ ( self: str , UpperCamelCase__: jax.random.KeyArray , UpperCamelCase__: Tuple , UpperCamelCase__: FrozenDict = None ): # init input tensor lowerCamelCase__ : str = jax.random.normal(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ , lowerCamelCase__ : Any = jax.random.split(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = {"""params""": params_rng, """dropout""": dropout_rng} lowerCamelCase__ : Optional[int] = self.module.init(UpperCamelCase__ , UpperCamelCase__ )["""params"""] return random_params def __call__( self: List[str] , UpperCamelCase__: int , UpperCamelCase__: dict = None , ): lowerCamelCase__ : int = jnp.transpose(UpperCamelCase__ , (0, 2, 3, 1) ) return self.module.apply( {"""params""": params or self.params} , jnp.array(UpperCamelCase__ , dtype=jnp.floataa ) , rngs={} , )
129
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _A : List[str] =logging.get_logger(__name__) class _lowercase ( _lowercase ): a = ["""input_features""", """attention_mask"""] def __init__( self: Optional[int] , UpperCamelCase__: str=80 , UpperCamelCase__: Union[str, Any]=16_000 , UpperCamelCase__: str=0.0 , UpperCamelCase__: Tuple=10 , UpperCamelCase__: Dict=25 , UpperCamelCase__: List[Any]="hamming_window" , UpperCamelCase__: Tuple=32_768.0 , UpperCamelCase__: List[Any]=0.97 , UpperCamelCase__: Optional[int]=1.0 , UpperCamelCase__: str=True , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Optional[int]=False , **UpperCamelCase__: Optional[int] , ): super().__init__(feature_size=UpperCamelCase__ , sampling_rate=UpperCamelCase__ , padding_value=UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase__ : Dict = feature_size lowerCamelCase__ : Tuple = sampling_rate lowerCamelCase__ : Union[str, Any] = padding_value lowerCamelCase__ : Union[str, Any] = hop_length lowerCamelCase__ : Any = win_length lowerCamelCase__ : Tuple = frame_signal_scale lowerCamelCase__ : Dict = preemphasis_coeff lowerCamelCase__ : Optional[Any] = mel_floor lowerCamelCase__ : Any = normalize_means lowerCamelCase__ : Optional[int] = normalize_vars lowerCamelCase__ : Dict = win_function lowerCamelCase__ : Dict = return_attention_mask lowerCamelCase__ : Optional[Any] = win_length * sampling_rate // 1_000 lowerCamelCase__ : Tuple = hop_length * sampling_rate // 1_000 lowerCamelCase__ : List[Any] = optimal_fft_length(self.sample_size ) lowerCamelCase__ : List[str] = (self.n_fft // 2) + 1 def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: np.array ): if self.win_function == "hamming_window": lowerCamelCase__ : Dict = window_function(window_length=self.sample_size , name=self.win_function , periodic=UpperCamelCase__ ) else: lowerCamelCase__ : Union[str, Any] = window_function(window_length=self.sample_size , name=self.win_function ) lowerCamelCase__ : List[Any] = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) lowerCamelCase__ : Tuple = spectrogram( one_waveform * self.frame_signal_scale , window=UpperCamelCase__ , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=UpperCamelCase__ , preemphasis=self.preemphasis_coeff , mel_filters=UpperCamelCase__ , mel_floor=self.mel_floor , log_mel="""log""" , ) return msfc_features.T def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: List[str] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): # make sure we normalize float32 arrays if self.normalize_means: lowerCamelCase__ : List[str] = x[:input_length].mean(axis=0 ) lowerCamelCase__ : List[str] = np.subtract(UpperCamelCase__ , UpperCamelCase__ ) if self.normalize_vars: lowerCamelCase__ : Union[str, Any] = x[:input_length].std(axis=0 ) lowerCamelCase__ : str = np.divide(UpperCamelCase__ , UpperCamelCase__ ) if input_length < x.shape[0]: lowerCamelCase__ : List[Any] = padding_value # make sure array is in float32 lowerCamelCase__ : str = x.astype(np.floataa ) return x def lowerCamelCase_ ( self: str , UpperCamelCase__: List[np.ndarray] , UpperCamelCase__: Optional[np.ndarray] = None ): lowerCamelCase__ : Optional[int] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(UpperCamelCase__ , UpperCamelCase__ , self.padding_value ) for x, n in zip(UpperCamelCase__ , UpperCamelCase__ )] def __call__( self: List[str] , UpperCamelCase__: Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase__: Union[bool, str, PaddingStrategy] = False , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: bool = False , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: Optional[bool] = None , UpperCamelCase__: Optional[Union[str, TensorType]] = None , UpperCamelCase__: Optional[int] = None , **UpperCamelCase__: Tuple , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) lowerCamelCase__ : List[str] = isinstance(UpperCamelCase__ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) lowerCamelCase__ : List[str] = is_batched_numpy or ( isinstance(UpperCamelCase__ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCamelCase__ : str = [np.asarray(UpperCamelCase__ , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase__ , np.ndarray ): lowerCamelCase__ : List[Any] = np.asarray(UpperCamelCase__ , dtype=np.floataa ) elif isinstance(UpperCamelCase__ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCamelCase__ : Optional[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCamelCase__ : Optional[Any] = [raw_speech] # extract fbank features lowerCamelCase__ : str = [self._extract_mfsc_features(UpperCamelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCamelCase__ : str = BatchFeature({"""input_features""": features} ) lowerCamelCase__ : str = self.pad( UpperCamelCase__ , padding=UpperCamelCase__ , max_length=UpperCamelCase__ , truncation=UpperCamelCase__ , pad_to_multiple_of=UpperCamelCase__ , return_attention_mask=UpperCamelCase__ , **UpperCamelCase__ , ) # make sure list is in array format lowerCamelCase__ : str = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] , UpperCamelCase__ ): lowerCamelCase__ : Dict = [np.asarray(UpperCamelCase__ , dtype=np.floataa ) for feature in input_features] lowerCamelCase__ : List[str] = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: lowerCamelCase__ : Any = [np.asarray(UpperCamelCase__ , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCamelCase__ : str = ( np.array(UpperCamelCase__ , dtype=np.intaa ) if self._get_padding_strategies(UpperCamelCase__ , max_length=UpperCamelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCamelCase__ : Union[str, Any] = self.normalize( padded_inputs["""input_features"""] , attention_mask=UpperCamelCase__ ) if return_tensors is not None: lowerCamelCase__ : Dict = padded_inputs.convert_to_tensors(UpperCamelCase__ ) return padded_inputs
129
1
'''simple docstring''' from collections import deque def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' A : Dict = len(snake_case__ ) A : str = deque() A : int = [False for _ in range(snake_case__ )] A : List[Any] = [-1 for _ in range(snake_case__ )] A : Optional[int] = index_of[:] def strong_connect(snake_case__ , snake_case__ , snake_case__ ): A : int = index # the number when this node is seen A : Tuple = index # lowest rank node reachable from here index += 1 stack.append(snake_case__ ) A : List[str] = True for w in g[v]: if index_of[w] == -1: A : List[str] = strong_connect(snake_case__ , snake_case__ , snake_case__ ) A : str = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) elif on_stack[w]: A : List[Any] = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) if lowlink_of[v] == index_of[v]: A : Optional[int] = [] A : List[str] = stack.pop() A : List[str] = False component.append(snake_case__ ) while w != v: A : Optional[Any] = stack.pop() A : List[str] = False component.append(snake_case__ ) components.append(snake_case__ ) return index A : Any = [] for v in range(snake_case__ ): if index_of[v] == -1: strong_connect(snake_case__ , 0 , snake_case__ ) return components def lowerCAmelCase_ ( snake_case__ , snake_case__ ): '''simple docstring''' A : Optional[Any] = [[] for _ in range(snake_case__ )] for u, v in edges: g[u].append(snake_case__ ) return g if __name__ == "__main__": # Test lowercase : str = 7 lowercase : Any = [0, 0, 1, 2, 3, 3, 4, 4, 6] lowercase : Tuple = [1, 3, 2, 0, 1, 4, 5, 6, 5] lowercase : Optional[int] = [(u, v) for u, v in zip(source, target)] lowercase : Any = create_graph(n_vertices, edges) assert [[5], [6], [4], [3, 2, 1, 0]] == tarjan(g)
3
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## UpperCamelCase__ : Union[str, Any] = 16 UpperCamelCase__ : Dict = 32 def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ = 1_6 ) -> Tuple: """simple docstring""" a = AutoTokenizer.from_pretrained('''bert-base-cased''' ) a = load_dataset('''glue''', '''mrpc''' ) def tokenize_function(snake_case_ ): # max_length=None => use the model max length (it's actually the default) a = tokenizer(examples['''sentence1'''], examples['''sentence2'''], truncation=snake_case_, max_length=snake_case_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): a = datasets.map( snake_case_, batched=snake_case_, remove_columns=['''idx''', '''sentence1''', '''sentence2'''], ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library a = tokenized_datasets.rename_column('''label''', '''labels''' ) def collate_fn(snake_case_ ): # On TPU it's best to pad everything to the same length or training will be very slow. a = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": a = 1_6 elif accelerator.mixed_precision != "no": a = 8 else: a = None return tokenizer.pad( snake_case_, padding='''longest''', max_length=snake_case_, pad_to_multiple_of=snake_case_, return_tensors='''pt''', ) # Instantiate dataloaders. a = DataLoader( tokenized_datasets['''train'''], shuffle=snake_case_, collate_fn=snake_case_, batch_size=snake_case_ ) a = DataLoader( tokenized_datasets['''validation'''], shuffle=snake_case_, collate_fn=snake_case_, batch_size=snake_case_ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders UpperCamelCase__ : int = mocked_dataloaders # noqa: F811 def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ ) -> List[Any]: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''', snake_case_ ) == "1": a = 2 # Initialize accelerator a = Accelerator(cpu=args.cpu, mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs a = config['''lr'''] a = int(config['''num_epochs'''] ) a = int(config['''seed'''] ) a = int(config['''batch_size'''] ) a = evaluate.load('''glue''', '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=snake_case_ ) def inner_training_loop(snake_case_ ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(snake_case_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) a = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''', return_dict=snake_case_ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). a = model.to(accelerator.device ) # Instantiate optimizer a = AdamW(params=model.parameters(), lr=snake_case_ ) a , a = get_dataloaders(snake_case_, snake_case_ ) # Instantiate scheduler a = get_linear_schedule_with_warmup( optimizer=snake_case_, num_warmup_steps=1_0_0, num_training_steps=(len(snake_case_ ) * num_epochs), ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. a , a , a , a , a = accelerator.prepare( snake_case_, snake_case_, snake_case_, snake_case_, snake_case_ ) # Now we train the model for epoch in range(snake_case_ ): model.train() for step, batch in enumerate(snake_case_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) a = model(**snake_case_ ) a = outputs.loss accelerator.backward(snake_case_ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(snake_case_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): a = model(**snake_case_ ) a = outputs.logits.argmax(dim=-1 ) a , a = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=snake_case_, references=snake_case_, ) a = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""", snake_case_ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def SCREAMING_SNAKE_CASE__ ( ) -> Tuple: """simple docstring""" a = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''', type=snake_case_, default=snake_case_, choices=['''no''', '''fp16''', '''bf16''', '''fp8'''], help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''', ) parser.add_argument('''--cpu''', action='''store_true''', help='''If passed, will train on the CPU.''' ) a = parser.parse_args() a = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 4_2, '''batch_size''': 1_6} training_function(snake_case_, snake_case_ ) if __name__ == "__main__": main()
330
def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ ) -> int: """simple docstring""" a = '''''' for i in table: res += inp[i - 1] return res def SCREAMING_SNAKE_CASE__ ( snake_case_ ) -> int: """simple docstring""" return data[1:] + data[0] def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ ) -> List[str]: """simple docstring""" a = '''''' for i in range(len(snake_case_ ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ ) -> Dict: """simple docstring""" a = int('''0b''' + data[0] + data[-1], 2 ) a = int('''0b''' + data[1:3], 2 ) return bin(s[row][col] )[2:] def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_, snake_case_, snake_case_, snake_case_ ) -> Optional[int]: """simple docstring""" a = message[:4] a = message[4:] a = apply_table(snake_case_, snake_case_ ) a = xor(snake_case_, snake_case_ ) a = apply_sbox(snake_case_, temp[:4] ) # noqa: E741 a = apply_sbox(snake_case_, temp[4:] ) a = '''0''' * (2 - len(snake_case_ )) + l # noqa: E741 a = '''0''' * (2 - len(snake_case_ )) + r a = apply_table(l + r, snake_case_ ) a = xor(snake_case_, snake_case_ ) return temp + right if __name__ == "__main__": UpperCamelCase__ : int = input("""Enter 10 bit key: """) UpperCamelCase__ : Union[str, Any] = input("""Enter 8 bit message: """) UpperCamelCase__ : Dict = [6, 3, 7, 4, 8, 5, 10, 9] UpperCamelCase__ : Union[str, Any] = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] UpperCamelCase__ : Optional[int] = [2, 4, 3, 1] UpperCamelCase__ : List[Any] = [2, 6, 3, 1, 4, 8, 5, 7] UpperCamelCase__ : str = [4, 1, 3, 5, 7, 2, 8, 6] UpperCamelCase__ : List[Any] = [4, 1, 2, 3, 2, 3, 4, 1] UpperCamelCase__ : int = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] UpperCamelCase__ : Dict = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation UpperCamelCase__ : Optional[Any] = apply_table(key, paa_table) UpperCamelCase__ : str = temp[:5] UpperCamelCase__ : List[Any] = temp[5:] UpperCamelCase__ : Dict = left_shift(left) UpperCamelCase__ : Any = left_shift(right) UpperCamelCase__ : Optional[Any] = apply_table(left + right, pa_table) UpperCamelCase__ : List[str] = left_shift(left) UpperCamelCase__ : int = left_shift(right) UpperCamelCase__ : List[str] = left_shift(left) UpperCamelCase__ : Dict = left_shift(right) UpperCamelCase__ : List[str] = apply_table(left + right, pa_table) # encryption UpperCamelCase__ : Tuple = apply_table(message, IP) UpperCamelCase__ : Optional[Any] = function(expansion, sa, sa, keya, temp) UpperCamelCase__ : Optional[int] = temp[4:] + temp[:4] UpperCamelCase__ : Any = function(expansion, sa, sa, keya, temp) UpperCamelCase__ : Tuple = apply_table(temp, IP_inv) print("""Cipher text is:""", CT) # decryption UpperCamelCase__ : Union[str, Any] = apply_table(CT, IP) UpperCamelCase__ : List[str] = function(expansion, sa, sa, keya, temp) UpperCamelCase__ : Optional[Any] = temp[4:] + temp[:4] UpperCamelCase__ : Optional[int] = function(expansion, sa, sa, keya, temp) UpperCamelCase__ : Any = apply_table(temp, IP_inv) print("""Plain text after decypting is:""", PT)
330
1
import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin __UpperCAmelCase = get_tests_dir('fixtures/test_sentencepiece_bpe_char.model') @require_sentencepiece @require_tokenizers class lowerCamelCase (_snake_case , unittest.TestCase ): '''simple docstring''' _snake_case : int = SpeechTaTokenizer _snake_case : int = False _snake_case : Dict = True def __UpperCAmelCase ( self ) -> str: super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase_ : Optional[Any] = SpeechTaTokenizer(_UpperCamelCase ) UpperCAmelCase_ : List[Any] = AddedToken('<mask>' , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) UpperCAmelCase_ : Optional[int] = mask_token tokenizer.add_special_tokens({'mask_token': mask_token} ) tokenizer.add_tokens(['<ctc_blank>'] ) tokenizer.save_pretrained(self.tmpdirname ) def __UpperCAmelCase ( self , _UpperCamelCase ) -> int: UpperCAmelCase_ : Dict = 'this is a test' UpperCAmelCase_ : Optional[Any] = 'this is a test' return input_text, output_text def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase=False , _UpperCamelCase=2_0 , _UpperCamelCase=5 ) -> Optional[int]: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = self.get_input_output_texts(_UpperCamelCase ) UpperCAmelCase_ : Optional[Any] = tokenizer.encode(_UpperCamelCase , add_special_tokens=_UpperCamelCase ) UpperCAmelCase_ : str = tokenizer.decode(_UpperCamelCase , clean_up_tokenization_spaces=_UpperCamelCase ) return text, ids def __UpperCAmelCase ( self ) -> Any: UpperCAmelCase_ : Optional[int] = '<pad>' UpperCAmelCase_ : Any = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCamelCase ) , _UpperCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCamelCase ) , _UpperCamelCase ) def __UpperCAmelCase ( self ) -> Any: UpperCAmelCase_ : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-4] , 'œ' ) self.assertEqual(vocab_keys[-2] , '<mask>' ) self.assertEqual(vocab_keys[-1] , '<ctc_blank>' ) self.assertEqual(len(_UpperCamelCase ) , 8_1 ) def __UpperCAmelCase ( self ) -> Union[str, Any]: self.assertEqual(self.get_tokenizer().vocab_size , 7_9 ) def __UpperCAmelCase ( self ) -> int: UpperCAmelCase_ : Optional[int] = self.get_tokenizers(do_lower_case=_UpperCamelCase ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): UpperCAmelCase_ : int = tokenizer.vocab_size UpperCAmelCase_ : int = len(_UpperCamelCase ) self.assertNotEqual(_UpperCamelCase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) UpperCAmelCase_ : List[str] = ['aaaaa bbbbbb', 'cccccccccdddddddd'] UpperCAmelCase_ : Dict = tokenizer.add_tokens(_UpperCamelCase ) UpperCAmelCase_ : Optional[Any] = tokenizer.vocab_size UpperCAmelCase_ : Any = len(_UpperCamelCase ) self.assertNotEqual(_UpperCamelCase , 0 ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) self.assertEqual(_UpperCamelCase , len(_UpperCamelCase ) ) self.assertEqual(_UpperCamelCase , all_size + len(_UpperCamelCase ) ) UpperCAmelCase_ : Dict = tokenizer.encode('aaaaa bbbbbb low cccccccccdddddddd l' , add_special_tokens=_UpperCamelCase ) self.assertGreaterEqual(len(_UpperCamelCase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) UpperCAmelCase_ : Tuple = {'eos_token': '>>>>|||<||<<|<<', 'pad_token': '<<<<<|||>|>>>>|>'} UpperCAmelCase_ : Tuple = tokenizer.add_special_tokens(_UpperCamelCase ) UpperCAmelCase_ : Dict = tokenizer.vocab_size UpperCAmelCase_ : List[str] = len(_UpperCamelCase ) self.assertNotEqual(_UpperCamelCase , 0 ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) self.assertEqual(_UpperCamelCase , len(_UpperCamelCase ) ) self.assertEqual(_UpperCamelCase , all_size_a + len(_UpperCamelCase ) ) UpperCAmelCase_ : Any = tokenizer.encode( '>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l' , add_special_tokens=_UpperCamelCase ) self.assertGreaterEqual(len(_UpperCamelCase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) def __UpperCAmelCase ( self ) -> Any: pass def __UpperCAmelCase ( self ) -> str: pass def __UpperCAmelCase ( self ) -> List[Any]: UpperCAmelCase_ : str = self.get_tokenizer() UpperCAmelCase_ : Optional[int] = tokenizer.tokenize('This is a test' ) # fmt: off self.assertListEqual(_UpperCamelCase , [SPIECE_UNDERLINE, 'T', 'h', 'i', 's', SPIECE_UNDERLINE, 'i', 's', SPIECE_UNDERLINE, 'a', SPIECE_UNDERLINE, 't', 'e', 's', 't'] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(_UpperCamelCase ) , [4, 3_2, 1_1, 1_0, 1_2, 4, 1_0, 1_2, 4, 7, 4, 6, 5, 1_2, 6] , ) UpperCAmelCase_ : Optional[Any] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _UpperCamelCase , [SPIECE_UNDERLINE, 'I', SPIECE_UNDERLINE, 'w', 'a', 's', SPIECE_UNDERLINE, 'b', 'o', 'r', 'n', SPIECE_UNDERLINE, 'i', 'n', SPIECE_UNDERLINE, '92000', ',', SPIECE_UNDERLINE, 'a', 'n', 'd', SPIECE_UNDERLINE, 't', 'h', 'i', 's', SPIECE_UNDERLINE, 'i', 's', SPIECE_UNDERLINE, 'f', 'a', 'l', 's', 'é', '.'] ) UpperCAmelCase_ : Any = tokenizer.convert_tokens_to_ids(_UpperCamelCase ) # fmt: off self.assertListEqual(_UpperCamelCase , [4, 3_0, 4, 2_0, 7, 1_2, 4, 2_5, 8, 1_3, 9, 4, 1_0, 9, 4, 3, 2_3, 4, 7, 9, 1_4, 4, 6, 1_1, 1_0, 1_2, 4, 1_0, 1_2, 4, 1_9, 7, 1_5, 1_2, 7_3, 2_6] ) # fmt: on UpperCAmelCase_ : Union[str, Any] = tokenizer.convert_ids_to_tokens(_UpperCamelCase ) self.assertListEqual( _UpperCamelCase , [SPIECE_UNDERLINE, 'I', SPIECE_UNDERLINE, 'w', 'a', 's', SPIECE_UNDERLINE, 'b', 'o', 'r', 'n', SPIECE_UNDERLINE, 'i', 'n', SPIECE_UNDERLINE, '<unk>', ',', SPIECE_UNDERLINE, 'a', 'n', 'd', SPIECE_UNDERLINE, 't', 'h', 'i', 's', SPIECE_UNDERLINE, 'i', 's', SPIECE_UNDERLINE, 'f', 'a', 'l', 's', 'é', '.'] ) @slow def __UpperCAmelCase ( self ) -> Union[str, Any]: # Use custom sequence because this tokenizer does not handle numbers. UpperCAmelCase_ : Optional[Any] = [ 'Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides ' 'general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural ' 'Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained ' 'models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.', 'BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly ' 'conditioning on both left and right context in all layers.', 'The quick brown fox jumps over the lazy dog.', ] # fmt: off UpperCAmelCase_ : Dict = { 'input_ids': [ [4, 3_2, 1_3, 7, 9, 1_2, 1_9, 8, 1_3, 1_8, 5, 1_3, 1_2, 4, 6_4, 1_9, 8, 1_3, 1_8, 5, 1_3, 1_5, 2_2, 4, 2_8, 9, 8, 2_0, 9, 4, 7, 1_2, 4, 2_4, 2_2, 6, 8, 1_3, 1_7, 1_1, 3_9, 6, 1_3, 7, 9, 1_2, 1_9, 8, 1_3, 1_8, 5, 1_3, 1_2, 4, 7, 9, 1_4, 4, 2_4, 2_2, 6, 8, 1_3, 1_7, 1_1, 3_9, 2_4, 1_3, 5, 6, 1_3, 7, 1_0, 9, 5, 1_4, 3_9, 2_5, 5, 1_3, 6, 6_3, 4, 2_4, 1_3, 8, 2_7, 1_0, 1_4, 5, 1_2, 4, 2_1, 5, 9, 5, 1_3, 7, 1_5, 3_9, 2_4, 1_6, 1_3, 2_4, 8, 1_2, 5, 4, 7, 1_3, 1_7, 1_1, 1_0, 6, 5, 1_7, 6, 1_6, 1_3, 5, 1_2, 4, 6_4, 4_0, 4_7, 5_4, 3_2, 2_3, 4, 5_3, 4_9, 3_2, 2_3, 4, 5_4, 8, 4_0, 4_7, 5_4, 3_2, 7, 2_3, 4, 6_9, 5_2, 4_3, 2_3, 4, 5_1, 1_0, 1_2, 6, 1_0, 1_5, 4_0, 5, 1_3, 6, 2_3, 4, 6_9, 5_2, 4_8, 5, 6, 2_6, 2_6, 2_6, 6_3, 4, 1_9, 8, 1_3, 4, 4_8, 7, 6, 1_6, 1_3, 7, 1_5, 4, 5_2, 7, 9, 2_1, 1_6, 7, 2_1, 5, 4, 6_1, 9, 1_4, 5, 1_3, 1_2, 6, 7, 9, 1_4, 1_0, 9, 2_1, 4, 6_4, 4_8, 5_2, 6_1, 6_3, 4, 7, 9, 1_4, 4, 4_8, 7, 6, 1_6, 1_3, 7, 1_5, 4, 5_2, 7, 9, 2_1, 1_6, 7, 2_1, 5, 4, 5_3, 5, 9, 5, 1_3, 7, 6, 1_0, 8, 9, 4, 6_4, 4_8, 5_2, 5_3, 6_3, 4, 2_0, 1_0, 6, 1_1, 4, 8, 2_7, 5, 1_3, 4, 6, 1_1, 1_0, 1_3, 6, 2_2, 3_9, 6, 2_0, 8, 4, 2_4, 1_3, 5, 6, 1_3, 7, 1_0, 9, 5, 1_4, 4, 1_8, 8, 1_4, 5, 1_5, 1_2, 4, 1_0, 9, 4, 8, 9, 5, 4, 1_1, 1_6, 9, 1_4, 1_3, 5, 1_4, 4, 2_4, 1_5, 1_6, 1_2, 4, 1_5, 7, 9, 2_1, 1_6, 7, 2_1, 5, 1_2, 4, 7, 9, 1_4, 4, 1_4, 5, 5, 2_4, 4, 1_0, 9, 6, 5, 1_3, 8, 2_4, 5, 1_3, 7, 2_5, 1_0, 1_5, 1_0, 6, 2_2, 4, 2_5, 5, 6, 2_0, 5, 5, 9, 4, 5_8, 7, 3_7, 2_3, 4, 4_9, 2_2, 3_2, 8, 1_3, 1_7, 1_1, 4, 7, 9, 1_4, 4, 3_2, 5, 9, 1_2, 8, 1_3, 5_5, 1_5, 8, 2_0, 2_6, 2], [4, 4_0, 4_7, 5_4, 3_2, 4, 1_0, 1_2, 4, 1_4, 5, 1_2, 1_0, 2_1, 9, 5, 1_4, 4, 6, 8, 4, 2_4, 1_3, 5, 3_9, 6, 1_3, 7, 1_0, 9, 4, 1_4, 5, 5, 2_4, 4, 2_5, 1_0, 1_4, 1_0, 1_3, 5, 1_7, 6, 1_0, 8, 9, 7, 1_5, 4, 1_3, 5, 2_4, 1_3, 5, 1_2, 5, 9, 6, 7, 6, 1_0, 8, 9, 1_2, 4, 1_9, 1_3, 8, 1_8, 4, 1_6, 9, 1_5, 7, 2_5, 5, 1_5, 5, 1_4, 4, 6, 5, 3_7, 6, 4, 2_5, 2_2, 4, 4_6, 8, 1_0, 9, 6, 1_5, 2_2, 4, 1_7, 8, 9, 1_4, 1_0, 6, 1_0, 8, 9, 1_0, 9, 2_1, 4, 8, 9, 4, 2_5, 8, 6, 1_1, 4, 1_5, 5, 1_9, 6, 4, 7, 9, 1_4, 4, 1_3, 1_0, 2_1, 1_1, 6, 4, 1_7, 8, 9, 6, 5, 3_7, 6, 4, 1_0, 9, 4, 7, 1_5, 1_5, 4, 1_5, 7, 2_2, 5, 1_3, 1_2, 2_6, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 3_2, 1_1, 5, 4, 4_5, 1_6, 1_0, 1_7, 2_8, 4, 2_5, 1_3, 8, 2_0, 9, 4, 1_9, 8, 3_7, 4, 4_6, 1_6, 1_8, 2_4, 1_2, 4, 8, 2_7, 5, 1_3, 4, 6, 1_1, 5, 4, 1_5, 7, 5_7, 2_2, 4, 1_4, 8, 2_1, 2_6, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], 'attention_mask': [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=_UpperCamelCase , model_name='microsoft/speecht5_asr' , revision='c5ef64c71905caeccde0e4462ef3f9077224c524' , sequences=_UpperCamelCase , )
29
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = {'configuration_vit_msn': ['VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ViTMSNConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST', 'ViTMSNModel', 'ViTMSNForImageClassification', 'ViTMSNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
29
1
UpperCAmelCase_ = [ 'VerificationMode', 'Version', 'disable_progress_bar', 'enable_progress_bar', 'is_progress_bar_enabled', 'experimental', ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
355
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore UpperCAmelCase_ = '\nHuman: <<task>>\n\nAssistant: ' UpperCAmelCase_ = 'huggingface-tools/default-prompts' UpperCAmelCase_ = {'chat': 'chat_prompt_template.txt', 'run': 'run_prompt_template.txt'} def lowerCAmelCase_ ( __UpperCAmelCase: Optional[Any] , __UpperCAmelCase: List[Any] , __UpperCAmelCase: Optional[Any]="run" ) -> int: if prompt_or_repo_id is None: UpperCamelCase__ : List[Any] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search('''\\s''' , __UpperCAmelCase ) is not None: return prompt_or_repo_id UpperCamelCase__ : Any = cached_file( __UpperCAmelCase , PROMPT_FILES[mode] , repo_type='''dataset''' , user_agent={'''agent''': agent_name} ) with open(__UpperCAmelCase , '''r''' , encoding='''utf-8''' ) as f: return f.read()
247
0
from __future__ import annotations lowerCAmelCase__ = 1.6021e-19 # units = C def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError("You cannot supply more or less than 2 values" ) elif conductivity < 0: raise ValueError("Conductivity cannot be negative" ) elif electron_conc < 0: raise ValueError("Electron concentration cannot be negative" ) elif mobility < 0: raise ValueError("mobility cannot be negative" ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
130
import os # Precomputes a list of the 100 first triangular numbers lowerCAmelCase__ = [int(0.5 * n * (n + 1)) for n in range(1, 1_0_1)] def __lowerCamelCase ( ): """simple docstring""" lowercase__ : str = os.path.dirname(os.path.realpath(lowerCamelCase__ ) ) lowercase__ : Optional[Any] = os.path.join(lowerCamelCase__ , "words.txt" ) lowercase__ : int = "" with open(lowerCamelCase__ ) as f: lowercase__ : Any = f.readline() lowercase__ : Any = [word.strip("\"" ) for word in words.strip("\r\n" ).split("," )] lowercase__ : str = [ word for word in [sum(ord(lowerCamelCase__ ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(lowerCamelCase__ ) if __name__ == "__main__": print(solution())
130
1
'''simple docstring''' from manim import * class snake_case__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" def lowerCAmelCase ( self : Dict ) -> Tuple: """simple docstring""" snake_case : Any = Rectangle(height=0.5 , width=0.5 ) snake_case : Union[str, Any] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) snake_case : Optional[int] = [mem.copy() for i in range(6 )] snake_case : List[Any] = [mem.copy() for i in range(6 )] snake_case : Optional[Any] = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) snake_case : List[Any] = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) snake_case : int = VGroup(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) snake_case : List[str] = Text('''CPU''' , font_size=24 ) snake_case : Tuple = Group(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0.5 , aligned_edge=UpperCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(UpperCamelCase__ ) snake_case : Union[str, Any] = [mem.copy() for i in range(1 )] snake_case : Any = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) snake_case : Any = Text('''GPU''' , font_size=24 ) snake_case : str = Group(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0.5 , aligned_edge=UpperCamelCase__ ) gpu.align_to(UpperCamelCase__ , UpperCamelCase__ ) gpu.set_x(gpu.get_x() - 1 ) self.add(UpperCamelCase__ ) snake_case : Optional[Any] = [mem.copy() for i in range(6 )] snake_case : Union[str, Any] = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) snake_case : Optional[Any] = Text('''Model''' , font_size=24 ) snake_case : Optional[int] = Group(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0.5 , aligned_edge=UpperCamelCase__ ) model.move_to([3, -1.0, 0] ) self.play( Create(UpperCamelCase__ , run_time=1 ) , Create(UpperCamelCase__ , run_time=1 ) , Create(UpperCamelCase__ , run_time=1 ) , ) snake_case : List[str] = MarkupText( f'First, an empty model skeleton is loaded\ninto <span fgcolor=\'{YELLOW}\'>memory</span> without using much RAM.' , font_size=24 , ) snake_case : Any = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) snake_case : Dict = MarkupText( f'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCamelCase__ , run_time=2.5 ) , Write(UpperCamelCase__ ) , Write(UpperCamelCase__ ) ) self.add(UpperCamelCase__ ) snake_case : Union[str, Any] = [] snake_case : str = [] snake_case : int = [] for i, rect in enumerate(UpperCamelCase__ ): snake_case : str = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(UpperCamelCase__ , opacity=0.7 ) cpu_target.move_to(UpperCamelCase__ ) cpu_target.generate_target() snake_case : Optional[int] = 0.46 / 4 snake_case : int = 0.46 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=UpperCamelCase__ ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=UpperCamelCase__ , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=UpperCamelCase__ , buff=0.0 ) cpu_targs.append(UpperCamelCase__ ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(UpperCamelCase__ ) ) second_animations.append(MoveToTarget(UpperCamelCase__ , run_time=1.5 ) ) self.play(*UpperCamelCase__ ) self.play(*UpperCamelCase__ ) self.wait()
83
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class snake_case__ ( unittest.TestCase ): """simple docstring""" lowerCamelCase = StableDiffusionLDMaDPipeline lowerCamelCase = TEXT_TO_IMAGE_PARAMS lowerCamelCase = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase = TEXT_TO_IMAGE_IMAGE_PARAMS def lowerCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0 ) snake_case : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) snake_case : Dict = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=UpperCamelCase__ , set_alpha_to_one=UpperCamelCase__ , ) torch.manual_seed(0 ) snake_case : int = AutoencoderKL( block_out_channels=[32, 64] , in_channels=6 , out_channels=6 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0 ) snake_case : Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) snake_case : int = CLIPTextModel(UpperCamelCase__ ) snake_case : Tuple = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case : str = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def lowerCAmelCase ( self : Optional[int] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str=0 ) -> Optional[int]: """simple docstring""" if str(UpperCamelCase__ ).startswith('''mps''' ): snake_case : Optional[int] = torch.manual_seed(UpperCamelCase__ ) else: snake_case : Optional[int] = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) snake_case : Tuple = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def lowerCAmelCase ( self : Any ) -> str: """simple docstring""" snake_case : int = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case : List[Any] = self.get_dummy_components() snake_case : str = StableDiffusionLDMaDPipeline(**UpperCamelCase__ ) snake_case : Union[str, Any] = ldmad_pipe.to(UpperCamelCase__ ) ldmad_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : List[str] = self.get_dummy_inputs(UpperCamelCase__ ) snake_case : Tuple = ldmad_pipe(**UpperCamelCase__ ) snake_case ,snake_case : int = output.rgb, output.depth snake_case : str = rgb[0, -3:, -3:, -1] snake_case : Union[str, Any] = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) snake_case : int = np.array( [0.37_338_176, 0.70_247, 0.74_203_193, 0.51_643_604, 0.58_256_793, 0.60_932_136, 0.4_181_095, 0.48_355_877, 0.46_535_262] ) snake_case : str = np.array([103.46_727, 85.812_004, 87.849_236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1e-2 def lowerCAmelCase ( self : Dict ) -> int: """simple docstring""" snake_case : int = self.get_dummy_components() snake_case : Any = StableDiffusionLDMaDPipeline(**UpperCamelCase__ ) snake_case : Optional[Any] = ldmad_pipe.to(UpperCamelCase__ ) ldmad_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : int = self.get_dummy_inputs(UpperCamelCase__ ) snake_case : str = 3 * [inputs['''prompt''']] # forward snake_case : Union[str, Any] = ldmad_pipe(**UpperCamelCase__ ) snake_case ,snake_case : Union[str, Any] = output.rgb, output.depth snake_case : Tuple = rgb_slice_a[0, -3:, -3:, -1] snake_case : List[str] = depth_slice_a[0, -3:, -1] snake_case : int = self.get_dummy_inputs(UpperCamelCase__ ) snake_case : Optional[int] = 3 * [inputs.pop('''prompt''' )] snake_case : Dict = ldmad_pipe.tokenizer( UpperCamelCase__ , padding='''max_length''' , max_length=ldmad_pipe.tokenizer.model_max_length , truncation=UpperCamelCase__ , return_tensors='''pt''' , ) snake_case : Optional[Any] = text_inputs['''input_ids'''].to(UpperCamelCase__ ) snake_case : Any = ldmad_pipe.text_encoder(UpperCamelCase__ )[0] snake_case : Tuple = prompt_embeds # forward snake_case : List[Any] = ldmad_pipe(**UpperCamelCase__ ) snake_case ,snake_case : Dict = output.rgb, output.depth snake_case : Any = rgb_slice_a[0, -3:, -3:, -1] snake_case : Dict = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1e-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1e-4 def lowerCAmelCase ( self : str ) -> List[Any]: """simple docstring""" snake_case : str = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case : Dict = self.get_dummy_components() snake_case : List[str] = PNDMScheduler(skip_prk_steps=UpperCamelCase__ ) snake_case : Optional[Any] = StableDiffusionLDMaDPipeline(**UpperCamelCase__ ) snake_case : Dict = ldmad_pipe.to(UpperCamelCase__ ) ldmad_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Dict = self.get_dummy_inputs(UpperCamelCase__ ) snake_case : str = '''french fries''' snake_case : List[str] = ldmad_pipe(**UpperCamelCase__ , negative_prompt=UpperCamelCase__ ) snake_case ,snake_case : Union[str, Any] = output.rgb, output.depth snake_case : Union[str, Any] = rgb[0, -3:, -3:, -1] snake_case : int = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) snake_case : Dict = np.array( [0.37_044, 0.71_811_503, 0.7_223_251, 0.48_603_675, 0.5_638_391, 0.6_364_948, 0.42_833_704, 0.4_901_315, 0.47_926_217] ) snake_case : Any = np.array([107.84_738, 84.62_802, 89.962_135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1e-2 @slow @require_torch_gpu class snake_case__ ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase ( self : Optional[int] ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Union[str, Any]="cpu" , UpperCamelCase__ : Optional[int]=torch.floataa , UpperCamelCase__ : Union[str, Any]=0 ) -> Any: """simple docstring""" snake_case : Any = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) snake_case : Optional[Any] = np.random.RandomState(UpperCamelCase__ ).standard_normal((1, 4, 64, 64) ) snake_case : Any = torch.from_numpy(UpperCamelCase__ ).to(device=UpperCamelCase__ , dtype=UpperCamelCase__ ) snake_case : List[Any] = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def lowerCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" snake_case : Dict = StableDiffusionLDMaDPipeline.from_pretrained('''Intel/ldm3d''' ) snake_case : Optional[Any] = ldmad_pipe.to(UpperCamelCase__ ) ldmad_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Any = self.get_inputs(UpperCamelCase__ ) snake_case : str = ldmad_pipe(**UpperCamelCase__ ) snake_case ,snake_case : List[Any] = output.rgb, output.depth snake_case : Optional[Any] = rgb[0, -3:, -3:, -1].flatten() snake_case : Tuple = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 512, 512, 3) assert depth.shape == (1, 512, 512) snake_case : Optional[Any] = np.array( [0.53_805_465, 0.56_707_305, 0.5_486_515, 0.57_012_236, 0.5_814_511, 0.56_253_487, 0.54_843_014, 0.55_092_263, 0.6_459_706] ) snake_case : str = np.array( [0.9_263_781, 0.6_678_672, 0.5_486_515, 0.92_202_145, 0.67_831_135, 0.56_253_487, 0.9_241_694, 0.7_551_478, 0.6_459_706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3e-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3e-3 @nightly @require_torch_gpu class snake_case__ ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase ( self : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Any="cpu" , UpperCamelCase__ : Optional[int]=torch.floataa , UpperCamelCase__ : Optional[int]=0 ) -> str: """simple docstring""" snake_case : Dict = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) snake_case : Optional[Any] = np.random.RandomState(UpperCamelCase__ ).standard_normal((1, 4, 64, 64) ) snake_case : int = torch.from_numpy(UpperCamelCase__ ).to(device=UpperCamelCase__ , dtype=UpperCamelCase__ ) snake_case : List[str] = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 50, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def lowerCAmelCase ( self : Optional[Any] ) -> Tuple: """simple docstring""" snake_case : Optional[Any] = StableDiffusionLDMaDPipeline.from_pretrained('''Intel/ldm3d''' ).to(UpperCamelCase__ ) ldmad_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : Dict = self.get_inputs(UpperCamelCase__ ) snake_case : str = ldmad_pipe(**UpperCamelCase__ ) snake_case ,snake_case : Union[str, Any] = output.rgb, output.depth snake_case : Union[str, Any] = 0.495_586 snake_case : Tuple = 0.33_795_515 snake_case : Dict = 112.48_518 snake_case : Optional[int] = 98.489_746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3 def lowerCAmelCase ( self : Dict ) -> Dict: """simple docstring""" snake_case : Optional[int] = StableDiffusionLDMaDPipeline.from_pretrained('''Intel/ldm3d-4c''' ).to(UpperCamelCase__ ) ldmad_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) snake_case : int = self.get_inputs(UpperCamelCase__ ) snake_case : List[Any] = ldmad_pipe(**UpperCamelCase__ ) snake_case ,snake_case : Union[str, Any] = output.rgb, output.depth snake_case : Tuple = 0.4_194_127 snake_case : Optional[Any] = 0.35_375_586 snake_case : Any = 0.5_638_502 snake_case : int = 0.34_686_103 assert rgb.shape == (1, 512, 512, 3) assert depth.shape == (1, 512, 512, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3
83
1
'''simple docstring''' import argparse import math import traceback import dateutil.parser as date_parser import requests def UpperCamelCase_( snake_case : int ): '''simple docstring''' snake_case_ = {} snake_case_ = job["started_at"] snake_case_ = job["completed_at"] snake_case_ = date_parser.parse(snake_case ) snake_case_ = date_parser.parse(snake_case ) snake_case_ = round((end_datetime - start_datetime).total_seconds() / 60.0 ) snake_case_ = start snake_case_ = end snake_case_ = duration_in_min return job_info def UpperCamelCase_( snake_case : Union[str, Any] , snake_case : Union[str, Any]=None ): '''simple docstring''' snake_case_ = None if token is not None: snake_case_ = {"Accept": "application/vnd.github+json", "Authorization": f'Bearer {token}'} snake_case_ = f'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' snake_case_ = requests.get(snake_case , headers=snake_case ).json() snake_case_ = {} try: job_time.update({job["name"]: extract_time_from_single_job(snake_case ) for job in result["jobs"]} ) snake_case_ = math.ceil((result["total_count"] - 1_0_0) / 1_0_0 ) for i in range(snake_case ): snake_case_ = requests.get(url + f'&page={i + 2}' , headers=snake_case ).json() job_time.update({job["name"]: extract_time_from_single_job(snake_case ) for job in result["jobs"]} ) return job_time except Exception: print(f'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("--workflow_run_id", type=str, required=True, help="A GitHub Actions workflow run id.") _SCREAMING_SNAKE_CASE : int = parser.parse_args() _SCREAMING_SNAKE_CASE : List[str] = get_job_time(args.workflow_run_id) _SCREAMING_SNAKE_CASE : Optional[int] = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(F"{k}: {v['duration']}")
85
'''simple docstring''' # Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _SCREAMING_SNAKE_CASE : List[str] = { "configuration_cpmant": ["CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP", "CpmAntConfig"], "tokenization_cpmant": ["CpmAntTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : Any = [ "CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST", "CpmAntForCausalLM", "CpmAntModel", "CpmAntPreTrainedModel", ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
85
1
'''simple docstring''' import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig A__ : Optional[Any] =logging.get_logger(__name__) # General docstring A__ : Dict ='''PoolFormerConfig''' # Base docstring A__ : List[str] ='''sail/poolformer_s12''' A__ : str =[1, 5_12, 7, 7] # Image classification docstring A__ : Any ='''sail/poolformer_s12''' A__ : str ='''tabby, tabby cat''' A__ : Optional[int] =[ '''sail/poolformer_s12''', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase = 0.0 , lowerCAmelCase = False ): """simple docstring""" if drop_prob == 0.0 or not training: return input _lowerCAmelCase = 1 - drop_prob _lowerCAmelCase = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets _lowerCAmelCase = keep_prob + torch.rand(lowerCAmelCase , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize _lowerCAmelCase = input.div(lowerCAmelCase ) * random_tensor return output class UpperCAmelCase ( nn.Module ): def __init__( self : Optional[int] , __snake_case : Optional[float] = None ) -> None: super().__init__() _lowerCAmelCase = drop_prob def lowercase__ ( self : Tuple , __snake_case : torch.Tensor ) -> torch.Tensor: return drop_path(__snake_case , self.drop_prob , self.training ) def lowercase__ ( self : Union[str, Any] ) -> str: return "p={}".format(self.drop_prob ) class UpperCAmelCase ( nn.Module ): def __init__( self : Optional[Any] , __snake_case : Tuple , __snake_case : Any , __snake_case : Optional[Any] , __snake_case : Dict , __snake_case : int , __snake_case : Optional[Any]=None ) -> Dict: super().__init__() _lowerCAmelCase = patch_size if isinstance(__snake_case , collections.abc.Iterable ) else (patch_size, patch_size) _lowerCAmelCase = stride if isinstance(__snake_case , collections.abc.Iterable ) else (stride, stride) _lowerCAmelCase = padding if isinstance(__snake_case , collections.abc.Iterable ) else (padding, padding) _lowerCAmelCase = nn.Convad(__snake_case , __snake_case , kernel_size=__snake_case , stride=__snake_case , padding=__snake_case ) _lowerCAmelCase = norm_layer(__snake_case ) if norm_layer else nn.Identity() def lowercase__ ( self : str , __snake_case : Dict ) -> int: _lowerCAmelCase = self.projection(__snake_case ) _lowerCAmelCase = self.norm(__snake_case ) return embeddings class UpperCAmelCase ( nn.GroupNorm ): def __init__( self : int , __snake_case : Tuple , **__snake_case : str ) -> Dict: super().__init__(1 , __snake_case , **__snake_case ) class UpperCAmelCase ( nn.Module ): def __init__( self : Optional[Any] , __snake_case : Optional[Any] ) -> int: super().__init__() _lowerCAmelCase = nn.AvgPoolad(__snake_case , stride=1 , padding=pool_size // 2 , count_include_pad=__snake_case ) def lowercase__ ( self : List[str] , __snake_case : str ) -> List[str]: return self.pool(__snake_case ) - hidden_states class UpperCAmelCase ( nn.Module ): def __init__( self : Dict , __snake_case : Union[str, Any] , __snake_case : List[Any] , __snake_case : Dict , __snake_case : Optional[Any] ) -> List[str]: super().__init__() _lowerCAmelCase = nn.Convad(__snake_case , __snake_case , 1 ) _lowerCAmelCase = nn.Convad(__snake_case , __snake_case , 1 ) _lowerCAmelCase = PoolFormerDropPath(__snake_case ) if isinstance(config.hidden_act , __snake_case ): _lowerCAmelCase = ACTaFN[config.hidden_act] else: _lowerCAmelCase = config.hidden_act def lowercase__ ( self : Union[str, Any] , __snake_case : Optional[int] ) -> Any: _lowerCAmelCase = self.conva(__snake_case ) _lowerCAmelCase = self.act_fn(__snake_case ) _lowerCAmelCase = self.drop(__snake_case ) _lowerCAmelCase = self.conva(__snake_case ) _lowerCAmelCase = self.drop(__snake_case ) return hidden_states class UpperCAmelCase ( nn.Module ): def __init__( self : str , __snake_case : str , __snake_case : int , __snake_case : Dict , __snake_case : str , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> str: super().__init__() _lowerCAmelCase = PoolFormerPooling(__snake_case ) _lowerCAmelCase = PoolFormerOutput(__snake_case , __snake_case , __snake_case , __snake_case ) _lowerCAmelCase = PoolFormerGroupNorm(__snake_case ) _lowerCAmelCase = PoolFormerGroupNorm(__snake_case ) # Useful for training neural nets _lowerCAmelCase = PoolFormerDropPath(__snake_case ) if drop_path > 0.0 else nn.Identity() _lowerCAmelCase = config.use_layer_scale if config.use_layer_scale: _lowerCAmelCase = nn.Parameter( config.layer_scale_init_value * torch.ones((__snake_case) ) , requires_grad=__snake_case ) _lowerCAmelCase = nn.Parameter( config.layer_scale_init_value * torch.ones((__snake_case) ) , requires_grad=__snake_case ) def lowercase__ ( self : Any , __snake_case : Union[str, Any] ) -> Optional[int]: if self.use_layer_scale: _lowerCAmelCase = self.pooling(self.before_norm(__snake_case ) ) _lowerCAmelCase = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection _lowerCAmelCase = hidden_states + self.drop_path(__snake_case ) _lowerCAmelCase = () _lowerCAmelCase = self.output(self.after_norm(__snake_case ) ) _lowerCAmelCase = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection _lowerCAmelCase = hidden_states + self.drop_path(__snake_case ) _lowerCAmelCase = (output,) + outputs return outputs else: _lowerCAmelCase = self.drop_path(self.pooling(self.before_norm(__snake_case ) ) ) # First residual connection _lowerCAmelCase = pooling_output + hidden_states _lowerCAmelCase = () # Second residual connection inside the PoolFormerOutput block _lowerCAmelCase = self.drop_path(self.output(self.after_norm(__snake_case ) ) ) _lowerCAmelCase = hidden_states + layer_output _lowerCAmelCase = (output,) + outputs return outputs class UpperCAmelCase ( nn.Module ): def __init__( self : List[str] , __snake_case : str ) -> Dict: super().__init__() _lowerCAmelCase = config # stochastic depth decay rule _lowerCAmelCase = [x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings _lowerCAmelCase = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) _lowerCAmelCase = nn.ModuleList(__snake_case ) # Transformer blocks _lowerCAmelCase = [] _lowerCAmelCase = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers _lowerCAmelCase = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( __snake_case , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(__snake_case ) ) _lowerCAmelCase = nn.ModuleList(__snake_case ) def lowercase__ ( self : List[str] , __snake_case : str , __snake_case : Any=False , __snake_case : Dict=True ) -> Any: _lowerCAmelCase = () if output_hidden_states else None _lowerCAmelCase = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): _lowerCAmelCase , _lowerCAmelCase = layers # Get patch embeddings from hidden_states _lowerCAmelCase = embedding_layer(__snake_case ) # Send the embeddings through the blocks for _, blk in enumerate(__snake_case ): _lowerCAmelCase = blk(__snake_case ) _lowerCAmelCase = layer_outputs[0] if output_hidden_states: _lowerCAmelCase = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=__snake_case , hidden_states=__snake_case ) class UpperCAmelCase ( snake_case_ ): _lowercase: Union[str, Any] = PoolFormerConfig _lowercase: Optional[int] = '''poolformer''' _lowercase: str = '''pixel_values''' _lowercase: Optional[Any] = True def lowercase__ ( self : int , __snake_case : int ) -> int: if isinstance(__snake_case , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(__snake_case , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def lowercase__ ( self : int , __snake_case : int , __snake_case : List[Any]=False ) -> Tuple: if isinstance(__snake_case , __snake_case ): _lowerCAmelCase = value A__ : Optional[Any] =r''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' A__ : List[Any] =r''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`PoolFormerImageProcessor.__call__`] for details. ''' @add_start_docstrings( '''The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.''' , snake_case_ , ) class UpperCAmelCase ( snake_case_ ): def __init__( self : Optional[int] , __snake_case : str ) -> List[str]: super().__init__(__snake_case ) _lowerCAmelCase = config _lowerCAmelCase = PoolFormerEncoder(__snake_case ) # Initialize weights and apply final processing self.post_init() def lowercase__ ( self : str ) -> int: return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(__snake_case ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=__snake_case , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowercase__ ( self : Tuple , __snake_case : Optional[torch.FloatTensor] = None , __snake_case : Optional[bool] = None , __snake_case : Optional[bool] = None , ) -> Union[Tuple, BaseModelOutputWithNoAttention]: _lowerCAmelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("""You have to specify pixel_values""" ) _lowerCAmelCase = self.encoder( __snake_case , output_hidden_states=__snake_case , return_dict=__snake_case , ) _lowerCAmelCase = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=__snake_case , hidden_states=encoder_outputs.hidden_states , ) class UpperCAmelCase ( nn.Module ): def __init__( self : Optional[int] , __snake_case : Union[str, Any] ) -> Dict: super().__init__() _lowerCAmelCase = nn.Linear(config.hidden_size , config.hidden_size ) def lowercase__ ( self : Any , __snake_case : Dict ) -> Dict: _lowerCAmelCase = self.dense(__snake_case ) return output @add_start_docstrings( ''' PoolFormer Model transformer with an image classification head on top ''' , snake_case_ , ) class UpperCAmelCase ( snake_case_ ): def __init__( self : str , __snake_case : List[Any] ) -> Optional[Any]: super().__init__(__snake_case ) _lowerCAmelCase = config.num_labels _lowerCAmelCase = PoolFormerModel(__snake_case ) # Final norm _lowerCAmelCase = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head _lowerCAmelCase = ( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__snake_case ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=__snake_case , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowercase__ ( self : List[str] , __snake_case : Optional[torch.FloatTensor] = None , __snake_case : Optional[torch.LongTensor] = None , __snake_case : Optional[bool] = None , __snake_case : Optional[bool] = None , ) -> Union[Tuple, ImageClassifierOutputWithNoAttention]: _lowerCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase = self.poolformer( __snake_case , output_hidden_states=__snake_case , return_dict=__snake_case , ) _lowerCAmelCase = outputs[0] _lowerCAmelCase = self.classifier(self.norm(__snake_case ).mean([-2, -1] ) ) _lowerCAmelCase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _lowerCAmelCase = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _lowerCAmelCase = """single_label_classification""" else: _lowerCAmelCase = """multi_label_classification""" if self.config.problem_type == "regression": _lowerCAmelCase = MSELoss() if self.num_labels == 1: _lowerCAmelCase = loss_fct(logits.squeeze() , labels.squeeze() ) else: _lowerCAmelCase = loss_fct(__snake_case , __snake_case ) elif self.config.problem_type == "single_label_classification": _lowerCAmelCase = CrossEntropyLoss() _lowerCAmelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _lowerCAmelCase = BCEWithLogitsLoss() _lowerCAmelCase = loss_fct(__snake_case , __snake_case ) if not return_dict: _lowerCAmelCase = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=__snake_case , logits=__snake_case , hidden_states=outputs.hidden_states )
365
'''simple docstring''' import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCAmelCase : @staticmethod def lowercase__ ( *__snake_case : Optional[Any] , **__snake_case : Any ) -> Tuple: pass @is_pipeline_test @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): _lowercase: Union[str, Any] = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def lowercase__ ( self : List[str] , __snake_case : Optional[Any] , __snake_case : str , __snake_case : List[str] ) -> int: _lowerCAmelCase = pipeline("""visual-question-answering""" , model="""hf-internal-testing/tiny-vilt-random-vqa""" ) _lowerCAmelCase = [ { """image""": Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ), """question""": """How many cats are there?""", }, { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """question""": """How many cats are there?""", }, ] return vqa_pipeline, examples def lowercase__ ( self : Any , __snake_case : List[Any] , __snake_case : List[Any] ) -> Union[str, Any]: _lowerCAmelCase = vqa_pipeline(__snake_case , top_k=1 ) self.assertEqual( __snake_case , [ [{"""score""": ANY(__snake_case ), """answer""": ANY(__snake_case )}], [{"""score""": ANY(__snake_case ), """answer""": ANY(__snake_case )}], ] , ) @require_torch def lowercase__ ( self : str ) -> int: _lowerCAmelCase = pipeline("""visual-question-answering""" , model="""hf-internal-testing/tiny-vilt-random-vqa""" ) _lowerCAmelCase = """./tests/fixtures/tests_samples/COCO/000000039769.png""" _lowerCAmelCase = """How many cats are there?""" _lowerCAmelCase = vqa_pipeline(image=__snake_case , question="""How many cats are there?""" , top_k=2 ) self.assertEqual( __snake_case , [{"""score""": ANY(__snake_case ), """answer""": ANY(__snake_case )}, {"""score""": ANY(__snake_case ), """answer""": ANY(__snake_case )}] ) _lowerCAmelCase = vqa_pipeline({"""image""": image, """question""": question} , top_k=2 ) self.assertEqual( __snake_case , [{"""score""": ANY(__snake_case ), """answer""": ANY(__snake_case )}, {"""score""": ANY(__snake_case ), """answer""": ANY(__snake_case )}] ) @slow @require_torch def lowercase__ ( self : List[Any] ) -> List[str]: _lowerCAmelCase = pipeline("""visual-question-answering""" , model="""dandelin/vilt-b32-finetuned-vqa""" ) _lowerCAmelCase = """./tests/fixtures/tests_samples/COCO/000000039769.png""" _lowerCAmelCase = """How many cats are there?""" _lowerCAmelCase = vqa_pipeline(image=__snake_case , question=__snake_case , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [{"""score""": 0.87_99, """answer""": """2"""}, {"""score""": 0.2_96, """answer""": """1"""}] ) _lowerCAmelCase = vqa_pipeline({"""image""": image, """question""": question} , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [{"""score""": 0.87_99, """answer""": """2"""}, {"""score""": 0.2_96, """answer""": """1"""}] ) _lowerCAmelCase = vqa_pipeline( [{"""image""": image, """question""": question}, {"""image""": image, """question""": question}] , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [[{"""score""": 0.87_99, """answer""": """2"""}, {"""score""": 0.2_96, """answer""": """1"""}]] * 2 , ) @require_tf @unittest.skip("""Visual question answering not implemented in TF""" ) def lowercase__ ( self : List[str] ) -> Union[str, Any]: pass
220
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ : Optional[int] =logging.get_logger(__name__) lowerCAmelCase__ : List[Any] ='''▁''' lowerCAmelCase__ : Optional[int] ={'''vocab_file''': '''sentencepiece.bpe.model'''} lowerCAmelCase__ : Tuple ={ '''vocab_file''': { '''facebook/xglm-564M''': '''https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model''', } } lowerCAmelCase__ : List[str] ={ '''facebook/xglm-564M''': 2048, } class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ : Tuple = VOCAB_FILES_NAMES UpperCamelCase__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : Optional[int] = ['''input_ids''', '''attention_mask'''] def __init__( self , _A , _A="<s>" , _A="</s>" , _A="</s>" , _A="<s>" , _A="<unk>" , _A="<pad>" , _A = None , **_A , ): '''simple docstring''' __SCREAMING_SNAKE_CASE = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer __SCREAMING_SNAKE_CASE = 7 __SCREAMING_SNAKE_CASE = [f"""<madeupword{i}>""" for i in range(self.num_madeup_words )] __SCREAMING_SNAKE_CASE = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=_A , eos_token=_A , unk_token=_A , sep_token=_A , cls_token=_A , pad_token=_A , sp_model_kwargs=self.sp_model_kwargs , **_A , ) __SCREAMING_SNAKE_CASE = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_A ) ) __SCREAMING_SNAKE_CASE = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab __SCREAMING_SNAKE_CASE = 1 # Mimic fairseq token-to-id alignment for the first 4 token __SCREAMING_SNAKE_CASE = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} __SCREAMING_SNAKE_CASE = len(self.sp_model ) __SCREAMING_SNAKE_CASE = {f"""<madeupword{i}>""": sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(_A ) __SCREAMING_SNAKE_CASE = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.__dict__.copy() __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = self.sp_model.serialized_model_proto() return state def __setstate__( self , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _A ( self , _A , _A = None ): '''simple docstring''' if token_ids_a is None: return [self.sep_token_id] + token_ids_a __SCREAMING_SNAKE_CASE = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def _A ( self , _A , _A = None , _A = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_A , token_ids_a=_A , already_has_special_tokens=_A ) if token_ids_a is None: return [1] + ([0] * len(_A )) return [1] + ([0] * len(_A )) + [1, 1] + ([0] * len(_A )) def _A ( self , _A , _A = None ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def _A ( self ): '''simple docstring''' return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = {self.convert_ids_to_tokens(_A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _A ( self , _A ): '''simple docstring''' return self.sp_model.encode(_A , out_type=_A ) def _A ( self , _A ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __SCREAMING_SNAKE_CASE = self.sp_model.PieceToId(_A ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _A ( self , _A ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _A ( self , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = ''.join(_A ).replace(_A , ' ' ).strip() return out_string def _A ( self , _A , _A = None ): '''simple docstring''' if not os.path.isdir(_A ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __SCREAMING_SNAKE_CASE = os.path.join( _A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _A ) elif not os.path.isfile(self.vocab_file ): with open(_A , 'wb' ) as fi: __SCREAMING_SNAKE_CASE = self.sp_model.serialized_model_proto() fi.write(_A ) return (out_vocab_file,)
257
from heapq import heappop, heappush import numpy as np def __lowercase ( a__ , a__ , a__ , a__ , ) -> tuple[float | int, list[tuple[int, int]]]: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = grid.shape __SCREAMING_SNAKE_CASE = [-1, 1, 0, 0] __SCREAMING_SNAKE_CASE = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = [(0, source)], set() __SCREAMING_SNAKE_CASE = np.full((rows, cols) , np.inf ) __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = np.empty((rows, cols) , dtype=a__ ) __SCREAMING_SNAKE_CASE = None while queue: ((__SCREAMING_SNAKE_CASE) , (__SCREAMING_SNAKE_CASE)) = heappop(a__ ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: __SCREAMING_SNAKE_CASE = [] while (x, y) != source: path.append((x, y) ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = predecessors[x, y] path.append(a__ ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(a__ ) ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: __SCREAMING_SNAKE_CASE = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(a__ , (dist + 1, (nx, ny)) ) __SCREAMING_SNAKE_CASE = dist + 1 __SCREAMING_SNAKE_CASE = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
257
1
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ = logging.get_logger(__name__) UpperCAmelCase__ = { 'asapp/sew-d-tiny-100k': 'https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class lowerCAmelCase__ ( A_ ): __a = """sew-d""" def __init__( self : Optional[int] , _lowerCamelCase : Tuple=32 , _lowerCamelCase : Union[str, Any]=768 , _lowerCamelCase : Tuple=12 , _lowerCamelCase : Any=12 , _lowerCamelCase : List[str]=3072 , _lowerCamelCase : Any=2 , _lowerCamelCase : str=512 , _lowerCamelCase : Union[str, Any]=256 , _lowerCamelCase : Optional[Any]=True , _lowerCamelCase : Optional[int]=True , _lowerCamelCase : int=("p2c", "c2p") , _lowerCamelCase : Any="layer_norm" , _lowerCamelCase : Dict="gelu_python" , _lowerCamelCase : Optional[int]=0.1 , _lowerCamelCase : Optional[Any]=0.1 , _lowerCamelCase : List[Any]=0.1 , _lowerCamelCase : Union[str, Any]=0.0 , _lowerCamelCase : int=0.1 , _lowerCamelCase : Optional[int]=0.0_2 , _lowerCamelCase : List[Any]=1e-7 , _lowerCamelCase : List[str]=1e-5 , _lowerCamelCase : Optional[Any]="group" , _lowerCamelCase : Dict="gelu" , _lowerCamelCase : Union[str, Any]=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , _lowerCamelCase : Optional[Any]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , _lowerCamelCase : Dict=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , _lowerCamelCase : List[str]=False , _lowerCamelCase : Dict=128 , _lowerCamelCase : Optional[int]=16 , _lowerCamelCase : List[Any]=True , _lowerCamelCase : Union[str, Any]=0.0_5 , _lowerCamelCase : int=10 , _lowerCamelCase : Any=2 , _lowerCamelCase : Any=0.0 , _lowerCamelCase : List[str]=10 , _lowerCamelCase : List[Any]=0 , _lowerCamelCase : Union[str, Any]="mean" , _lowerCamelCase : str=False , _lowerCamelCase : int=False , _lowerCamelCase : Union[str, Any]=256 , _lowerCamelCase : int=0 , _lowerCamelCase : Tuple=1 , _lowerCamelCase : Any=2 , **_lowerCamelCase : str , ): super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase ) _snake_case = hidden_size _snake_case = feat_extract_norm _snake_case = feat_extract_activation _snake_case = list(_lowerCamelCase ) _snake_case = list(_lowerCamelCase ) _snake_case = list(_lowerCamelCase ) _snake_case = conv_bias _snake_case = num_conv_pos_embeddings _snake_case = num_conv_pos_embedding_groups _snake_case = len(self.conv_dim ) _snake_case = num_hidden_layers _snake_case = intermediate_size _snake_case = squeeze_factor _snake_case = max_position_embeddings _snake_case = position_buckets _snake_case = share_att_key _snake_case = relative_attention _snake_case = norm_rel_ebd _snake_case = list(_lowerCamelCase ) _snake_case = hidden_act _snake_case = num_attention_heads _snake_case = hidden_dropout _snake_case = attention_dropout _snake_case = activation_dropout _snake_case = feat_proj_dropout _snake_case = final_dropout _snake_case = layer_norm_eps _snake_case = feature_layer_norm_eps _snake_case = initializer_range _snake_case = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect.''' '''It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,''' f'''but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)''' f'''= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _snake_case = apply_spec_augment _snake_case = mask_time_prob _snake_case = mask_time_length _snake_case = mask_time_min_masks _snake_case = mask_feature_prob _snake_case = mask_feature_length _snake_case = mask_feature_min_masks # ctc loss _snake_case = ctc_loss_reduction _snake_case = ctc_zero_infinity # sequence classification _snake_case = use_weighted_layer_sum _snake_case = classifier_proj_size @property def lowercase ( self : int ): return functools.reduce(operator.mul , self.conv_stride , 1 )
355
"""simple docstring""" def _UpperCAmelCase ( __lowerCamelCase : float , __lowerCamelCase : float ) -> float: if mass < 0: raise ValueError('''The mass of a body cannot be negative''' ) return 0.5 * mass * abs(__lowerCamelCase ) * abs(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
40
0
"""simple docstring""" def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : Optional[int] = 2**power lowercase__ : Tuple = 0 while n: lowercase__ , lowercase__ : str = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
77
from __future__ import annotations from typing import Any class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_ ) -> None: UpperCamelCase : Any = num_of_nodes UpperCamelCase : list[list[int]] = [] UpperCamelCase : dict[int, int] = {} def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> None: self.m_edges.append([u_node, v_node, weight] ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> int: if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> None: if self.m_component[u_node] != u_node: for k in self.m_component: UpperCamelCase : Dict = self.find_component(SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> None: if component_size[u_node] <= component_size[v_node]: UpperCamelCase : Tuple = v_node component_size[v_node] += component_size[u_node] self.set_component(SCREAMING_SNAKE_CASE_ ) elif component_size[u_node] >= component_size[v_node]: UpperCamelCase : Union[str, Any] = self.find_component(SCREAMING_SNAKE_CASE_ ) component_size[u_node] += component_size[v_node] self.set_component(SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self ) -> None: UpperCamelCase : int = [] UpperCamelCase : int = 0 UpperCamelCase : list[Any] = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) UpperCamelCase : List[str] = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: UpperCamelCase , UpperCamelCase , UpperCamelCase : List[Any] = edge UpperCamelCase : str = self.m_component[u] UpperCamelCase : Any = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): UpperCamelCase : Any = [u, v, w] for edge in minimum_weight_edge: if isinstance(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): UpperCamelCase , UpperCamelCase , UpperCamelCase : int = edge UpperCamelCase : List[Any] = self.m_component[u] UpperCamelCase : Tuple = self.m_component[v] if u_component != v_component: mst_weight += w self.union(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) print(F"""Added edge [{u} - {v}]\nAdded weight: {w}\n""" ) num_of_components -= 1 UpperCamelCase : Optional[Any] = [-1] * self.m_num_of_nodes print(F"""The total weight of the minimal spanning tree is: {mst_weight}""" ) def UpperCamelCase ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
119
0
def A ( _UpperCAmelCase : list ) -> list: '''simple docstring''' if len(_UpperCAmelCase ) < 2: return collection def circle_sort_util(_UpperCAmelCase : list , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> bool: _UpperCAmelCase = False if low == high: return swapped _UpperCAmelCase = low _UpperCAmelCase = high while left < right: if collection[left] > collection[right]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right], collection[left], ) _UpperCAmelCase = True left += 1 right -= 1 if left == right and collection[left] > collection[right + 1]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right + 1], collection[left], ) _UpperCAmelCase = True _UpperCAmelCase = low + int((high - low) / 2 ) _UpperCAmelCase = circle_sort_util(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) _UpperCAmelCase = circle_sort_util(_UpperCAmelCase , mid + 1 , _UpperCAmelCase ) return swapped or left_swap or right_swap _UpperCAmelCase = True while is_not_sorted is True: _UpperCAmelCase = circle_sort_util(_UpperCAmelCase , 0 , len(_UpperCAmelCase ) - 1 ) return collection if __name__ == "__main__": UpperCAmelCase__ = input("Enter numbers separated by a comma:\n").strip() UpperCAmelCase__ = [int(item) for item in user_input.split(",")] print(circle_sort(unsorted))
290
import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings UpperCAmelCase__ = r"\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information.\n\n Args:\n title_sep (`str`, *optional*, defaults to `\" / \"`):\n Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`].\n doc_sep (`str`, *optional*, defaults to `\" // \"`):\n Separator inserted between the text of the retrieved document and the original input when calling\n [`RagRetriever`].\n n_docs (`int`, *optional*, defaults to 5):\n Number of documents to retrieve.\n max_combined_length (`int`, *optional*, defaults to 300):\n Max length of contextualized input returned by [`~RagRetriever.__call__`].\n retrieval_vector_size (`int`, *optional*, defaults to 768):\n Dimensionality of the document embeddings indexed by [`RagRetriever`].\n retrieval_batch_size (`int`, *optional*, defaults to 8):\n Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated\n [`RagRetriever`].\n dataset (`str`, *optional*, defaults to `\"wiki_dpr\"`):\n A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids\n using `datasets.list_datasets()`).\n dataset_split (`str`, *optional*, defaults to `\"train\"`)\n Which split of the `dataset` to load.\n index_name (`str`, *optional*, defaults to `\"compressed\"`)\n The index name of the index associated with the `dataset`. One can choose between `\"legacy\"`, `\"exact\"` and\n `\"compressed\"`.\n index_path (`str`, *optional*)\n The path to the serialized faiss index on disk.\n passages_path (`str`, *optional*):\n A path to text passages compatible with the faiss index. Required if using\n [`~models.rag.retrieval_rag.LegacyIndex`]\n use_dummy_dataset (`bool`, *optional*, defaults to `False`)\n Whether to load a \"dummy\" variant of the dataset specified by `dataset`.\n label_smoothing (`float`, *optional*, defaults to 0.0):\n Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing\n in the loss calculation. If set to 0, no label smoothing is performed.\n do_marginalize (`bool`, *optional*, defaults to `False`):\n If `True`, the logits are marginalized over all documents by making use of\n `torch.nn.functional.log_softmax`.\n reduce_loss (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation.\n do_deduplication (`bool`, *optional*, defaults to `True`):\n Whether or not to deduplicate the generations from different context documents for a given input. Has to be\n set to `False` if used while training with distributed backend.\n exclude_bos_score (`bool`, *optional*, defaults to `False`):\n Whether or not to disregard the BOS token when computing the loss.\n output_retrieved(`bool`, *optional*, defaults to `False`):\n If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and\n `context_attention_mask` are returned. See returned tensors for more detail.\n use_cache (`bool`, *optional*, defaults to `True`):\n Whether or not the model should return the last key/values attentions (not used by all models).\n forced_eos_token_id (`int`, *optional*):\n The id of the token to force as the last generated token when `max_length` is reached. Usually set to\n `eos_token_id`.\n" @add_start_docstrings(A ) class __lowerCAmelCase ( A ): UpperCamelCase = '''rag''' UpperCamelCase = True def __init__( self : str , A : int=None , A : Dict=True , A : int=None , A : Any=None , A : Any=None , A : Any=None , A : int=None , A : str=" / " , A : int=" // " , A : Any=5 , A : Optional[Any]=3_00 , A : Optional[Any]=7_68 , A : Optional[int]=8 , A : Union[str, Any]="wiki_dpr" , A : int="train" , A : Tuple="compressed" , A : str=None , A : Any=None , A : Tuple=False , A : List[str]=False , A : Tuple=0.0 , A : Any=True , A : int=False , A : Dict=False , A : List[str]=False , A : Optional[Any]=True , A : Tuple=None , **A : Optional[Any] , ) -> Optional[Any]: """simple docstring""" super().__init__( bos_token_id=A , pad_token_id=A , eos_token_id=A , decoder_start_token_id=A , forced_eos_token_id=A , is_encoder_decoder=A , prefix=A , vocab_size=A , **A , ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" _UpperCAmelCase = kwargs.pop('question_encoder') _UpperCAmelCase = question_encoder_config.pop('model_type') _UpperCAmelCase = kwargs.pop('generator') _UpperCAmelCase = decoder_config.pop('model_type') from ..auto.configuration_auto import AutoConfig _UpperCAmelCase = AutoConfig.for_model(A , **A) _UpperCAmelCase = AutoConfig.for_model(A , **A) _UpperCAmelCase = reduce_loss _UpperCAmelCase = label_smoothing _UpperCAmelCase = exclude_bos_score _UpperCAmelCase = do_marginalize _UpperCAmelCase = title_sep _UpperCAmelCase = doc_sep _UpperCAmelCase = n_docs _UpperCAmelCase = max_combined_length _UpperCAmelCase = dataset _UpperCAmelCase = dataset_split _UpperCAmelCase = index_name _UpperCAmelCase = retrieval_vector_size _UpperCAmelCase = retrieval_batch_size _UpperCAmelCase = passages_path _UpperCAmelCase = index_path _UpperCAmelCase = use_dummy_dataset _UpperCAmelCase = output_retrieved _UpperCAmelCase = do_deduplication _UpperCAmelCase = use_cache if self.forced_eos_token_id is None: _UpperCAmelCase = getattr(self.generator , 'forced_eos_token_id' , A) @classmethod def _lowerCamelCase ( cls : Union[str, Any] , A : PretrainedConfig , A : PretrainedConfig , **A : Optional[Any]) -> PretrainedConfig: """simple docstring""" return cls(question_encoder=question_encoder_config.to_dict() , generator=generator_config.to_dict() , **A) def _lowerCamelCase ( self : Dict) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = copy.deepcopy(self.__dict__) _UpperCAmelCase = self.question_encoder.to_dict() _UpperCAmelCase = self.generator.to_dict() _UpperCAmelCase = self.__class__.model_type return output
290
1
from __future__ import annotations import math from collections.abc import Callable def UpperCAmelCase ( a_ , a_ , a_ , a_ = 1_0_0 , ) -> float: """simple docstring""" __A = x_start __A = fnc(a_ ) __A = 0.0 for _ in range(a_ ): # Approximates curve as a sequence of linear lines and sums their length __A = (x_end - x_start) / steps + xa __A = fnc(a_ ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step __A = xa __A = fxa return length if __name__ == "__main__": def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" return math.sin(1_0 * x ) print('f(x) = sin(10 * x)') print('The length of the curve from x = -10 to x = 10 is:') SCREAMING_SNAKE_CASE :Tuple = 10 while i <= 10_0000: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
15
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowercase = { """configuration_bert""": ["""BERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BertConfig""", """BertOnnxConfig"""], """tokenization_bert""": ["""BasicTokenizer""", """BertTokenizer""", """WordpieceTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ["""BertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ """BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BertForMaskedLM""", """BertForMultipleChoice""", """BertForNextSentencePrediction""", """BertForPreTraining""", """BertForQuestionAnswering""", """BertForSequenceClassification""", """BertForTokenClassification""", """BertLayer""", """BertLMHeadModel""", """BertModel""", """BertPreTrainedModel""", """load_tf_weights_in_bert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ """TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBertEmbeddings""", """TFBertForMaskedLM""", """TFBertForMultipleChoice""", """TFBertForNextSentencePrediction""", """TFBertForPreTraining""", """TFBertForQuestionAnswering""", """TFBertForSequenceClassification""", """TFBertForTokenClassification""", """TFBertLMHeadModel""", """TFBertMainLayer""", """TFBertModel""", """TFBertPreTrainedModel""", ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ["""TFBertTokenizer"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ """FlaxBertForCausalLM""", """FlaxBertForMaskedLM""", """FlaxBertForMultipleChoice""", """FlaxBertForNextSentencePrediction""", """FlaxBertForPreTraining""", """FlaxBertForQuestionAnswering""", """FlaxBertForSequenceClassification""", """FlaxBertForTokenClassification""", """FlaxBertModel""", """FlaxBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
40
0
'''simple docstring''' import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class a__ ( lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Tuple = (EulerDiscreteScheduler,) __UpperCamelCase : str = 10 def _snake_case (self , **__lowercase ): __lowerCAmelCase = { '''num_train_timesteps''': 11_00, '''beta_start''': 0.0_0_0_1, '''beta_end''': 0.0_2, '''beta_schedule''': '''linear''', } config.update(**__A ) return config def _snake_case (self ): for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=__A ) def _snake_case (self ): for beta_start, beta_end in zip([0.0_0_0_0_1, 0.0_0_0_1, 0.0_0_1] , [0.0_0_0_2, 0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=__A , beta_end=__A ) def _snake_case (self ): for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=__A ) def _snake_case (self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__A ) def _snake_case (self ): __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**__A ) scheduler.set_timesteps(self.num_inference_steps ) __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma __lowerCAmelCase = sample.to(__A ) for i, t in enumerate(scheduler.timesteps ): __lowerCAmelCase = scheduler.scale_model_input(__A , __A ) __lowerCAmelCase = model(__A , __A ) __lowerCAmelCase = scheduler.step(__A , __A , __A , generator=__A ) __lowerCAmelCase = output.prev_sample __lowerCAmelCase = torch.sum(torch.abs(__A ) ) __lowerCAmelCase = torch.mean(torch.abs(__A ) ) assert abs(result_sum.item() - 10.08_07 ) < 1e-2 assert abs(result_mean.item() - 0.0_1_3_1 ) < 1e-3 def _snake_case (self ): __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config(prediction_type='''v_prediction''' ) __lowerCAmelCase = scheduler_class(**__A ) scheduler.set_timesteps(self.num_inference_steps ) __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma __lowerCAmelCase = sample.to(__A ) for i, t in enumerate(scheduler.timesteps ): __lowerCAmelCase = scheduler.scale_model_input(__A , __A ) __lowerCAmelCase = model(__A , __A ) __lowerCAmelCase = scheduler.step(__A , __A , __A , generator=__A ) __lowerCAmelCase = output.prev_sample __lowerCAmelCase = torch.sum(torch.abs(__A ) ) __lowerCAmelCase = torch.mean(torch.abs(__A ) ) assert abs(result_sum.item() - 0.0_0_0_2 ) < 1e-2 assert abs(result_mean.item() - 2.2_676e-06 ) < 1e-3 def _snake_case (self ): __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**__A ) scheduler.set_timesteps(self.num_inference_steps , device=__A ) __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() __lowerCAmelCase = sample.to(__A ) for t in scheduler.timesteps: __lowerCAmelCase = scheduler.scale_model_input(__A , __A ) __lowerCAmelCase = model(__A , __A ) __lowerCAmelCase = scheduler.step(__A , __A , __A , generator=__A ) __lowerCAmelCase = output.prev_sample __lowerCAmelCase = torch.sum(torch.abs(__A ) ) __lowerCAmelCase = torch.mean(torch.abs(__A ) ) assert abs(result_sum.item() - 10.08_07 ) < 1e-2 assert abs(result_mean.item() - 0.0_1_3_1 ) < 1e-3 def _snake_case (self ): __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**__A , use_karras_sigmas=__A ) scheduler.set_timesteps(self.num_inference_steps , device=__A ) __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() __lowerCAmelCase = sample.to(__A ) for t in scheduler.timesteps: __lowerCAmelCase = scheduler.scale_model_input(__A , __A ) __lowerCAmelCase = model(__A , __A ) __lowerCAmelCase = scheduler.step(__A , __A , __A , generator=__A ) __lowerCAmelCase = output.prev_sample __lowerCAmelCase = torch.sum(torch.abs(__A ) ) __lowerCAmelCase = torch.mean(torch.abs(__A ) ) assert abs(result_sum.item() - 1_24.52_29_94_99_51_17_19 ) < 1e-2 assert abs(result_mean.item() - 0.1_6_2_1_3_9_3_2_6_3_3_3_9_9_9_6_3 ) < 1e-3
360
'''simple docstring''' import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class a__ ( __A , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Optional[Any] = ConsistencyModelPipeline __UpperCamelCase : Optional[int] = UNCONDITIONAL_IMAGE_GENERATION_PARAMS __UpperCamelCase : int = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt __UpperCamelCase : List[Any] = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _snake_case (self ): __lowerCAmelCase = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet''' , ) return unet @property def _snake_case (self ): __lowerCAmelCase = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet_class_cond''' , ) return unet def _snake_case (self , __lowercase=False ): if class_cond: __lowerCAmelCase = self.dummy_cond_unet else: __lowerCAmelCase = self.dummy_uncond_unet # Default to CM multistep sampler __lowerCAmelCase = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) __lowerCAmelCase = { '''unet''': unet, '''scheduler''': scheduler, } return components def _snake_case (self , __lowercase , __lowercase=0 ): if str(__lowercase ).startswith('''mps''' ): __lowerCAmelCase = torch.manual_seed(__lowercase ) else: __lowerCAmelCase = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __lowerCAmelCase = { '''batch_size''': 1, '''num_inference_steps''': None, '''timesteps''': [22, 0], '''generator''': generator, '''output_type''': '''np''', } return inputs def _snake_case (self ): __lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = ConsistencyModelPipeline(**__lowercase ) __lowerCAmelCase = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = pipe(**__lowercase ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _snake_case (self ): __lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components(class_cond=__lowercase ) __lowerCAmelCase = ConsistencyModelPipeline(**__lowercase ) __lowerCAmelCase = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = 0 __lowerCAmelCase = pipe(**__lowercase ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _snake_case (self ): __lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = ConsistencyModelPipeline(**__lowercase ) __lowerCAmelCase = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = 1 __lowerCAmelCase = None __lowerCAmelCase = pipe(**__lowercase ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _snake_case (self ): __lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components(class_cond=__lowercase ) __lowerCAmelCase = ConsistencyModelPipeline(**__lowercase ) __lowerCAmelCase = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_dummy_inputs(__lowercase ) __lowerCAmelCase = 1 __lowerCAmelCase = None __lowerCAmelCase = 0 __lowerCAmelCase = pipe(**__lowercase ).images assert image.shape == (1, 32, 32, 3) __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class a__ ( unittest.TestCase ): """simple docstring""" def _snake_case (self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case (self , __lowercase=0 , __lowercase=False , __lowercase="cpu" , __lowercase=torch.floataa , __lowercase=(1, 3, 64, 64) ): __lowerCAmelCase = torch.manual_seed(__lowercase ) __lowerCAmelCase = { '''num_inference_steps''': None, '''timesteps''': [22, 0], '''class_labels''': 0, '''generator''': generator, '''output_type''': '''np''', } if get_fixed_latents: __lowerCAmelCase = self.get_fixed_latents(seed=__lowercase , device=__lowercase , dtype=__lowercase , shape=__lowercase ) __lowerCAmelCase = latents return inputs def _snake_case (self , __lowercase=0 , __lowercase="cpu" , __lowercase=torch.floataa , __lowercase=(1, 3, 64, 64) ): if type(__lowercase ) == str: __lowerCAmelCase = torch.device(__lowercase ) __lowerCAmelCase = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) __lowerCAmelCase = randn_tensor(__lowercase , generator=__lowercase , device=__lowercase , dtype=__lowercase ) return latents def _snake_case (self ): __lowerCAmelCase = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) __lowerCAmelCase = ConsistencyModelPipeline(unet=__lowercase , scheduler=__lowercase ) pipe.to(torch_device=__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**__lowercase ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def _snake_case (self ): __lowerCAmelCase = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) __lowerCAmelCase = ConsistencyModelPipeline(unet=__lowercase , scheduler=__lowercase ) pipe.to(torch_device=__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = 1 __lowerCAmelCase = None __lowerCAmelCase = pipe(**__lowercase ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def _snake_case (self ): __lowerCAmelCase = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) __lowerCAmelCase = ConsistencyModelPipeline(unet=__lowercase , scheduler=__lowercase ) pipe.to(torch_device=__lowercase , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_inputs(get_fixed_latents=__lowercase , device=__lowercase ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=__lowercase , enable_math=__lowercase , enable_mem_efficient=__lowercase ): __lowerCAmelCase = pipe(**__lowercase ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def _snake_case (self ): __lowerCAmelCase = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __lowerCAmelCase = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) __lowerCAmelCase = ConsistencyModelPipeline(unet=__lowercase , scheduler=__lowercase ) pipe.to(torch_device=__lowercase , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=__lowercase ) __lowerCAmelCase = self.get_inputs(get_fixed_latents=__lowercase , device=__lowercase ) __lowerCAmelCase = 1 __lowerCAmelCase = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=__lowercase , enable_math=__lowercase , enable_mem_efficient=__lowercase ): __lowerCAmelCase = pipe(**__lowercase ).images assert image.shape == (1, 64, 64, 3) __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
9
0
'''simple docstring''' import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __a: Optional[int] = 'platform' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class UpperCAmelCase : '''simple docstring''' SCREAMING_SNAKE_CASE = PegasusConfig SCREAMING_SNAKE_CASE = {} SCREAMING_SNAKE_CASE = '''gelu''' def __init__( self , __lowerCAmelCase , __lowerCAmelCase=13 , __lowerCAmelCase=7 , __lowerCAmelCase=True , __lowerCAmelCase=False , __lowerCAmelCase=99 , __lowerCAmelCase=32 , __lowerCAmelCase=5 , __lowerCAmelCase=4 , __lowerCAmelCase=37 , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=20 , __lowerCAmelCase=2 , __lowerCAmelCase=1 , __lowerCAmelCase=0 , ) -> Optional[Any]: lowercase__ : List[Any] = parent lowercase__ : Optional[Any] = batch_size lowercase__ : Dict = seq_length lowercase__ : Optional[int] = is_training lowercase__ : List[Any] = use_labels lowercase__ : List[str] = vocab_size lowercase__ : str = hidden_size lowercase__ : Tuple = num_hidden_layers lowercase__ : List[str] = num_attention_heads lowercase__ : Tuple = intermediate_size lowercase__ : List[str] = hidden_dropout_prob lowercase__ : List[Any] = attention_probs_dropout_prob lowercase__ : List[Any] = max_position_embeddings lowercase__ : List[str] = eos_token_id lowercase__ : Optional[Any] = pad_token_id lowercase__ : Union[str, Any] = bos_token_id def _lowerCAmelCase( self ) -> Dict: lowercase__ : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ).clip(3 , self.vocab_size ) lowercase__ : List[Any] = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) , 1 ) lowercase__ : List[str] = np.concatenate([input_ids, eos_tensor] , axis=1 ) lowercase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : Optional[Any] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowercase__ : Optional[Any] = prepare_pegasus_inputs_dict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return config, inputs_dict def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Optional[Any]: lowercase__ : Any = 20 lowercase__ : int = model_class_name(SCREAMING_SNAKE_CASE__ ) lowercase__ : Union[str, Any] = model.encode(inputs_dict['''input_ids'''] ) lowercase__ : List[Any] = ( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) lowercase__ : List[Any] = model.init_cache(decoder_input_ids.shape[0] , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase__ : Dict = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype='''i4''' ) lowercase__ : List[Any] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase__ : Union[str, Any] = model.decode( decoder_input_ids[:, :-1] , SCREAMING_SNAKE_CASE__ , decoder_attention_mask=SCREAMING_SNAKE_CASE__ , past_key_values=SCREAMING_SNAKE_CASE__ , decoder_position_ids=SCREAMING_SNAKE_CASE__ , ) lowercase__ : int = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='''i4''' ) lowercase__ : Union[str, Any] = model.decode( decoder_input_ids[:, -1:] , SCREAMING_SNAKE_CASE__ , decoder_attention_mask=SCREAMING_SNAKE_CASE__ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=SCREAMING_SNAKE_CASE__ , ) lowercase__ : Dict = model.decode(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase__ : Optional[int] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F"""Max diff is {diff}""" ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Optional[Any]: lowercase__ : Tuple = 20 lowercase__ : Optional[int] = model_class_name(SCREAMING_SNAKE_CASE__ ) lowercase__ : int = model.encode(inputs_dict['''input_ids'''] ) lowercase__ : Dict = ( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) lowercase__ : Optional[int] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) lowercase__ : str = model.init_cache(decoder_input_ids.shape[0] , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase__ : int = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase__ : List[str] = model.decode( decoder_input_ids[:, :-1] , SCREAMING_SNAKE_CASE__ , decoder_attention_mask=SCREAMING_SNAKE_CASE__ , past_key_values=SCREAMING_SNAKE_CASE__ , decoder_position_ids=SCREAMING_SNAKE_CASE__ , ) lowercase__ : Optional[int] = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='''i4''' ) lowercase__ : Any = model.decode( decoder_input_ids[:, -1:] , SCREAMING_SNAKE_CASE__ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=SCREAMING_SNAKE_CASE__ , decoder_position_ids=SCREAMING_SNAKE_CASE__ , ) lowercase__ : Optional[int] = model.decode(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , decoder_attention_mask=SCREAMING_SNAKE_CASE__ ) lowercase__ : List[Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F"""Max diff is {diff}""" ) def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None , ): if attention_mask is None: lowercase__ : List[str] = np.not_equal(__A , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: lowercase__ : str = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class UpperCAmelCase ( lowercase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) SCREAMING_SNAKE_CASE = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _lowerCAmelCase( self ) -> Any: lowercase__ : Any = FlaxPegasusModelTester(self ) lowercase__ : Tuple = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ ) def _lowerCAmelCase( self ) -> Tuple: self.config_tester.run_common_tests() def _lowerCAmelCase( self ) -> Dict: lowercase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _lowerCAmelCase( self ) -> str: lowercase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _lowerCAmelCase( self ) -> List[Any]: lowercase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowercase__ : Dict = self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase__ : Union[str, Any] = model_class(SCREAMING_SNAKE_CASE__ ) @jax.jit def encode_jitted(__lowerCAmelCase , __lowerCAmelCase=None , **__lowerCAmelCase ): return model.encode(input_ids=SCREAMING_SNAKE_CASE__ , attention_mask=SCREAMING_SNAKE_CASE__ ) with self.subTest('''JIT Enabled''' ): lowercase__ : Any = encode_jitted(**SCREAMING_SNAKE_CASE__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): lowercase__ : Optional[int] = encode_jitted(**SCREAMING_SNAKE_CASE__ ).to_tuple() self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , len(SCREAMING_SNAKE_CASE__ ) ) for jitted_output, output in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): self.assertEqual(jitted_output.shape , output.shape ) def _lowerCAmelCase( self ) -> Dict: lowercase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowercase__ : List[str] = model_class(SCREAMING_SNAKE_CASE__ ) lowercase__ : int = model.encode(inputs_dict['''input_ids'''] , inputs_dict['''attention_mask'''] ) lowercase__ : str = { 'decoder_input_ids': inputs_dict['decoder_input_ids'], 'decoder_attention_mask': inputs_dict['decoder_attention_mask'], 'encoder_outputs': encoder_outputs, } @jax.jit def decode_jitted(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): return model.decode( decoder_input_ids=SCREAMING_SNAKE_CASE__ , decoder_attention_mask=SCREAMING_SNAKE_CASE__ , encoder_outputs=SCREAMING_SNAKE_CASE__ , ) with self.subTest('''JIT Enabled''' ): lowercase__ : List[str] = decode_jitted(**SCREAMING_SNAKE_CASE__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): lowercase__ : Dict = decode_jitted(**SCREAMING_SNAKE_CASE__ ).to_tuple() self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , len(SCREAMING_SNAKE_CASE__ ) ) for jitted_output, output in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def _lowerCAmelCase( self ) -> str: for model_class_name in self.all_model_classes: lowercase__ : Dict = model_class_name.from_pretrained('''google/pegasus-large''' , from_pt=SCREAMING_SNAKE_CASE__ ) lowercase__ : Optional[Any] = np.ones((1, 1) ) lowercase__ : Union[str, Any] = model(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) @slow def _lowerCAmelCase( self ) -> Tuple: lowercase__ : List[Any] = FlaxPegasusForConditionalGeneration.from_pretrained('''google/pegasus-xsum''' ) lowercase__ : List[Any] = PegasusTokenizer.from_pretrained('''google/pegasus-xsum''' ) lowercase__ : Optional[Any] = [ ' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.', ' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ', ] lowercase__ : Union[str, Any] = [ 'California\'s largest electricity provider has turned off power to hundreds of thousands of customers.', 'Pop group N-Dubz have revealed they were surprised to get four nominations for this year\'s Mobo Awards.', ] lowercase__ : Any = tokenizer(SCREAMING_SNAKE_CASE__ , return_tensors='''np''' , truncation=SCREAMING_SNAKE_CASE__ , max_length=512 , padding=SCREAMING_SNAKE_CASE__ ) lowercase__ : List[str] = model.generate(**SCREAMING_SNAKE_CASE__ , num_beams=2 ).sequences lowercase__ : Tuple = tokenizer.batch_decode(SCREAMING_SNAKE_CASE__ , skip_special_tokens=SCREAMING_SNAKE_CASE__ ) assert tgt_text == decoded
198
from __future__ import annotations UpperCAmelCase_ : Tuple = [] def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] , __A : int , __A : int ) -> bool: """simple docstring""" for i in range(len(__A ) ): if board[row][i] == 1: return False for i in range(len(__A ) ): if board[i][column] == 1: return False for i, j in zip(range(__A , -1 , -1 ) , range(__A , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(__A , -1 , -1 ) , range(__A , len(__A ) ) ): if board[i][j] == 1: return False return True def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] , __A : int ) -> bool: """simple docstring""" if row >= len(__A ): solution.append(__A ) printboard(__A ) print() return True for i in range(len(__A ) ): if is_safe(__A , __A , __A ): a_ : Any = 1 solve(__A , row + 1 ) a_ : Tuple = 0 return False def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] ) -> None: """simple docstring""" for i in range(len(__A ) ): for j in range(len(__A ) ): if board[i][j] == 1: print('Q' , end=' ' ) else: print('.' , end=' ' ) print() # n=int(input("The no. of queens")) UpperCAmelCase_ : List[str] = 8 UpperCAmelCase_ : str = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('The total no. of solutions are :', len(solution))
32
0
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __SCREAMING_SNAKE_CASE = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __SCREAMING_SNAKE_CASE = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __SCREAMING_SNAKE_CASE = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): return float((preds == labels).mean() ) def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase="binary" ): A : Optional[Any] = simple_accuracy(_lowerCamelCase , _lowerCamelCase ) A : Tuple = float(fa_score(y_true=_lowerCamelCase , y_pred=_lowerCamelCase , average=_lowerCamelCase ) ) return { "accuracy": acc, "f1": fa, } def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): A : Optional[int] = {} for id_pred, label in zip(_lowerCamelCase , _lowerCamelCase ): A : List[str] = f"""{id_pred["idx"]["paragraph"]}-{id_pred["idx"]["question"]}""" A : List[Any] = id_pred["prediction"] if question_id in question_map: question_map[question_id].append((pred, label) ) else: A : str = [(pred, label)] A , A : Any = [], [] for question, preds_labels in question_map.items(): A , A : Tuple = zip(*_lowerCamelCase ) A : Optional[Any] = fa_score(y_true=_lowerCamelCase , y_pred=_lowerCamelCase , average="macro" ) fas.append(_lowerCamelCase ) A : int = int(sum(pred == label for pred, label in preds_labels ) == len(_lowerCamelCase ) ) ems.append(_lowerCamelCase ) A : List[Any] = float(sum(_lowerCamelCase ) / len(_lowerCamelCase ) ) A : Any = sum(_lowerCamelCase ) / len(_lowerCamelCase ) A : str = float(fa_score(y_true=_lowerCamelCase , y_pred=[id_pred["prediction"] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class lowerCamelCase_ ( datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> List[str]: if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="numpy" if not self.config_name == "record" and not self.config_name == "multirc" else None , ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ) -> str: if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "prediction_text": datasets.Value("string" ), }, "references": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "answers": datasets.Sequence(datasets.Value("string" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("int64" ), "paragraph": datasets.Value("int64" ), "question": datasets.Value("int64" ), }, "prediction": datasets.Value("int64" ), }, "references": datasets.Value("int64" ), } else: return { "predictions": datasets.Value("int64" ), "references": datasets.Value("int64" ), } def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : List[Any] ) -> Optional[Any]: if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="macro" ) elif self.config_name == "record": A : Union[str, Any] = [ { "qas": [ {"id": ref["idx"]["query"], "answers": [{"text": ans} for ans in ref["answers"]]} for ref in references ] } ] A : Optional[int] = {pred["idx"]["query"]: pred["prediction_text"] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" )
256
from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class lowerCamelCase_ : '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : Any , __lowerCamelCase : Any ) -> Optional[Any]: raise NotImplementedError() def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> Optional[Any]: raise NotImplementedError() class lowerCamelCase_ ( _A ): '''simple docstring''' def __init__( self : Optional[Any] , __lowerCamelCase : "AutoTokenizer" , __lowerCamelCase : bool = False , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: A : str = tokenizer A : Tuple = skip_prompt A : Optional[Any] = decode_kwargs # variables used in the streaming process A : Any = [] A : Tuple = 0 A : List[str] = True def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , __lowerCamelCase : Optional[Any] ) -> int: if len(value.shape ) > 1 and value.shape[0] > 1: raise ValueError("TextStreamer only supports batch size 1" ) elif len(value.shape ) > 1: A : List[str] = value[0] if self.skip_prompt and self.next_tokens_are_prompt: A : Tuple = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist() ) A : str = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) # After the symbol for a new line, we flush the cache. if text.endswith("\n" ): A : int = text[self.print_len :] A : Union[str, Any] = [] A : Any = 0 # If the last token is a CJK character, we print the characters. elif len(__lowerCamelCase ) > 0 and self._is_chinese_char(ord(text[-1] ) ): A : Optional[int] = text[self.print_len :] self.print_len += len(__lowerCamelCase ) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: A : Union[str, Any] = text[self.print_len : text.rfind(" " ) + 1] self.print_len += len(__lowerCamelCase ) self.on_finalized_text(__lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ) -> List[str]: # Flush the cache, if it exists if len(self.token_cache ) > 0: A : Optional[int] = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) A : Optional[Any] = text[self.print_len :] A : Optional[int] = [] A : List[str] = 0 else: A : List[Any] = "" A : Union[str, Any] = True self.on_finalized_text(__lowerCamelCase , stream_end=__lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : List[str] , __lowerCamelCase : str , __lowerCamelCase : bool = False ) -> List[str]: print(__lowerCamelCase , flush=__lowerCamelCase , end="" if not stream_end else None ) def SCREAMING_SNAKE_CASE__ ( self : Any , __lowerCamelCase : int ) -> Dict: # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0X4_e_0_0 and cp <= 0X9_f_f_f) or (cp >= 0X3_4_0_0 and cp <= 0X4_d_b_f) # or (cp >= 0X2_0_0_0_0 and cp <= 0X2_a_6_d_f) # or (cp >= 0X2_a_7_0_0 and cp <= 0X2_b_7_3_f) # or (cp >= 0X2_b_7_4_0 and cp <= 0X2_b_8_1_f) # or (cp >= 0X2_b_8_2_0 and cp <= 0X2_c_e_a_f) # or (cp >= 0Xf_9_0_0 and cp <= 0Xf_a_f_f) or (cp >= 0X2_f_8_0_0 and cp <= 0X2_f_a_1_f) # ): # return True return False class lowerCamelCase_ ( _A ): '''simple docstring''' def __init__( self : Any , __lowerCamelCase : "AutoTokenizer" , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[float] = None , **__lowerCamelCase : Dict ) -> Any: super().__init__(__lowerCamelCase , __lowerCamelCase , **__lowerCamelCase ) A : Tuple = Queue() A : Dict = None A : List[str] = timeout def SCREAMING_SNAKE_CASE__ ( self : Dict , __lowerCamelCase : str , __lowerCamelCase : bool = False ) -> Any: self.text_queue.put(__lowerCamelCase , timeout=self.timeout ) if stream_end: self.text_queue.put(self.stop_signal , timeout=self.timeout ) def __iter__( self : int ) -> Tuple: return self def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Dict: A : int = self.text_queue.get(timeout=self.timeout ) if value == self.stop_signal: raise StopIteration() else: return value
256
1
"""simple docstring""" import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class UpperCamelCase ( datasets.BeamBasedBuilder ): def _lowercase ( self : List[Any] ) -> List[str]: return datasets.DatasetInfo( features=datasets.Features({"""content""": datasets.Value("""string""" )} ) , supervised_keys=UpperCAmelCase__ , ) def _lowercase ( self : str , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] ) -> List[Any]: return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""examples""": get_test_dummy_examples()} )] def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Dict ) -> int: import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(UpperCAmelCase__ ) class UpperCamelCase ( datasets.BeamBasedBuilder ): def _lowercase ( self : Optional[int] ) -> Optional[Any]: return datasets.DatasetInfo( features=datasets.Features({"""a""": datasets.Sequence({"""b""": datasets.Value("""string""" )} )} ) , supervised_keys=UpperCAmelCase__ , ) def _lowercase ( self : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Any ) -> Tuple: return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""examples""": get_test_nested_examples()} ) ] def _lowercase ( self : List[str] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] ) -> int: import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(UpperCAmelCase__ ) def lowerCAmelCase__ ( ): '''simple docstring''' return [(i, {"content": content}) for i, content in enumerate(["""foo""", """bar""", """foobar"""] )] def lowerCAmelCase__ ( ): '''simple docstring''' return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["""foo""", """bar""", """foobar"""] )] class UpperCamelCase ( __UpperCAmelCase ): @require_beam def _lowercase ( self : Tuple ) -> Any: _a : int = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _a : Tuple = DummyBeamDataset(cache_dir=UpperCAmelCase__ , beam_runner="""DirectRunner""" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(UpperCAmelCase__ , builder.name , """default""" , """0.0.0""" , f"""{builder.name}-train.arrow""" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"""content""": datasets.Value("""string""" )} ) ) _a : Dict = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , UpperCAmelCase__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , UpperCAmelCase__ ) self.assertDictEqual(dset["""train"""][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["""train"""][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(UpperCAmelCase__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset @require_beam def _lowercase ( self : Optional[Any] ) -> List[Any]: import apache_beam as beam _a : int = beam.io.parquetio.WriteToParquet _a : str = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _a : Union[str, Any] = DummyBeamDataset(cache_dir=UpperCAmelCase__ , beam_runner="""DirectRunner""" ) with patch("""apache_beam.io.parquetio.WriteToParquet""" ) as write_parquet_mock: _a : Tuple = partial(UpperCAmelCase__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( UpperCAmelCase__ , builder.name , """default""" , """0.0.0""" , f"""{builder.name}-train-00000-of-00002.arrow""" ) ) ) self.assertTrue( os.path.exists( os.path.join( UpperCAmelCase__ , builder.name , """default""" , """0.0.0""" , f"""{builder.name}-train-00000-of-00002.arrow""" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"""content""": datasets.Value("""string""" )} ) ) _a : Any = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , UpperCAmelCase__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , UpperCAmelCase__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["""train"""]["""content"""] ) , sorted(["""foo""", """bar""", """foobar"""] ) ) self.assertTrue( os.path.exists(os.path.join(UpperCAmelCase__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset @require_beam def _lowercase ( self : str ) -> List[str]: with tempfile.TemporaryDirectory() as tmp_cache_dir: _a : List[Any] = DummyBeamDataset(cache_dir=UpperCAmelCase__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def _lowercase ( self : str ) -> Optional[int]: _a : str = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _a : str = NestedBeamDataset(cache_dir=UpperCAmelCase__ , beam_runner="""DirectRunner""" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(UpperCAmelCase__ , builder.name , """default""" , """0.0.0""" , f"""{builder.name}-train.arrow""" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"""a""": datasets.Sequence({"""b""": datasets.Value("""string""" )} )} ) ) _a : Any = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , UpperCAmelCase__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , UpperCAmelCase__ ) self.assertDictEqual(dset["""train"""][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["""train"""][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(UpperCAmelCase__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset
294
'''simple docstring''' # Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar lowerCamelCase :str = TypeVar('''T''') class _lowerCAmelCase ( Generic[T] ): def __init__(self , lowercase = True ): A_ : dict[T, list[T]] = {} # dictionary of lists A_ : Any = directed def _a (self , lowercase , lowercase ): if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) self.adj_list[destination_vertex].append(lowercase ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) A_ : Dict = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(lowercase ) A_ : int = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: A_ : Optional[Any] = [destination_vertex] A_ : Tuple = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) A_ : Tuple = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: A_ : Tuple = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: A_ : int = [destination_vertex] A_ : List[str] = [] return self def __repr__(self ): return pformat(self.adj_list )
206
0
import logging import os from .state import PartialState class A ( logging.LoggerAdapter ): @staticmethod def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__, *UpperCamelCase__, **UpperCamelCase__ ): """simple docstring""" if PartialState._shared_state == {}: raise RuntimeError( '''You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.''' ) lowerCAmelCase_ = kwargs.pop('''main_process_only''', UpperCamelCase__ ) lowerCAmelCase_ = kwargs.pop('''in_order''', UpperCamelCase__ ) if self.isEnabledFor(UpperCamelCase__ ): if self._should_log(UpperCamelCase__ ): lowerCAmelCase_ , lowerCAmelCase_ = self.process(UpperCamelCase__, UpperCamelCase__ ) self.logger.log(UpperCamelCase__, UpperCamelCase__, *UpperCamelCase__, **UpperCamelCase__ ) elif in_order: lowerCAmelCase_ = PartialState() for i in range(state.num_processes ): if i == state.process_index: lowerCAmelCase_ , lowerCAmelCase_ = self.process(UpperCamelCase__, UpperCamelCase__ ) self.logger.log(UpperCamelCase__, UpperCamelCase__, *UpperCamelCase__, **UpperCamelCase__ ) state.wait_for_everyone() def __UpperCamelCase ( _A , _A = None ): if log_level is None: lowerCAmelCase_ = os.environ.get('''ACCELERATE_LOG_LEVEL''' , _A ) lowerCAmelCase_ = logging.getLogger(_A ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(_A , {} )
167
def __UpperCamelCase ( _A ): lowerCAmelCase_ = [int(_A ) for i in ip_va_address.split('''.''' ) if i.isdigit()] return len(_A ) == 4 and all(0 <= int(_A ) <= 254 for octet in octets ) if __name__ == "__main__": _A = input().strip() _A = '''valid''' if is_ip_va_address_valid(ip) else '''invalid''' print(f"{ip} is a {valid_or_invalid} IP v4 address.")
167
1
'''simple docstring''' UpperCamelCase_ = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} UpperCamelCase_ = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def lowercase__( __UpperCamelCase: dict[int, list[int]] ,__UpperCamelCase: int ,__UpperCamelCase: list[bool] ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = True SCREAMING_SNAKE_CASE : Dict = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) order.append(__UpperCamelCase ) return order def lowercase__( __UpperCamelCase: dict[int, list[int]] ,__UpperCamelCase: int ,__UpperCamelCase: list[bool] ): """simple docstring""" SCREAMING_SNAKE_CASE : str = True SCREAMING_SNAKE_CASE : Dict = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) return component def lowercase__( __UpperCamelCase: dict[int, list[int]] ): """simple docstring""" SCREAMING_SNAKE_CASE : str = len(__UpperCamelCase ) * [False] SCREAMING_SNAKE_CASE : dict[int, list[int]] = {vert: [] for vert in range(len(__UpperCamelCase ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(__UpperCamelCase ) SCREAMING_SNAKE_CASE : int = [] for i, was_visited in enumerate(__UpperCamelCase ): if not was_visited: order += topology_sort(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) SCREAMING_SNAKE_CASE : Optional[Any] = [] SCREAMING_SNAKE_CASE : Dict = len(__UpperCamelCase ) * [False] for i in range(len(__UpperCamelCase ) ): SCREAMING_SNAKE_CASE : List[str] = order[len(__UpperCamelCase ) - i - 1] if not visited[vert]: SCREAMING_SNAKE_CASE : Tuple = find_components(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) components_list.append(__UpperCamelCase ) return components_list
251
'''simple docstring''' from pathlib import Path import numpy as np from PIL import Image def lowercase__( __UpperCamelCase: np.ndarray ): """simple docstring""" SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.2_9_8_9 * r + 0.5_8_7_0 * g + 0.1_1_4_0 * b def lowercase__( __UpperCamelCase: np.ndarray ): """simple docstring""" return (gray > 1_27) & (gray <= 2_55) def lowercase__( __UpperCamelCase: np.ndarray ,__UpperCamelCase: np.ndarray ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = np.zeros_like(__UpperCamelCase ) SCREAMING_SNAKE_CASE : Union[str, Any] = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image SCREAMING_SNAKE_CASE : List[str] = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): SCREAMING_SNAKE_CASE : Union[str, Any] = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() SCREAMING_SNAKE_CASE : Tuple = int(summation > 0 ) return output if __name__ == "__main__": # read original image UpperCamelCase_ = Path(__file__).resolve().parent / "image_data" / "lena.jpg" UpperCamelCase_ = np.array(Image.open(lena_path)) # kernel to be applied UpperCamelCase_ = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) UpperCamelCase_ = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image UpperCamelCase_ = Image.fromarray(output).convert("RGB") pil_img.save("result_dilation.png")
251
1
import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaTokenizer from transformers.models.mam_aaa.modeling_mam_aaa import MaMaaaDecoder, MaMaaaEncoder def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) ->Optional[int]: """simple docstring""" if attention_mask is None: lowerCAmelCase__ :int = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCAmelCase__ :Optional[Any] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCAmelCase__ :Union[str, Any] = torch.ones(config.encoder_layers , config.encoder_attention_heads , device=_UpperCAmelCase ) if decoder_head_mask is None: lowerCAmelCase__ :Any = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=_UpperCAmelCase ) if cross_attn_head_mask is None: lowerCAmelCase__ :Union[str, Any] = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=_UpperCAmelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } class _lowerCAmelCase : """simple docstring""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=1_3 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=9_9 , __UpperCAmelCase=1_6 , __UpperCAmelCase=2 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase="relu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase=2_0 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , ): '''simple docstring''' lowerCAmelCase__ :List[str] = parent lowerCAmelCase__ :Union[str, Any] = batch_size lowerCAmelCase__ :Union[str, Any] = seq_length lowerCAmelCase__ :str = is_training lowerCAmelCase__ :Optional[Any] = use_labels lowerCAmelCase__ :Dict = vocab_size lowerCAmelCase__ :int = hidden_size lowerCAmelCase__ :Dict = num_hidden_layers lowerCAmelCase__ :List[str] = num_attention_heads lowerCAmelCase__ :Dict = intermediate_size lowerCAmelCase__ :Dict = hidden_act lowerCAmelCase__ :Optional[int] = hidden_dropout_prob lowerCAmelCase__ :Union[str, Any] = attention_probs_dropout_prob lowerCAmelCase__ :str = encoder_layerdrop lowerCAmelCase__ :Union[str, Any] = decoder_layerdrop lowerCAmelCase__ :List[Any] = max_position_embeddings lowerCAmelCase__ :str = eos_token_id lowerCAmelCase__ :List[Any] = pad_token_id lowerCAmelCase__ :str = bos_token_id def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ :Any = self.eos_token_id # Eos Token lowerCAmelCase__ :Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for M2M100 the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCAmelCase__ :List[Any] = input_ids.clamp(self.pad_token_id + 1 ) lowerCAmelCase__ :Tuple = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCAmelCase__ :Any = self.get_config() lowerCAmelCase__ :int = prepare_mam_aaa_inputs_dict(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return config, inputs_dict def snake_case ( self ): '''simple docstring''' return MaMaaaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , encoder_layerdrop=self.encoder_layerdrop , decoder_layerdrop=self.decoder_layerdrop , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = self.prepare_config_and_inputs() return config, inputs_dict def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :int = MaMaaaModel(config=__UpperCAmelCase ).get_decoder().to(__UpperCAmelCase ).eval() lowerCAmelCase__ :Tuple = inputs_dict['input_ids'] lowerCAmelCase__ :str = inputs_dict['attention_mask'] lowerCAmelCase__ :Tuple = inputs_dict['head_mask'] # first forward pass lowerCAmelCase__ :int = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , head_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase ) lowerCAmelCase__ :str = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids lowerCAmelCase__ :str = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCAmelCase__ :Dict = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and lowerCAmelCase__ :Optional[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCAmelCase__ :List[Any] = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) lowerCAmelCase__ :str = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )['last_hidden_state'] lowerCAmelCase__ :Any = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase )[ 'last_hidden_state' ] # select random slice lowerCAmelCase__ :Optional[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCAmelCase__ :Tuple = output_from_no_past[:, -3:, random_slice_idx].detach() lowerCAmelCase__ :Optional[int] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-2 ) ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = MaMaaaModel(config=__UpperCAmelCase ).to(__UpperCAmelCase ).eval() lowerCAmelCase__ :Any = model(**__UpperCAmelCase ) lowerCAmelCase__ :Optional[Any] = outputs.encoder_last_hidden_state lowerCAmelCase__ :List[str] = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ :List[Any] = model.get_encoder() encoder.save_pretrained(__UpperCAmelCase ) lowerCAmelCase__ :Optional[int] = MaMaaaEncoder.from_pretrained(__UpperCAmelCase ).to(__UpperCAmelCase ) lowerCAmelCase__ :List[Any] = encoder(inputs_dict['input_ids'] , attention_mask=inputs_dict['attention_mask'] )[ 0 ] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1E-3 ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ :int = model.get_decoder() decoder.save_pretrained(__UpperCAmelCase ) lowerCAmelCase__ :Dict = MaMaaaDecoder.from_pretrained(__UpperCAmelCase ).to(__UpperCAmelCase ) lowerCAmelCase__ :List[Any] = decoder( input_ids=inputs_dict['decoder_input_ids'] , attention_mask=inputs_dict['decoder_attention_mask'] , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=inputs_dict['attention_mask'] , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1E-3 ) @require_torch class _lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , unittest.TestCase ): """simple docstring""" __magic_name__ :Optional[Any] = ( ( MaMaaaModel, MaMaaaForConditionalGeneration, ) if is_torch_available() else () ) __magic_name__ :List[Any] = (MaMaaaForConditionalGeneration,) if is_torch_available() else () __magic_name__ :List[str] = ( { """conversational""": MaMaaaForConditionalGeneration, """feature-extraction""": MaMaaaModel, """summarization""": MaMaaaForConditionalGeneration, """text2text-generation""": MaMaaaForConditionalGeneration, """translation""": MaMaaaForConditionalGeneration, } if is_torch_available() else {} ) __magic_name__ :List[Any] = True __magic_name__ :Optional[Any] = True __magic_name__ :List[Any] = False __magic_name__ :List[Any] = False def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if pipeline_test_casse_name == "TranslationPipelineTests": # Get `ValueError: Translation requires a `src_lang` and a `tgt_lang` for this model`. # `M2M100Config` was never used in pipeline tests: cannot create a simple tokenizer. return True return False def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = MaMaaaModelTester(self ) lowerCAmelCase__ :Optional[Any] = ConfigTester(self , config_class=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' self.config_tester.run_common_tests() def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: lowerCAmelCase__ :Optional[Any] = model_class(__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCAmelCase ) lowerCAmelCase__ :Optional[Any] = model_class.from_pretrained(__UpperCAmelCase , output_loading_info=__UpperCAmelCase ) self.assertEqual(info['missing_keys'] , [] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in (MaMaaaModel, MaMaaaForConditionalGeneration): lowerCAmelCase__ :Dict = model_class(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ :List[Any] = copy.deepcopy(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) if not self.is_encoder_decoder: lowerCAmelCase__ :Union[str, Any] = inputs['input_ids'] del inputs["input_ids"] else: lowerCAmelCase__ :Optional[int] = inputs['input_ids'] lowerCAmelCase__ :Any = inputs.get('decoder_input_ids' , __UpperCAmelCase ) del inputs["input_ids"] inputs.pop('decoder_input_ids' , __UpperCAmelCase ) lowerCAmelCase__ :Any = model.get_input_embeddings() if not self.is_encoder_decoder: lowerCAmelCase__ :List[str] = wte(__UpperCAmelCase ) else: lowerCAmelCase__ :Optional[Any] = wte(__UpperCAmelCase ) lowerCAmelCase__ :Any = wte(__UpperCAmelCase ) with torch.no_grad(): model(**__UpperCAmelCase )[0] def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = self.model_tester.prepare_config_and_inputs() lowerCAmelCase__ :Optional[Any] = input_dict['input_ids'] lowerCAmelCase__ :Optional[int] = input_ids.ne(1 ).to(__UpperCAmelCase ) lowerCAmelCase__ :List[Any] = MaMaaaForConditionalGeneration(__UpperCAmelCase ).eval().to(__UpperCAmelCase ) if torch_device == "cuda": model.half() model.generate(__UpperCAmelCase , attention_mask=__UpperCAmelCase ) model.generate(num_beams=4 , do_sample=__UpperCAmelCase , early_stopping=__UpperCAmelCase , num_return_sequences=3 ) def __A (_SCREAMING_SNAKE_CASE ) ->Optional[Any]: """simple docstring""" return torch.tensor(_UpperCAmelCase , dtype=torch.long , device=_UpperCAmelCase ) __A = 1e-4 @require_torch @require_sentencepiece @require_tokenizers @slow class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case ( self ): '''simple docstring''' return MaMaaaTokenizer.from_pretrained('facebook/m2m100_418M' ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = MaMaaaModel.from_pretrained('facebook/m2m100_418M' ).to(__UpperCAmelCase ) lowerCAmelCase__ :List[str] = _long_tensor([[1_2_8_0_2_8, 9_8, 1_2, 3_0_5_2_7, 2_7_3_2, 1_5_9, 7_7_5_5, 6_1_9_0_4, 3_9_1_4_4, 3_8, 2]] ) lowerCAmelCase__ :Dict = _long_tensor([[2, 1_2_8_0_2_8, 9_8, 1_2, 3_0_5_2_7, 2_7_3_2, 1_5_9, 7_7_5_5, 6_1_9_0_4, 3_9_1_4_4, 3_8]] ) lowerCAmelCase__ :str = prepare_mam_aaa_inputs_dict(model.config , __UpperCAmelCase , __UpperCAmelCase ) with torch.no_grad(): lowerCAmelCase__ :Union[str, Any] = model(**__UpperCAmelCase )[0] lowerCAmelCase__ :str = torch.Size((1, 1_1, 1_0_2_4) ) self.assertEqual(output.shape , __UpperCAmelCase ) # change to expected output here lowerCAmelCase__ :int = torch.tensor( [[-0.77_80, -0.16_76, 0.10_38], [-6.75_56, -1.39_92, 0.05_67], [-7.53_83, -0.59_20, -0.27_79]] , device=__UpperCAmelCase ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=__UpperCAmelCase ) ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Any = MaMaaaForConditionalGeneration.from_pretrained('facebook/m2m100_418M' ).to(__UpperCAmelCase ) # change to intended input lowerCAmelCase__ :List[Any] = _long_tensor([[1_2_8_0_2_8, 9_8, 1_2, 3_0_5_2_7, 2_7_3_2, 1_5_9, 7_7_5_5, 6_1_9_0_4, 3_9_1_4_4, 3_8, 2]] ) lowerCAmelCase__ :Tuple = _long_tensor([[2, 1_2_8_0_2_8, 9_8, 1_2, 3_0_5_2_7, 2_7_3_2, 1_5_9, 7_7_5_5, 6_1_9_0_4, 3_9_1_4_4, 3_8]] ) lowerCAmelCase__ :Tuple = prepare_mam_aaa_inputs_dict(model.config , __UpperCAmelCase , __UpperCAmelCase ) with torch.no_grad(): lowerCAmelCase__ :Optional[int] = model(**__UpperCAmelCase )[0] lowerCAmelCase__ :List[str] = torch.Size((1, 1_1, model.config.vocab_size) ) self.assertEqual(output.shape , __UpperCAmelCase ) # change to expected output here lowerCAmelCase__ :Dict = torch.tensor( [[-1.04_48, -1.04_11, 3.79_92], [-3.21_91, -3.23_86, -1.34_51], [-3.62_10, -3.59_93, 0.49_25]] , device=__UpperCAmelCase ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=__UpperCAmelCase ) ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = MaMaaaForConditionalGeneration.from_pretrained('facebook/m2m100_418M' ).to(__UpperCAmelCase ) lowerCAmelCase__ :Dict = MaMaaaTokenizer.from_pretrained('facebook/m2m100_418M' , src_lang='fr' , tgt_lang='en' ) lowerCAmelCase__ :int = [ 'L\'affaire NSA souligne l\'absence totale de débat sur le renseignement', 'Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.', 'Lorsque François Hollande téléphone à Barack Obama ou quand le ministre des affaires étrangères Laurent' ' Fabius convoque l\'ambassadeur des Etats-Unis, ils réagissent à une vraie découverte, qui est celle de' ' l\'ampleur de la surveillance américaine sur l\'ensemble des communications en France.', ] # The below article tests that we don't add any hypotheses outside of the top n_beams lowerCAmelCase__ :Dict = tokenizer(__UpperCAmelCase , padding=__UpperCAmelCase , return_tensors='pt' ) lowerCAmelCase__ :Optional[Any] = model.generate( input_ids=dct['input_ids'].to(__UpperCAmelCase ) , attention_mask=dct['attention_mask'].to(__UpperCAmelCase ) , num_beams=5 , forced_bos_token_id=tokenizer.get_lang_id('en' ) , ) lowerCAmelCase__ :Optional[Any] = [ 'The NSA case highlights the total absence of intelligence debate', 'I think there are two levels of response from the French government.', 'When François Hollande calls Barack Obama or when Foreign Minister Laurent Fabius calls the U.S.' ' Ambassador, they respond to a real discovery, which is that of the scale of U.S. surveillance on all' ' communications in France.', ] lowerCAmelCase__ :str = tokenizer.batch_decode( hypotheses_batch.tolist() , clean_up_tokenization_spaces=__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) assert generated == expected_en
351
"""simple docstring""" import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaTokenizer from transformers.models.mam_aaa.modeling_mam_aaa import MaMaaaDecoder, MaMaaaEncoder def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) ->str: """simple docstring""" if attention_mask is None: lowerCAmelCase__ :List[str] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCAmelCase__ :Tuple = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCAmelCase__ :Union[str, Any] = torch.ones(config.encoder_layers , config.encoder_attention_heads , device=_SCREAMING_SNAKE_CASE ) if decoder_head_mask is None: lowerCAmelCase__ :List[str] = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=_SCREAMING_SNAKE_CASE ) if cross_attn_head_mask is None: lowerCAmelCase__ :List[str] = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=_SCREAMING_SNAKE_CASE ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } class _lowerCAmelCase : """simple docstring""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=1_3 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=9_9 , __UpperCAmelCase=1_6 , __UpperCAmelCase=2 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase="relu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase=2_0 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , ): '''simple docstring''' lowerCAmelCase__ :List[Any] = parent lowerCAmelCase__ :Any = batch_size lowerCAmelCase__ :Optional[Any] = seq_length lowerCAmelCase__ :int = is_training lowerCAmelCase__ :Tuple = use_labels lowerCAmelCase__ :Union[str, Any] = vocab_size lowerCAmelCase__ :Tuple = hidden_size lowerCAmelCase__ :Tuple = num_hidden_layers lowerCAmelCase__ :Tuple = num_attention_heads lowerCAmelCase__ :Dict = intermediate_size lowerCAmelCase__ :Optional[int] = hidden_act lowerCAmelCase__ :Any = hidden_dropout_prob lowerCAmelCase__ :Dict = attention_probs_dropout_prob lowerCAmelCase__ :Tuple = encoder_layerdrop lowerCAmelCase__ :Tuple = decoder_layerdrop lowerCAmelCase__ :Tuple = max_position_embeddings lowerCAmelCase__ :Any = eos_token_id lowerCAmelCase__ :str = pad_token_id lowerCAmelCase__ :Tuple = bos_token_id def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ :Tuple = self.eos_token_id # Eos Token lowerCAmelCase__ :Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for M2M100 the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCAmelCase__ :List[Any] = input_ids.clamp(self.pad_token_id + 1 ) lowerCAmelCase__ :Dict = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCAmelCase__ :Optional[Any] = self.get_config() lowerCAmelCase__ :Any = prepare_mam_aaa_inputs_dict(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return config, inputs_dict def snake_case ( self ): '''simple docstring''' return MaMaaaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , encoder_layerdrop=self.encoder_layerdrop , decoder_layerdrop=self.decoder_layerdrop , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :List[Any] = self.prepare_config_and_inputs() return config, inputs_dict def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :List[Any] = MaMaaaModel(config=__UpperCAmelCase ).get_decoder().to(__UpperCAmelCase ).eval() lowerCAmelCase__ :Optional[int] = inputs_dict['input_ids'] lowerCAmelCase__ :Any = inputs_dict['attention_mask'] lowerCAmelCase__ :Tuple = inputs_dict['head_mask'] # first forward pass lowerCAmelCase__ :int = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , head_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase ) lowerCAmelCase__ , lowerCAmelCase__ :Union[str, Any] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids lowerCAmelCase__ :Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCAmelCase__ :int = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and lowerCAmelCase__ :Tuple = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCAmelCase__ :Union[str, Any] = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) lowerCAmelCase__ :Dict = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )['last_hidden_state'] lowerCAmelCase__ :Optional[Any] = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase )[ 'last_hidden_state' ] # select random slice lowerCAmelCase__ :Any = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCAmelCase__ :List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() lowerCAmelCase__ :Union[str, Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-2 ) ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Tuple = MaMaaaModel(config=__UpperCAmelCase ).to(__UpperCAmelCase ).eval() lowerCAmelCase__ :List[Any] = model(**__UpperCAmelCase ) lowerCAmelCase__ :int = outputs.encoder_last_hidden_state lowerCAmelCase__ :Any = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ :Union[str, Any] = model.get_encoder() encoder.save_pretrained(__UpperCAmelCase ) lowerCAmelCase__ :Any = MaMaaaEncoder.from_pretrained(__UpperCAmelCase ).to(__UpperCAmelCase ) lowerCAmelCase__ :Any = encoder(inputs_dict['input_ids'] , attention_mask=inputs_dict['attention_mask'] )[ 0 ] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1E-3 ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ :Optional[int] = model.get_decoder() decoder.save_pretrained(__UpperCAmelCase ) lowerCAmelCase__ :Dict = MaMaaaDecoder.from_pretrained(__UpperCAmelCase ).to(__UpperCAmelCase ) lowerCAmelCase__ :int = decoder( input_ids=inputs_dict['decoder_input_ids'] , attention_mask=inputs_dict['decoder_attention_mask'] , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=inputs_dict['attention_mask'] , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1E-3 ) @require_torch class _lowerCAmelCase ( a , a , a , unittest.TestCase ): """simple docstring""" __magic_name__ :Optional[int] = ( ( MaMaaaModel, MaMaaaForConditionalGeneration, ) if is_torch_available() else () ) __magic_name__ :str = (MaMaaaForConditionalGeneration,) if is_torch_available() else () __magic_name__ :str = ( { """conversational""": MaMaaaForConditionalGeneration, """feature-extraction""": MaMaaaModel, """summarization""": MaMaaaForConditionalGeneration, """text2text-generation""": MaMaaaForConditionalGeneration, """translation""": MaMaaaForConditionalGeneration, } if is_torch_available() else {} ) __magic_name__ :Any = True __magic_name__ :Union[str, Any] = True __magic_name__ :Tuple = False __magic_name__ :List[str] = False def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if pipeline_test_casse_name == "TranslationPipelineTests": # Get `ValueError: Translation requires a `src_lang` and a `tgt_lang` for this model`. # `M2M100Config` was never used in pipeline tests: cannot create a simple tokenizer. return True return False def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = MaMaaaModelTester(self ) lowerCAmelCase__ :Tuple = ConfigTester(self , config_class=__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' self.config_tester.run_common_tests() def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :str = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: lowerCAmelCase__ :str = model_class(__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCAmelCase ) lowerCAmelCase__ , lowerCAmelCase__ :List[Any] = model_class.from_pretrained(__UpperCAmelCase , output_loading_info=__UpperCAmelCase ) self.assertEqual(info['missing_keys'] , [] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in (MaMaaaModel, MaMaaaForConditionalGeneration): lowerCAmelCase__ :Optional[int] = model_class(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ :List[Any] = copy.deepcopy(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) if not self.is_encoder_decoder: lowerCAmelCase__ :List[str] = inputs['input_ids'] del inputs["input_ids"] else: lowerCAmelCase__ :int = inputs['input_ids'] lowerCAmelCase__ :str = inputs.get('decoder_input_ids' , __UpperCAmelCase ) del inputs["input_ids"] inputs.pop('decoder_input_ids' , __UpperCAmelCase ) lowerCAmelCase__ :List[Any] = model.get_input_embeddings() if not self.is_encoder_decoder: lowerCAmelCase__ :Tuple = wte(__UpperCAmelCase ) else: lowerCAmelCase__ :List[Any] = wte(__UpperCAmelCase ) lowerCAmelCase__ :Dict = wte(__UpperCAmelCase ) with torch.no_grad(): model(**__UpperCAmelCase )[0] def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ :int = self.model_tester.prepare_config_and_inputs() lowerCAmelCase__ :Any = input_dict['input_ids'] lowerCAmelCase__ :Optional[Any] = input_ids.ne(1 ).to(__UpperCAmelCase ) lowerCAmelCase__ :List[Any] = MaMaaaForConditionalGeneration(__UpperCAmelCase ).eval().to(__UpperCAmelCase ) if torch_device == "cuda": model.half() model.generate(__UpperCAmelCase , attention_mask=__UpperCAmelCase ) model.generate(num_beams=4 , do_sample=__UpperCAmelCase , early_stopping=__UpperCAmelCase , num_return_sequences=3 ) def __A (_SCREAMING_SNAKE_CASE ) ->Dict: """simple docstring""" return torch.tensor(_SCREAMING_SNAKE_CASE , dtype=torch.long , device=_SCREAMING_SNAKE_CASE ) __A = 1e-4 @require_torch @require_sentencepiece @require_tokenizers @slow class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case ( self ): '''simple docstring''' return MaMaaaTokenizer.from_pretrained('facebook/m2m100_418M' ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = MaMaaaModel.from_pretrained('facebook/m2m100_418M' ).to(__UpperCAmelCase ) lowerCAmelCase__ :str = _long_tensor([[1_2_8_0_2_8, 9_8, 1_2, 3_0_5_2_7, 2_7_3_2, 1_5_9, 7_7_5_5, 6_1_9_0_4, 3_9_1_4_4, 3_8, 2]] ) lowerCAmelCase__ :Optional[int] = _long_tensor([[2, 1_2_8_0_2_8, 9_8, 1_2, 3_0_5_2_7, 2_7_3_2, 1_5_9, 7_7_5_5, 6_1_9_0_4, 3_9_1_4_4, 3_8]] ) lowerCAmelCase__ :Union[str, Any] = prepare_mam_aaa_inputs_dict(model.config , __UpperCAmelCase , __UpperCAmelCase ) with torch.no_grad(): lowerCAmelCase__ :Any = model(**__UpperCAmelCase )[0] lowerCAmelCase__ :List[str] = torch.Size((1, 1_1, 1_0_2_4) ) self.assertEqual(output.shape , __UpperCAmelCase ) # change to expected output here lowerCAmelCase__ :int = torch.tensor( [[-0.77_80, -0.16_76, 0.10_38], [-6.75_56, -1.39_92, 0.05_67], [-7.53_83, -0.59_20, -0.27_79]] , device=__UpperCAmelCase ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=__UpperCAmelCase ) ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = MaMaaaForConditionalGeneration.from_pretrained('facebook/m2m100_418M' ).to(__UpperCAmelCase ) # change to intended input lowerCAmelCase__ :str = _long_tensor([[1_2_8_0_2_8, 9_8, 1_2, 3_0_5_2_7, 2_7_3_2, 1_5_9, 7_7_5_5, 6_1_9_0_4, 3_9_1_4_4, 3_8, 2]] ) lowerCAmelCase__ :Any = _long_tensor([[2, 1_2_8_0_2_8, 9_8, 1_2, 3_0_5_2_7, 2_7_3_2, 1_5_9, 7_7_5_5, 6_1_9_0_4, 3_9_1_4_4, 3_8]] ) lowerCAmelCase__ :List[Any] = prepare_mam_aaa_inputs_dict(model.config , __UpperCAmelCase , __UpperCAmelCase ) with torch.no_grad(): lowerCAmelCase__ :List[Any] = model(**__UpperCAmelCase )[0] lowerCAmelCase__ :Any = torch.Size((1, 1_1, model.config.vocab_size) ) self.assertEqual(output.shape , __UpperCAmelCase ) # change to expected output here lowerCAmelCase__ :List[Any] = torch.tensor( [[-1.04_48, -1.04_11, 3.79_92], [-3.21_91, -3.23_86, -1.34_51], [-3.62_10, -3.59_93, 0.49_25]] , device=__UpperCAmelCase ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=__UpperCAmelCase ) ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = MaMaaaForConditionalGeneration.from_pretrained('facebook/m2m100_418M' ).to(__UpperCAmelCase ) lowerCAmelCase__ :Union[str, Any] = MaMaaaTokenizer.from_pretrained('facebook/m2m100_418M' , src_lang='fr' , tgt_lang='en' ) lowerCAmelCase__ :Tuple = [ 'L\'affaire NSA souligne l\'absence totale de débat sur le renseignement', 'Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.', 'Lorsque François Hollande téléphone à Barack Obama ou quand le ministre des affaires étrangères Laurent' ' Fabius convoque l\'ambassadeur des Etats-Unis, ils réagissent à une vraie découverte, qui est celle de' ' l\'ampleur de la surveillance américaine sur l\'ensemble des communications en France.', ] # The below article tests that we don't add any hypotheses outside of the top n_beams lowerCAmelCase__ :Union[str, Any] = tokenizer(__UpperCAmelCase , padding=__UpperCAmelCase , return_tensors='pt' ) lowerCAmelCase__ :List[Any] = model.generate( input_ids=dct['input_ids'].to(__UpperCAmelCase ) , attention_mask=dct['attention_mask'].to(__UpperCAmelCase ) , num_beams=5 , forced_bos_token_id=tokenizer.get_lang_id('en' ) , ) lowerCAmelCase__ :Optional[Any] = [ 'The NSA case highlights the total absence of intelligence debate', 'I think there are two levels of response from the French government.', 'When François Hollande calls Barack Obama or when Foreign Minister Laurent Fabius calls the U.S.' ' Ambassador, they respond to a real discovery, which is that of the scale of U.S. surveillance on all' ' communications in France.', ] lowerCAmelCase__ :Any = tokenizer.batch_decode( hypotheses_batch.tolist() , clean_up_tokenization_spaces=__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) assert generated == expected_en
254
0
from ..utils import DummyObject, requires_backends class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[Any] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : int = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[Any] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : int = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[int] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Union[str, Any] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[int] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[Any] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : str = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Any = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Any = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Any = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Union[str, Any] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Any = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[int] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[Any] = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : str = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : int = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""sentencepiece"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(self , ['''sentencepiece'''] )
348
from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def lowerCAmelCase_ ( )-> int: '''simple docstring''' UpperCAmelCase : str ={ '''repo_name''': ['''test_repo1''', '''test_repo2''', '''test_repo3'''], '''path''': ['''test_1.py''', '''test_2.py''', '''unit_test.py'''], '''content''': ['''a ''' * 20, '''a ''' * 30, '''b ''' * 7], } UpperCAmelCase : Union[str, Any] =Dataset.from_dict(__lowerCAmelCase ) return dataset class __snake_case ( lowerCamelCase__ ): def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : List[str] =get_dataset() UpperCAmelCase : Optional[int] =make_duplicate_clusters(snake_case__ , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : str =get_dataset() UpperCAmelCase , UpperCAmelCase : Tuple =deduplicate_dataset(snake_case__ ) self.assertEqual(len(snake_case__ ) , 2 ) print(snake_case__ ) self.assertEqual(duplicate_clusters[0][0]['''copies'''] , 2 ) self.assertEqual(duplicate_clusters[0][0]['''is_extreme'''] , snake_case__ )
348
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ : Any = { '''configuration_clipseg''': [ '''CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CLIPSegConfig''', '''CLIPSegTextConfig''', '''CLIPSegVisionConfig''', ], '''processing_clipseg''': ['''CLIPSegProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : List[Any] = [ '''CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CLIPSegModel''', '''CLIPSegPreTrainedModel''', '''CLIPSegTextModel''', '''CLIPSegVisionModel''', '''CLIPSegForImageSegmentation''', ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys lowercase__ : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
190
'''simple docstring''' import argparse lowercase__ : Any = '''docs/source/_static/js/custom.js''' def _lowerCAmelCase ( __snake_case : Union[str, Any] ) -> str: with open(__snake_case , encoding='utf-8' , newline='\n' ) as f: __A : Optional[Any] = f.readlines() __A : List[str] = 0 # First let's put the right version while not lines[index].startswith('const stableVersion =' ): index += 1 __A : Tuple = f'const stableVersion = "v{version}"\n' # Then update the dictionary while not lines[index].startswith('const versionMapping = {' ): index += 1 # We go until the end while not lines[index].startswith('}' ): index += 1 # We add the new version at the end lines[index - 1] += f' "v{version}": "v{version}",\n' with open(__snake_case , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(__snake_case ) if __name__ == "__main__": lowercase__ : Optional[Any] = argparse.ArgumentParser() parser.add_argument('''--version''', help='''Release version.''') lowercase__ : List[str] = parser.parse_args() update_custom_js(args.version)
190
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig __snake_case : Any = { 'google/tapas-base-finetuned-sqa': ( 'https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json' ), 'google/tapas-base-finetuned-wtq': ( 'https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json' ), 'google/tapas-base-finetuned-wikisql-supervised': ( 'https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json' ), 'google/tapas-base-finetuned-tabfact': ( 'https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json' ), } class lowerCamelCase ( lowercase_ ): '''simple docstring''' __snake_case = 'tapas' def __init__( self : List[Any] , lowerCAmelCase_ : Union[str, Any]=3_05_22 , lowerCAmelCase_ : Tuple=7_68 , lowerCAmelCase_ : Dict=12 , lowerCAmelCase_ : str=12 , lowerCAmelCase_ : Tuple=30_72 , lowerCAmelCase_ : Optional[Any]="gelu" , lowerCAmelCase_ : str=0.1 , lowerCAmelCase_ : Optional[int]=0.1 , lowerCAmelCase_ : List[Any]=10_24 , lowerCAmelCase_ : List[Any]=[3, 2_56, 2_56, 2, 2_56, 2_56, 10] , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : List[Any]=1e-12 , lowerCAmelCase_ : Any=0 , lowerCAmelCase_ : Dict=10.0 , lowerCAmelCase_ : str=0 , lowerCAmelCase_ : Optional[Any]=1.0 , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : int=1.0 , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : Optional[Any]=1.0 , lowerCAmelCase_ : List[str]=1.0 , lowerCAmelCase_ : Optional[int]=False , lowerCAmelCase_ : Dict=False , lowerCAmelCase_ : Dict="ratio" , lowerCAmelCase_ : int=None , lowerCAmelCase_ : str=None , lowerCAmelCase_ : Tuple=64 , lowerCAmelCase_ : Dict=32 , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : int=True , lowerCAmelCase_ : List[str]=False , lowerCAmelCase_ : Dict=False , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : int=False , lowerCAmelCase_ : Optional[Any]=None , lowerCAmelCase_ : Tuple=None , **lowerCAmelCase_ : List[str] , ) -> Dict: '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) A__ : Tuple =vocab_size A__ : Dict =hidden_size A__ : int =num_hidden_layers A__ : Optional[Any] =num_attention_heads A__ : str =hidden_act A__ : str =intermediate_size A__ : Union[str, Any] =hidden_dropout_prob A__ : str =attention_probs_dropout_prob A__ : int =max_position_embeddings A__ : Tuple =type_vocab_sizes A__ : List[str] =initializer_range A__ : List[str] =layer_norm_eps # Fine-tuning task hyperparameters A__ : Optional[int] =positive_label_weight A__ : Optional[Any] =num_aggregation_labels A__ : Tuple =aggregation_loss_weight A__ : Optional[Any] =use_answer_as_supervision A__ : List[Any] =answer_loss_importance A__ : Union[str, Any] =use_normalized_answer_loss A__ : Any =huber_loss_delta A__ : Any =temperature A__ : Tuple =aggregation_temperature A__ : str =use_gumbel_for_cells A__ : str =use_gumbel_for_aggregation A__ : Any =average_approximation_function A__ : Dict =cell_selection_preference A__ : str =answer_loss_cutoff A__ : Dict =max_num_rows A__ : Tuple =max_num_columns A__ : int =average_logits_per_cell A__ : int =select_one_column A__ : str =allow_empty_column_selection A__ : List[str] =init_cell_selection_weights_to_zero A__ : List[Any] =reset_position_index_per_cell A__ : List[str] =disable_per_token_loss # Aggregation hyperparameters A__ : int =aggregation_labels A__ : Any =no_aggregation_label_index if isinstance(self.aggregation_labels , lowerCAmelCase_ ): A__ : str ={int(lowerCAmelCase_ ): v for k, v in aggregation_labels.items()}
134
'''simple docstring''' from graphs.minimum_spanning_tree_kruskal import kruskal def __lowerCamelCase ( ) -> int: """simple docstring""" A__ : int =9 A__ : int =[ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] A__ : Optional[Any] =kruskal(__snake_case, __snake_case ) A__ : List[str] =[ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] assert sorted(__snake_case ) == sorted(__snake_case )
134
1
"""simple docstring""" import unittest from transformers.testing_utils import require_bsa from transformers.utils import is_bsa_available from ...test_feature_extraction_common import FeatureExtractionSavingTestMixin if is_bsa_available(): from transformers import MarkupLMFeatureExtractor class UpperCamelCase_ ( unittest.TestCase): """simple docstring""" def __init__( self : Any , UpperCAmelCase__ : Optional[Any] ) -> Tuple: __SCREAMING_SNAKE_CASE = parent def UpperCAmelCase_ ( self : int ) -> Tuple: return {} def UpperCAmelCase__ (): '''simple docstring''' __SCREAMING_SNAKE_CASE = "<HTML>\n\n <HEAD>\n <TITLE>sample document</TITLE>\n </HEAD>\n\n <BODY BGCOLOR=\"FFFFFF\">\n <HR>\n <a href=\"http://google.com\">Goog</a>\n <H1>This is one header</H1>\n <H2>This is a another Header</H2>\n <P>Travel from\n <P>\n <B>SFO to JFK</B>\n <BR>\n <B><I>on May 2, 2015 at 2:00 pm. For details go to confirm.com </I></B>\n <HR>\n <div style=\"color:#0000FF\">\n <h3>Traveler <b> name </b> is\n <p> John Doe </p>\n </div>" __SCREAMING_SNAKE_CASE = "\n <!DOCTYPE html>\n <html>\n <body>\n\n <h1>My First Heading</h1>\n <p>My first paragraph.</p>\n\n </body>\n </html>\n " return [html_string_a, html_string_a] @require_bsa class UpperCamelCase_ ( UpperCamelCase , unittest.TestCase): """simple docstring""" snake_case__ : Optional[int] = MarkupLMFeatureExtractor if is_bsa_available() else None def UpperCAmelCase_ ( self : List[Any] ) -> Any: __SCREAMING_SNAKE_CASE = MarkupLMFeatureExtractionTester(self ) @property def UpperCAmelCase_ ( self : List[str] ) -> Dict: return self.feature_extract_tester.prepare_feat_extract_dict() def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: # Initialize feature_extractor __SCREAMING_SNAKE_CASE = self.feature_extraction_class() # Test not batched input __SCREAMING_SNAKE_CASE = get_html_strings()[0] __SCREAMING_SNAKE_CASE = feature_extractor(UpperCAmelCase__ ) # fmt: off __SCREAMING_SNAKE_CASE = [["sample document", "Goog", "This is one header", "This is a another Header", "Travel from", "SFO to JFK", "on May 2, 2015 at 2:00 pm. For details go to confirm.com", "Traveler", "name", "is", "John Doe"]] __SCREAMING_SNAKE_CASE = [["/html/head/title", "/html/body/a", "/html/body/h1", "/html/body/h2", "/html/body/p", "/html/body/p/p/b[1]", "/html/body/p/p/b[2]/i", "/html/body/p/p/div/h3", "/html/body/p/p/div/h3/b", "/html/body/p/p/div/h3", "/html/body/p/p/div/h3/p"]] # fmt: on self.assertEqual(encoding.nodes , UpperCAmelCase__ ) self.assertEqual(encoding.xpaths , UpperCAmelCase__ ) # Test batched __SCREAMING_SNAKE_CASE = get_html_strings() __SCREAMING_SNAKE_CASE = feature_extractor(UpperCAmelCase__ ) # fmt: off __SCREAMING_SNAKE_CASE = expected_nodes + [["My First Heading", "My first paragraph."]] __SCREAMING_SNAKE_CASE = expected_xpaths + [["/html/body/h1", "/html/body/p"]] self.assertEqual(len(encoding.nodes ) , 2 ) self.assertEqual(len(encoding.xpaths ) , 2 ) self.assertEqual(encoding.nodes , UpperCAmelCase__ ) self.assertEqual(encoding.xpaths , UpperCAmelCase__ )
195
"""simple docstring""" from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch a__ : Optional[int] = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Optional[Any] = ["pixel_values"] def __init__( self : Dict , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Union[int, float] = 1 / 2_5_5 , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : bool = True , **UpperCAmelCase__ : Dict , ) -> None: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = size if size is not None else {"shortest_edge": 2_2_4} __SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase__ , default_to_square=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = crop_size if crop_size is not None else {"height": 2_5_6, "width": 2_5_6} __SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase__ , param_name="crop_size" ) __SCREAMING_SNAKE_CASE = do_resize __SCREAMING_SNAKE_CASE = size __SCREAMING_SNAKE_CASE = resample __SCREAMING_SNAKE_CASE = do_rescale __SCREAMING_SNAKE_CASE = rescale_factor __SCREAMING_SNAKE_CASE = do_center_crop __SCREAMING_SNAKE_CASE = crop_size __SCREAMING_SNAKE_CASE = do_flip_channel_order def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Dict[str, int] , UpperCAmelCase__ : PILImageResampling = PIL.Image.BILINEAR , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase__ : Optional[Any] , ) -> np.ndarray: __SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase__ , default_to_square=UpperCAmelCase__ ) if "shortest_edge" not in size: raise ValueError(F"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) __SCREAMING_SNAKE_CASE = get_resize_output_image_size(UpperCAmelCase__ , size=size["shortest_edge"] , default_to_square=UpperCAmelCase__ ) return resize(UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ , data_format=UpperCAmelCase__ , **UpperCAmelCase__ ) def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Dict[str, int] , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase__ : Any , ) -> np.ndarray: __SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(UpperCAmelCase__ , size=(size["height"], size["width"]) , data_format=UpperCAmelCase__ , **UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Union[int, float] , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase__ : Dict , ) -> Dict: return rescale(UpperCAmelCase__ , scale=UpperCAmelCase__ , data_format=UpperCAmelCase__ , **UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None ) -> np.ndarray: return flip_channel_order(UpperCAmelCase__ , data_format=UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] , UpperCAmelCase__ : ImageInput , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : float = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : ChannelDimension = ChannelDimension.FIRST , **UpperCAmelCase__ : str , ) -> PIL.Image.Image: __SCREAMING_SNAKE_CASE = do_resize if do_resize is not None else self.do_resize __SCREAMING_SNAKE_CASE = resample if resample is not None else self.resample __SCREAMING_SNAKE_CASE = do_rescale if do_rescale is not None else self.do_rescale __SCREAMING_SNAKE_CASE = rescale_factor if rescale_factor is not None else self.rescale_factor __SCREAMING_SNAKE_CASE = do_center_crop if do_center_crop is not None else self.do_center_crop __SCREAMING_SNAKE_CASE = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) __SCREAMING_SNAKE_CASE = size if size is not None else self.size __SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase__ , default_to_square=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = crop_size if crop_size is not None else self.crop_size __SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase__ , param_name="crop_size" ) __SCREAMING_SNAKE_CASE = make_list_of_images(UpperCAmelCase__ ) if not valid_images(UpperCAmelCase__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) # All transformations expect numpy arrays. __SCREAMING_SNAKE_CASE = [to_numpy_array(UpperCAmelCase__ ) for image in images] if do_resize: __SCREAMING_SNAKE_CASE = [self.resize(image=UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ ) for image in images] if do_center_crop: __SCREAMING_SNAKE_CASE = [self.center_crop(image=UpperCAmelCase__ , size=UpperCAmelCase__ ) for image in images] if do_rescale: __SCREAMING_SNAKE_CASE = [self.rescale(image=UpperCAmelCase__ , scale=UpperCAmelCase__ ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: __SCREAMING_SNAKE_CASE = [self.flip_channel_order(image=UpperCAmelCase__ ) for image in images] __SCREAMING_SNAKE_CASE = [to_channel_dimension_format(UpperCAmelCase__ , UpperCAmelCase__ ) for image in images] __SCREAMING_SNAKE_CASE = {"pixel_values": images} return BatchFeature(data=UpperCAmelCase__ , tensor_type=UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[Tuple] = None ) -> Optional[int]: __SCREAMING_SNAKE_CASE = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(UpperCAmelCase__ ) != len(UpperCAmelCase__ ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(UpperCAmelCase__ ): __SCREAMING_SNAKE_CASE = target_sizes.numpy() __SCREAMING_SNAKE_CASE = [] for idx in range(len(UpperCAmelCase__ ) ): __SCREAMING_SNAKE_CASE = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(UpperCAmelCase__ ) else: __SCREAMING_SNAKE_CASE = logits.argmax(dim=1 ) __SCREAMING_SNAKE_CASE = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
195
1
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import Callable, Dict, List, Tuple import timm import torch import torch.nn as nn from classy_vision.models.regnet import RegNet, RegNetParams, RegNetYaagf, RegNetYaagf, RegNetYaaagf from huggingface_hub import cached_download, hf_hub_url from torch import Tensor from vissl.models.model_helpers import get_trunk_forward_outputs from transformers import AutoImageProcessor, RegNetConfig, RegNetForImageClassification, RegNetModel from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger() @dataclass class lowerCamelCase : '''simple docstring''' _snake_case : nn.Module _snake_case : List[nn.Module] = field(default_factory=_lowercase ) _snake_case : list = field(default_factory=_lowercase ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Any: UpperCAmelCase_ : str = len(list(m.modules() ) ) == 1 or isinstance(_SCREAMING_SNAKE_CASE , nn.Convad ) or isinstance(_SCREAMING_SNAKE_CASE , nn.BatchNormad ) if has_not_submodules: self.traced.append(_SCREAMING_SNAKE_CASE ) def __call__( self , _UpperCamelCase ) -> str: for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(_SCREAMING_SNAKE_CASE ) [x.remove() for x in self.handles] return self @property def __UpperCAmelCase ( self ) -> Union[str, Any]: # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda _UpperCamelCase : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class lowerCamelCase : '''simple docstring''' _snake_case : nn.Module _snake_case : nn.Module _snake_case : int = 1 _snake_case : List = field(default_factory=_lowercase ) _snake_case : List = field(default_factory=_lowercase ) _snake_case : bool = True def __call__( self , _UpperCamelCase ) -> str: UpperCAmelCase_ : List[Any] = Tracker(self.dest )(_SCREAMING_SNAKE_CASE ).parametrized UpperCAmelCase_ : str = Tracker(self.src )(_SCREAMING_SNAKE_CASE ).parametrized UpperCAmelCase_ : List[str] = list(filter(lambda _UpperCamelCase : type(_SCREAMING_SNAKE_CASE ) not in self.src_skip , _SCREAMING_SNAKE_CASE ) ) UpperCAmelCase_ : str = list(filter(lambda _UpperCamelCase : type(_SCREAMING_SNAKE_CASE ) not in self.dest_skip , _SCREAMING_SNAKE_CASE ) ) if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ) and self.raise_if_mismatch: raise Exception( f"Numbers of operations are different. Source module has {len(_SCREAMING_SNAKE_CASE )} operations while" f" destination module has {len(_SCREAMING_SNAKE_CASE )}." ) for dest_m, src_m in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(f"Transfered from={src_m} to={dest_m}" ) class lowerCamelCase (nn.Module ): '''simple docstring''' def __init__( self , _UpperCamelCase ) -> Optional[int]: super().__init__() UpperCAmelCase_ : List[Tuple[str, nn.Module]] = [] # - get the stem feature_blocks.append(('conv1', model.stem) ) # - get all the feature blocks for k, v in model.trunk_output.named_children(): assert k.startswith('block' ), f"Unexpected layer name {k}" UpperCAmelCase_ : Optional[int] = len(_SCREAMING_SNAKE_CASE ) + 1 feature_blocks.append((f"res{block_index}", v) ) UpperCAmelCase_ : List[str] = nn.ModuleDict(_SCREAMING_SNAKE_CASE ) def __UpperCAmelCase ( self , _UpperCamelCase ) -> List[str]: return get_trunk_forward_outputs( _SCREAMING_SNAKE_CASE , out_feat_keys=_SCREAMING_SNAKE_CASE , feature_blocks=self._feature_blocks , ) class lowerCamelCase (_lowercase ): '''simple docstring''' def __UpperCAmelCase ( self , _UpperCamelCase ) -> str: UpperCAmelCase_ : int = x.split('-' ) return x_split[0] + x_split[1] + "_" + "".join(x_split[2:] ) def __getitem__( self , _UpperCamelCase ) -> Callable[[], Tuple[nn.Module, Dict]]: # default to timm! if x not in self: UpperCAmelCase_ : Optional[Any] = self.convert_name_to_timm(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = partial(lambda: (timm.create_model(_SCREAMING_SNAKE_CASE , pretrained=_SCREAMING_SNAKE_CASE ).eval(), None) ) else: UpperCAmelCase_ : Any = super().__getitem__(_SCREAMING_SNAKE_CASE ) return val class lowerCamelCase (_lowercase ): '''simple docstring''' def __getitem__( self , _UpperCamelCase ) -> Callable[[], nn.Module]: if "seer" in x and "in1k" not in x: UpperCAmelCase_ : int = RegNetModel else: UpperCAmelCase_ : List[str] = RegNetForImageClassification return val def lowercase__ ( __snake_case : Dict , __snake_case : Union[str, Any] , __snake_case : Any ): '''simple docstring''' for from_key, to_key in keys: UpperCAmelCase_ : Optional[Any] = from_state_dict[from_key].clone() print(F"Copied key={from_key} to={to_key}" ) return to_state_dict def lowercase__ ( __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : int , __snake_case : Optional[int] , __snake_case : Optional[int] , __snake_case : Optional[Any] = True , ): '''simple docstring''' print(F"Converting {name}..." ) with torch.no_grad(): UpperCAmelCase_ : int = from_model_func() UpperCAmelCase_ : Optional[Any] = our_model_func(_a ).eval() UpperCAmelCase_ : int = ModuleTransfer(src=_a , dest=_a , raise_if_mismatch=_a ) UpperCAmelCase_ : str = torch.randn((1, 3, 224, 224) ) module_transfer(_a ) if from_state_dict is not None: UpperCAmelCase_ : Any = [] # for seer - in1k finetuned we have to manually copy the head if "seer" in name and "in1k" in name: UpperCAmelCase_ : List[Any] = [('''0.clf.0.weight''', '''classifier.1.weight'''), ('''0.clf.0.bias''', '''classifier.1.bias''')] UpperCAmelCase_ : int = manually_copy_vissl_head(_a , our_model.state_dict() , _a ) our_model.load_state_dict(_a ) UpperCAmelCase_ : List[str] = our_model(_a , output_hidden_states=_a ) UpperCAmelCase_ : Dict = ( our_outputs.logits if isinstance(_a , _a ) else our_outputs.last_hidden_state ) UpperCAmelCase_ : Tuple = from_model(_a ) UpperCAmelCase_ : int = from_output[-1] if type(_a ) is list else from_output # now since I don't want to use any config files, vissl seer model doesn't actually have an head, so let's just check the last hidden state if "seer" in name and "in1k" in name: UpperCAmelCase_ : Optional[int] = our_outputs.hidden_states[-1] assert torch.allclose(_a , _a ), "The model logits don't match the original one." if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / name , commit_message='Add model' , use_temp_dir=_a , ) UpperCAmelCase_ : Optional[int] = 224 if '''seer''' not in name else 384 # we can use the convnext one UpperCAmelCase_ : int = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' , size=_a ) image_processor.push_to_hub( repo_path_or_name=save_directory / name , commit_message='Add image processor' , use_temp_dir=_a , ) print(F"Pushed {name}" ) def lowercase__ ( __snake_case : Optional[int] , __snake_case : str = None , __snake_case : Tuple = True ): '''simple docstring''' UpperCAmelCase_ : str = '''imagenet-1k-id2label.json''' UpperCAmelCase_ : Dict = 1_000 UpperCAmelCase_ : Optional[int] = (1, num_labels) UpperCAmelCase_ : Optional[int] = '''huggingface/label-files''' UpperCAmelCase_ : Tuple = num_labels UpperCAmelCase_ : List[Any] = json.load(open(cached_download(hf_hub_url(_a , _a , repo_type='dataset' ) ) , 'r' ) ) UpperCAmelCase_ : Dict = {int(_a ): v for k, v in idalabel.items()} UpperCAmelCase_ : List[Any] = idalabel UpperCAmelCase_ : Union[str, Any] = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : Dict = partial(_a , num_labels=_a , idalabel=_a , labelaid=_a ) UpperCAmelCase_ : Tuple = { '''regnet-x-002''': ImageNetPreTrainedConfig( depths=[1, 1, 4, 7] , hidden_sizes=[24, 56, 152, 368] , groups_width=8 , layer_type='x' ), '''regnet-x-004''': ImageNetPreTrainedConfig( depths=[1, 2, 7, 12] , hidden_sizes=[32, 64, 160, 384] , groups_width=16 , layer_type='x' ), '''regnet-x-006''': ImageNetPreTrainedConfig( depths=[1, 3, 5, 7] , hidden_sizes=[48, 96, 240, 528] , groups_width=24 , layer_type='x' ), '''regnet-x-008''': ImageNetPreTrainedConfig( depths=[1, 3, 7, 5] , hidden_sizes=[64, 128, 288, 672] , groups_width=16 , layer_type='x' ), '''regnet-x-016''': ImageNetPreTrainedConfig( depths=[2, 4, 10, 2] , hidden_sizes=[72, 168, 408, 912] , groups_width=24 , layer_type='x' ), '''regnet-x-032''': ImageNetPreTrainedConfig( depths=[2, 6, 15, 2] , hidden_sizes=[96, 192, 432, 1_008] , groups_width=48 , layer_type='x' ), '''regnet-x-040''': ImageNetPreTrainedConfig( depths=[2, 5, 14, 2] , hidden_sizes=[80, 240, 560, 1_360] , groups_width=40 , layer_type='x' ), '''regnet-x-064''': ImageNetPreTrainedConfig( depths=[2, 4, 10, 1] , hidden_sizes=[168, 392, 784, 1_624] , groups_width=56 , layer_type='x' ), '''regnet-x-080''': ImageNetPreTrainedConfig( depths=[2, 5, 15, 1] , hidden_sizes=[80, 240, 720, 1_920] , groups_width=120 , layer_type='x' ), '''regnet-x-120''': ImageNetPreTrainedConfig( depths=[2, 5, 11, 1] , hidden_sizes=[224, 448, 896, 2_240] , groups_width=112 , layer_type='x' ), '''regnet-x-160''': ImageNetPreTrainedConfig( depths=[2, 6, 13, 1] , hidden_sizes=[256, 512, 896, 2_048] , groups_width=128 , layer_type='x' ), '''regnet-x-320''': ImageNetPreTrainedConfig( depths=[2, 7, 13, 1] , hidden_sizes=[336, 672, 1_344, 2_520] , groups_width=168 , layer_type='x' ), # y variant '''regnet-y-002''': ImageNetPreTrainedConfig(depths=[1, 1, 4, 7] , hidden_sizes=[24, 56, 152, 368] , groups_width=8 ), '''regnet-y-004''': ImageNetPreTrainedConfig( depths=[1, 3, 6, 6] , hidden_sizes=[48, 104, 208, 440] , groups_width=8 ), '''regnet-y-006''': ImageNetPreTrainedConfig( depths=[1, 3, 7, 4] , hidden_sizes=[48, 112, 256, 608] , groups_width=16 ), '''regnet-y-008''': ImageNetPreTrainedConfig( depths=[1, 3, 8, 2] , hidden_sizes=[64, 128, 320, 768] , groups_width=16 ), '''regnet-y-016''': ImageNetPreTrainedConfig( depths=[2, 6, 17, 2] , hidden_sizes=[48, 120, 336, 888] , groups_width=24 ), '''regnet-y-032''': ImageNetPreTrainedConfig( depths=[2, 5, 13, 1] , hidden_sizes=[72, 216, 576, 1_512] , groups_width=24 ), '''regnet-y-040''': ImageNetPreTrainedConfig( depths=[2, 6, 12, 2] , hidden_sizes=[128, 192, 512, 1_088] , groups_width=64 ), '''regnet-y-064''': ImageNetPreTrainedConfig( depths=[2, 7, 14, 2] , hidden_sizes=[144, 288, 576, 1_296] , groups_width=72 ), '''regnet-y-080''': ImageNetPreTrainedConfig( depths=[2, 4, 10, 1] , hidden_sizes=[168, 448, 896, 2_016] , groups_width=56 ), '''regnet-y-120''': ImageNetPreTrainedConfig( depths=[2, 5, 11, 1] , hidden_sizes=[224, 448, 896, 2_240] , groups_width=112 ), '''regnet-y-160''': ImageNetPreTrainedConfig( depths=[2, 4, 11, 1] , hidden_sizes=[224, 448, 1_232, 3_024] , groups_width=112 ), '''regnet-y-320''': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1_392, 3_712] , groups_width=232 ), # models created by SEER -> https://arxiv.org/abs/2202.08360 '''regnet-y-320-seer''': RegNetConfig(depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1_392, 3_712] , groups_width=232 ), '''regnet-y-640-seer''': RegNetConfig(depths=[2, 5, 12, 1] , hidden_sizes=[328, 984, 1_968, 4_920] , groups_width=328 ), '''regnet-y-1280-seer''': RegNetConfig( depths=[2, 7, 17, 1] , hidden_sizes=[528, 1_056, 2_904, 7_392] , groups_width=264 ), '''regnet-y-2560-seer''': RegNetConfig( depths=[3, 7, 16, 1] , hidden_sizes=[640, 1_696, 2_544, 5_088] , groups_width=640 ), '''regnet-y-10b-seer''': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[2_020, 4_040, 11_110, 28_280] , groups_width=1_010 ), # finetuned on imagenet '''regnet-y-320-seer-in1k''': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1_392, 3_712] , groups_width=232 ), '''regnet-y-640-seer-in1k''': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[328, 984, 1_968, 4_920] , groups_width=328 ), '''regnet-y-1280-seer-in1k''': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[528, 1_056, 2_904, 7_392] , groups_width=264 ), '''regnet-y-2560-seer-in1k''': ImageNetPreTrainedConfig( depths=[3, 7, 16, 1] , hidden_sizes=[640, 1_696, 2_544, 5_088] , groups_width=640 ), '''regnet-y-10b-seer-in1k''': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[2_020, 4_040, 11_110, 28_280] , groups_width=1_010 ), } UpperCAmelCase_ : Optional[Any] = NameToOurModelFuncMap() UpperCAmelCase_ : Optional[Any] = NameToFromModelFuncMap() # add seer weights logic def load_using_classy_vision(__snake_case : Any , __snake_case : Dict ) -> Tuple[nn.Module, Dict]: UpperCAmelCase_ : Tuple = torch.hub.load_state_dict_from_url(_a , model_dir=str(_a ) , map_location='cpu' ) UpperCAmelCase_ : int = model_func() # check if we have a head, if yes add it UpperCAmelCase_ : int = files['''classy_state_dict''']['''base_model''']['''model'''] UpperCAmelCase_ : Tuple = model_state_dict['''trunk'''] model.load_state_dict(_a ) return model.eval(), model_state_dict["heads"] # pretrained UpperCAmelCase_ : int = partial( _a , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet32d/seer_regnet32gf_model_iteration244000.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) UpperCAmelCase_ : Optional[int] = partial( _a , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet64/seer_regnet64gf_model_final_checkpoint_phase0.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) UpperCAmelCase_ : Optional[int] = partial( _a , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/swav_ig1b_regnet128Gf_cnstant_bs32_node16_sinkhorn10_proto16k_syncBN64_warmup8k/model_final_checkpoint_phase0.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ) , ) UpperCAmelCase_ : Tuple = partial( _a , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet10B/model_iteration124500_conso.torch' , lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27 , group_width=1_010 , w_a=1_744 , w_a=620.83 , w_m=2.52 ) ) ) , ) # IN1K finetuned UpperCAmelCase_ : List[Any] = partial( _a , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet32_finetuned_in1k_model_final_checkpoint_phase78.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) UpperCAmelCase_ : Optional[int] = partial( _a , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet64_finetuned_in1k_model_final_checkpoint_phase78.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) UpperCAmelCase_ : Union[str, Any] = partial( _a , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet128_finetuned_in1k_model_final_checkpoint_phase78.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ) , ) UpperCAmelCase_ : Union[str, Any] = partial( _a , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_10b_finetuned_in1k_model_phase28_conso.torch' , lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27 , group_width=1_010 , w_a=1_744 , w_a=620.83 , w_m=2.52 ) ) ) , ) if model_name: convert_weight_and_push( _a , names_to_from_model_map[model_name] , names_to_ours_model_map[model_name] , names_to_config[model_name] , _a , _a , ) else: for model_name, config in names_to_config.items(): convert_weight_and_push( _a , names_to_from_model_map[model_name] , names_to_ours_model_map[model_name] , _a , _a , _a , ) return config, expected_shape if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help=( 'The name of the model you wish to convert, it must be one of the supported regnet* architecture,' ' currently: regnetx-*, regnety-*. If `None`, all of them will the converted.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=Path, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=True, type=bool, required=False, help='If True, push model and image processor to the hub.', ) __UpperCAmelCase = parser.parse_args() __UpperCAmelCase = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
29
import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class _a : _a : Dict = None def UpperCAmelCase__( self : Union[str, Any] )-> int: lowerCAmelCase__ : str = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase__ : List[str] = json.loads(feat_extract.to_json_string() ) for key, value in self.feat_extract_dict.items(): self.assertEqual(obj[key] , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Optional[Any] )-> Dict: lowerCAmelCase__ : int = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ : Optional[Any] = os.path.join(_SCREAMING_SNAKE_CASE , '''feat_extract.json''' ) feat_extract_first.to_json_file(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = self.feature_extraction_class.from_json_file(_SCREAMING_SNAKE_CASE ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def UpperCAmelCase__( self : Optional[int] )-> Union[str, Any]: lowerCAmelCase__ : List[Any] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ : Union[str, Any] = feat_extract_first.save_pretrained(_SCREAMING_SNAKE_CASE )[0] check_json_file_has_correct_format(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Union[str, Any] = self.feature_extraction_class.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def UpperCAmelCase__( self : int )-> str: lowerCAmelCase__ : Optional[Any] = self.feature_extraction_class() self.assertIsNotNone(_SCREAMING_SNAKE_CASE )
131
0
'''simple docstring''' import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class UpperCAmelCase__ ( _UpperCamelCase , unittest.TestCase): __SCREAMING_SNAKE_CASE = BioGptTokenizer __SCREAMING_SNAKE_CASE = False def __lowerCamelCase ( self ) -> Any: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __UpperCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] __UpperCamelCase = dict(zip(_UpperCAmelCase , range(len(_UpperCAmelCase ) ) ) ) __UpperCamelCase = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] __UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(_UpperCAmelCase ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(_UpperCAmelCase ) ) def __lowerCamelCase ( self , lowercase ) -> Optional[int]: __UpperCamelCase = 'lower newer' __UpperCamelCase = 'lower newer' return input_text, output_text def __lowerCamelCase ( self ) -> Dict: __UpperCamelCase = BioGptTokenizer(self.vocab_file , self.merges_file ) __UpperCamelCase = 'lower' __UpperCamelCase = ['low', 'er</w>'] __UpperCamelCase = tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) __UpperCamelCase = tokens + ['<unk>'] __UpperCamelCase = [1_4, 1_5, 2_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , _UpperCAmelCase ) @slow def __lowerCamelCase ( self ) -> List[str]: __UpperCamelCase = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) __UpperCamelCase = tokenizer.encode("""sequence builders""" , add_special_tokens=_UpperCAmelCase ) __UpperCamelCase = tokenizer.encode("""multi-sequence build""" , add_special_tokens=_UpperCAmelCase ) __UpperCamelCase = tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase ) __UpperCamelCase = tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase , _UpperCAmelCase ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
369
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = '''Salesforce/blip-image-captioning-base''' __SCREAMING_SNAKE_CASE = ( '''This is a tool that generates a description of an image. It takes an input named `image` which should be the ''' '''image to caption, and returns a text that contains the description in English.''' ) __SCREAMING_SNAKE_CASE = '''image_captioner''' __SCREAMING_SNAKE_CASE = AutoModelForVisionaSeq __SCREAMING_SNAKE_CASE = ['''image'''] __SCREAMING_SNAKE_CASE = ['''text'''] def __init__( self , *lowercase , **lowercase ) -> Optional[int]: requires_backends(self , ["""vision"""] ) super().__init__(*lowercase , **lowercase ) def __lowerCamelCase ( self , lowercase ) -> Any: return self.pre_processor(images=lowercase , return_tensors="""pt""" ) def __lowerCamelCase ( self , lowercase ) -> Optional[int]: return self.model.generate(**lowercase ) def __lowerCamelCase ( self , lowercase ) -> List[str]: return self.pre_processor.batch_decode(lowercase , skip_special_tokens=lowercase )[0].strip()
243
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase = { """configuration_bridgetower""": [ """BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BridgeTowerConfig""", """BridgeTowerTextConfig""", """BridgeTowerVisionConfig""", ], """processing_bridgetower""": ["""BridgeTowerProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ["""BridgeTowerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ """BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST""", """BridgeTowerForContrastiveLearning""", """BridgeTowerForImageAndTextRetrieval""", """BridgeTowerForMaskedLM""", """BridgeTowerModel""", """BridgeTowerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
166
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCamelCase = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ lowerCamelCase = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ lowerCamelCase = R"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCamelCase ( datasets.Metric ): '''simple docstring''' def __lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('float'), 'references': datasets.Value('float'), }) , reference_urls=['https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'] , ) def __lowerCamelCase ( self : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : int=False): '''simple docstring''' __lowercase =spearmanr(_lowerCAmelCase , _lowerCAmelCase) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
166
1
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class __lowerCAmelCase : """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=100 , lowerCamelCase__=13 , lowerCamelCase__=30 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=32 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=37 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=10 , lowerCamelCase__=0.02 , lowerCamelCase__=3 , lowerCamelCase__=None , lowerCamelCase__=[0, 1, 2, 3] , ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = 100 __lowerCamelCase = batch_size __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = scope __lowerCamelCase = out_indices __lowerCamelCase = num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCamelCase = (image_size // patch_size) ** 2 __lowerCamelCase = num_patches + 1 def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __lowerCamelCase = self.get_config() return config, pixel_values, labels, pixel_labels def lowercase_ ( self ) -> Dict: '''simple docstring''' return BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , out_indices=self.out_indices , ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> List[Any]: '''simple docstring''' __lowerCamelCase = BeitModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = BeitForMaskedImageModeling(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Dict: '''simple docstring''' __lowerCamelCase = self.type_sequence_label_size __lowerCamelCase = BeitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCamelCase = 1 __lowerCamelCase = BeitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> str: '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = BeitForSemanticSegmentation(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) __lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = config_and_inputs __lowerCamelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( __magic_name__ , __magic_name__ , unittest.TestCase ): """simple docstring""" snake_case_ = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) snake_case_ = ( { '''feature-extraction''': BeitModel, '''image-classification''': BeitForImageClassification, '''image-segmentation''': BeitForSemanticSegmentation, } if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False def lowercase_ ( self ) -> Any: '''simple docstring''' __lowerCamelCase = BeitModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=37 ) def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='BEiT does not use inputs_embeds' ) def lowercase_ ( self ) -> str: '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='BEiT has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def lowercase_ ( self ) -> Dict: '''simple docstring''' pass def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(lowerCamelCase__ ) __lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase__ ) def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase__ ) def lowercase_ ( self ) -> Dict: '''simple docstring''' if not self.model_tester.is_training: return __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(lowerCamelCase__ ), BeitForMaskedImageModeling]: continue __lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.train() __lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ , return_labels=lowerCamelCase__ ) __lowerCamelCase = model(**lowerCamelCase__ ).loss loss.backward() def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __lowerCamelCase = False __lowerCamelCase = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(lowerCamelCase__ ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue __lowerCamelCase = model_class(lowerCamelCase__ ) model.gradient_checkpointing_enable() model.to(lowerCamelCase__ ) model.train() __lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ , return_labels=lowerCamelCase__ ) __lowerCamelCase = model(**lowerCamelCase__ ).loss loss.backward() def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = _config_zero_init(lowerCamelCase__ ) for model_class in self.all_model_classes: __lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @slow def lowercase_ ( self ) -> List[Any]: '''simple docstring''' for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = BeitModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCamelCase_ ( ) -> Optional[int]: """simple docstring""" __lowerCamelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def lowercase_ ( self ) -> Dict: '''simple docstring''' return BeitImageProcessor.from_pretrained('microsoft/beit-base-patch16-224' ) if is_vision_available() else None @slow def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = BeitForMaskedImageModeling.from_pretrained('microsoft/beit-base-patch16-224-pt22k' ).to(lowerCamelCase__ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='pt' ).pixel_values.to(lowerCamelCase__ ) # prepare bool_masked_pos __lowerCamelCase = torch.ones((1, 196) , dtype=torch.bool ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(pixel_values=lowerCamelCase__ , bool_masked_pos=lowerCamelCase__ ) __lowerCamelCase = outputs.logits # verify the logits __lowerCamelCase = torch.Size((1, 196, 8_192) ) self.assertEqual(logits.shape , lowerCamelCase__ ) __lowerCamelCase = torch.tensor( [[-3.24_37, 0.50_72, -13.91_74], [-3.24_56, 0.49_48, -13.94_01], [-3.20_33, 0.51_21, -13.85_50]] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] , lowerCamelCase__ , atol=1e-2 ) ) @slow def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = BeitForImageClassification.from_pretrained('microsoft/beit-base-patch16-224' ).to(lowerCamelCase__ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='pt' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**lowerCamelCase__ ) __lowerCamelCase = outputs.logits # verify the logits __lowerCamelCase = torch.Size((1, 1_000) ) self.assertEqual(logits.shape , lowerCamelCase__ ) __lowerCamelCase = torch.tensor([-1.23_85, -1.09_87, -1.01_08] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) __lowerCamelCase = 281 self.assertEqual(logits.argmax(-1 ).item() , lowerCamelCase__ ) @slow def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = BeitForImageClassification.from_pretrained('microsoft/beit-large-patch16-224-pt22k-ft22k' ).to( lowerCamelCase__ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='pt' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**lowerCamelCase__ ) __lowerCamelCase = outputs.logits # verify the logits __lowerCamelCase = torch.Size((1, 21_841) ) self.assertEqual(logits.shape , lowerCamelCase__ ) __lowerCamelCase = torch.tensor([1.68_81, -0.27_87, 0.59_01] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) __lowerCamelCase = 2_396 self.assertEqual(logits.argmax(-1 ).item() , lowerCamelCase__ ) @slow def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' __lowerCamelCase = BeitForSemanticSegmentation.from_pretrained('microsoft/beit-base-finetuned-ade-640-640' ) __lowerCamelCase = model.to(lowerCamelCase__ ) __lowerCamelCase = BeitImageProcessor(do_resize=lowerCamelCase__ , size=640 , do_center_crop=lowerCamelCase__ ) __lowerCamelCase = load_dataset('hf-internal-testing/fixtures_ade20k' , split='test' ) __lowerCamelCase = Image.open(ds[0]['file'] ) __lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='pt' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**lowerCamelCase__ ) __lowerCamelCase = outputs.logits # verify the logits __lowerCamelCase = torch.Size((1, 150, 160, 160) ) self.assertEqual(logits.shape , lowerCamelCase__ ) __lowerCamelCase = version.parse(PIL.__version__ ) < version.parse('9.0.0' ) if is_pillow_less_than_a: __lowerCamelCase = torch.tensor( [ [[-4.92_25, -2.39_54, -3.05_22], [-2.88_22, -1.00_46, -1.75_61], [-2.95_49, -1.32_28, -2.13_47]], [[-5.81_68, -3.41_29, -4.07_78], [-3.86_51, -2.22_14, -3.02_77], [-3.83_56, -2.46_43, -3.35_35]], [[-0.00_78, 3.99_52, 4.07_54], [2.98_56, 4.69_44, 5.00_35], [3.24_13, 4.78_13, 4.99_69]], ] , device=lowerCamelCase__ , ) else: __lowerCamelCase = torch.tensor( [ [[-4.89_60, -2.36_88, -3.03_55], [-2.84_78, -0.98_36, -1.74_18], [-2.94_49, -1.33_32, -2.14_56]], [[-5.80_81, -3.41_24, -4.10_06], [-3.85_61, -2.20_81, -3.03_23], [-3.83_65, -2.46_01, -3.36_69]], [[-0.03_09, 3.98_68, 4.05_40], [2.96_40, 4.68_77, 4.99_76], [3.20_81, 4.76_90, 4.99_42]], ] , device=lowerCamelCase__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase__ , atol=1e-4 ) ) @slow def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = BeitForSemanticSegmentation.from_pretrained('microsoft/beit-base-finetuned-ade-640-640' ) __lowerCamelCase = model.to(lowerCamelCase__ ) __lowerCamelCase = BeitImageProcessor(do_resize=lowerCamelCase__ , size=640 , do_center_crop=lowerCamelCase__ ) __lowerCamelCase = load_dataset('hf-internal-testing/fixtures_ade20k' , split='test' ) __lowerCamelCase = Image.open(ds[0]['file'] ) __lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='pt' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**lowerCamelCase__ ) __lowerCamelCase = outputs.logits.detach().cpu() __lowerCamelCase = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase__ , target_sizes=[(500, 300)] ) __lowerCamelCase = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , lowerCamelCase__ ) __lowerCamelCase = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase__ ) __lowerCamelCase = torch.Size((160, 160) ) self.assertEqual(segmentation[0].shape , lowerCamelCase__ )
348
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal __A = logging.get_logger(__name__) __A = TypeVar("DatasetType", Dataset, IterableDataset) def lowerCamelCase_ ( UpperCamelCase__ : List[DatasetType] , UpperCamelCase__ : Optional[List[float]] = None , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : Optional[DatasetInfo] = None , UpperCamelCase__ : Optional[NamedSplit] = None , UpperCamelCase__ : Literal["first_exhausted", "all_exhausted"] = "first_exhausted" , ) -> DatasetType: """simple docstring""" from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError('Unable to interleave an empty list of datasets.' ) for i, dataset in enumerate(UpperCamelCase__ ): if not isinstance(UpperCamelCase__ , (Dataset, IterableDataset) ): if isinstance(UpperCamelCase__ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} """ 'is an empty dataset dictionary.' ) raise ValueError( F"""Dataset at position {i} has at least one split: {list(UpperCamelCase__ )}\n""" F"""Please pick one to interleave with the other datasets, for example: dataset['{next(iter(UpperCamelCase__ ) )}']""" ) raise ValueError( F"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(UpperCamelCase__ ).__name__}.""" ) if i == 0: __lowerCamelCase , __lowerCamelCase = ( (Dataset, IterableDataset) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else (IterableDataset, Dataset) ) elif not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError( F"""Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.""" ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(F"""{stopping_strategy} is not supported. Please enter a valid stopping_strategy.""" ) if dataset_type is Dataset: return _interleave_map_style_datasets( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , info=UpperCamelCase__ , split=UpperCamelCase__ , stopping_strategy=UpperCamelCase__ ) else: return _interleave_iterable_datasets( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , info=UpperCamelCase__ , split=UpperCamelCase__ , stopping_strategy=UpperCamelCase__ ) def lowerCamelCase_ ( UpperCamelCase__ : List[DatasetType] , UpperCamelCase__ : Optional[DatasetInfo] = None , UpperCamelCase__ : Optional[NamedSplit] = None , UpperCamelCase__ : int = 0 , ) -> DatasetType: """simple docstring""" if not dsets: raise ValueError('Unable to concatenate an empty list of datasets.' ) for i, dataset in enumerate(UpperCamelCase__ ): if not isinstance(UpperCamelCase__ , (Dataset, IterableDataset) ): if isinstance(UpperCamelCase__ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} """ 'is an empty dataset dictionary.' ) raise ValueError( F"""Dataset at position {i} has at least one split: {list(UpperCamelCase__ )}\n""" F"""Please pick one to interleave with the other datasets, for example: dataset['{next(iter(UpperCamelCase__ ) )}']""" ) raise ValueError( F"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(UpperCamelCase__ ).__name__}.""" ) if i == 0: __lowerCamelCase , __lowerCamelCase = ( (Dataset, IterableDataset) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else (IterableDataset, Dataset) ) elif not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError( F"""Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.""" ) if dataset_type is Dataset: return _concatenate_map_style_datasets(UpperCamelCase__ , info=UpperCamelCase__ , split=UpperCamelCase__ , axis=UpperCamelCase__ ) else: return _concatenate_iterable_datasets(UpperCamelCase__ , info=UpperCamelCase__ , split=UpperCamelCase__ , axis=UpperCamelCase__ )
348
1
'''simple docstring''' from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def a ( __a , __a , __a=1e-12 ) -> str: '''simple docstring''' UpperCamelCase__ :Optional[Any] = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(__a , axis=1 ) , a_min=__a ) ).T UpperCamelCase__ :Optional[Any] = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(__a , axis=1 ) , a_min=__a ) ).T return jnp.matmul(__a , norm_emb_a.T ) class lowercase ( nn.Module ): """simple docstring""" _a = 42 _a = jnp.floataa def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[int] = FlaxCLIPVisionModule(self.config.vision_config ) UpperCamelCase__ :Tuple = nn.Dense(self.config.projection_dim , use_bias=UpperCamelCase_ , dtype=self.dtype ) UpperCamelCase__ :Optional[Any] = self.param('''concept_embeds''' , jax.nn.initializers.ones , (17, self.config.projection_dim) ) UpperCamelCase__ :Optional[Any] = self.param( '''special_care_embeds''' , jax.nn.initializers.ones , (3, self.config.projection_dim) ) UpperCamelCase__ :Tuple = self.param('''concept_embeds_weights''' , jax.nn.initializers.ones , (17,) ) UpperCamelCase__ :List[str] = self.param('''special_care_embeds_weights''' , jax.nn.initializers.ones , (3,) ) def __call__( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[str] = self.vision_model(UpperCamelCase_ )[1] UpperCamelCase__ :Optional[int] = self.visual_projection(UpperCamelCase_ ) UpperCamelCase__ :List[Any] = jax_cosine_distance(UpperCamelCase_ , self.special_care_embeds ) UpperCamelCase__ :List[str] = jax_cosine_distance(UpperCamelCase_ , self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs UpperCamelCase__ :int = 0.0 UpperCamelCase__ :int = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment UpperCamelCase__ :Optional[int] = jnp.round(UpperCamelCase_ , 3 ) UpperCamelCase__ :Union[str, Any] = jnp.any(special_scores > 0 , axis=1 , keepdims=UpperCamelCase_ ) # Use a lower threshold if an image has any special care concept UpperCamelCase__ :Union[str, Any] = is_special_care * 0.01 UpperCamelCase__ :List[str] = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment UpperCamelCase__ :Optional[int] = jnp.round(UpperCamelCase_ , 3 ) UpperCamelCase__ :int = jnp.any(concept_scores > 0 , axis=1 ) return has_nsfw_concepts class lowercase ( A__ ): """simple docstring""" _a = CLIPConfig _a = 'clip_input' _a = FlaxStableDiffusionSafetyCheckerModule def __init__( self , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = 0 , UpperCamelCase_ = jnp.floataa , UpperCamelCase_ = True , **UpperCamelCase_ , ): '''simple docstring''' if input_shape is None: UpperCamelCase__ :Union[str, Any] = (1, 224, 224, 3) UpperCamelCase__ :Dict = self.module_class(config=UpperCamelCase_ , dtype=UpperCamelCase_ , **UpperCamelCase_ ) super().__init__(UpperCamelCase_ , UpperCamelCase_ , input_shape=UpperCamelCase_ , seed=UpperCamelCase_ , dtype=UpperCamelCase_ , _do_init=_do_init ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' UpperCamelCase__ :Dict = jax.random.normal(UpperCamelCase_ , UpperCamelCase_ ) UpperCamelCase__ , UpperCamelCase__ :List[Any] = jax.random.split(UpperCamelCase_ ) UpperCamelCase__ :Tuple = {'''params''': params_rng, '''dropout''': dropout_rng} UpperCamelCase__ :List[str] = self.module.init(UpperCamelCase_ , UpperCamelCase_ )['''params'''] return random_params def __call__( self , UpperCamelCase_ , UpperCamelCase_ = None , ): '''simple docstring''' UpperCamelCase__ :Any = jnp.transpose(UpperCamelCase_ , (0, 2, 3, 1) ) return self.module.apply( {'''params''': params or self.params} , jnp.array(UpperCamelCase_ , dtype=jnp.floataa ) , rngs={} , )
97
import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case__ : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=32 , lowerCAmelCase__=2 , lowerCAmelCase__=3 , lowerCAmelCase__=16 , lowerCAmelCase__=[32, 64, 1_28] , lowerCAmelCase__=[1, 2, 1] , lowerCAmelCase__=[2, 2, 4] , lowerCAmelCase__=2 , lowerCAmelCase__=2.0 , lowerCAmelCase__=True , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.1 , lowerCAmelCase__="gelu" , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=0.0_2 , lowerCAmelCase__=1e-5 , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=10 , lowerCAmelCase__=8 , lowerCAmelCase__=["stage1", "stage2"] , lowerCAmelCase__=[1, 2] , ) -> str: __magic_name__ : Optional[int] = parent __magic_name__ : Any = batch_size __magic_name__ : Union[str, Any] = image_size __magic_name__ : Optional[int] = patch_size __magic_name__ : Union[str, Any] = num_channels __magic_name__ : str = embed_dim __magic_name__ : int = hidden_sizes __magic_name__ : Union[str, Any] = depths __magic_name__ : List[str] = num_heads __magic_name__ : str = window_size __magic_name__ : Optional[Any] = mlp_ratio __magic_name__ : Dict = qkv_bias __magic_name__ : Dict = hidden_dropout_prob __magic_name__ : Optional[Any] = attention_probs_dropout_prob __magic_name__ : List[Any] = drop_path_rate __magic_name__ : Optional[Any] = hidden_act __magic_name__ : int = use_absolute_embeddings __magic_name__ : Dict = patch_norm __magic_name__ : Tuple = layer_norm_eps __magic_name__ : List[str] = initializer_range __magic_name__ : Optional[int] = is_training __magic_name__ : Optional[Any] = scope __magic_name__ : Union[str, Any] = use_labels __magic_name__ : Optional[Any] = type_sequence_label_size __magic_name__ : Union[str, Any] = encoder_stride __magic_name__ : List[Any] = out_features __magic_name__ : Union[str, Any] = out_indices def __magic_name__ ( self ) -> str: __magic_name__ : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __magic_name__ : Optional[Any] = None if self.use_labels: __magic_name__ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __magic_name__ : Dict = self.get_config() return config, pixel_values, labels def __magic_name__ ( self ) -> List[Any]: return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: __magic_name__ : Any = FocalNetModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __magic_name__ : Optional[int] = model(lowerCAmelCase__ ) __magic_name__ : Union[str, Any] = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __magic_name__ : Optional[Any] = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: __magic_name__ : List[str] = FocalNetBackbone(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __magic_name__ : Tuple = model(lowerCAmelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None __magic_name__ : Optional[Any] = None __magic_name__ : List[str] = FocalNetBackbone(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __magic_name__ : Union[str, Any] = model(lowerCAmelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: __magic_name__ : Optional[int] = FocalNetForMaskedImageModeling(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __magic_name__ : str = model(lowerCAmelCase__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __magic_name__ : Optional[int] = 1 __magic_name__ : int = FocalNetForMaskedImageModeling(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __magic_name__ : int = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __magic_name__ : List[Any] = model(lowerCAmelCase__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: __magic_name__ : int = self.type_sequence_label_size __magic_name__ : Tuple = FocalNetForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __magic_name__ : int = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __magic_name__ : Optional[int] = 1 __magic_name__ : Dict = FocalNetForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() __magic_name__ : Union[str, Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __magic_name__ : Dict = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __magic_name__ ( self ) -> int: __magic_name__ : int = self.prepare_config_and_inputs() __magic_name__ ,__magic_name__ ,__magic_name__ : Dict = config_and_inputs __magic_name__ : Optional[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class snake_case__ ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): lowercase__ : str = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) lowercase__ : Any = ( {'''feature-extraction''': FocalNetModel, '''image-classification''': FocalNetForImageClassification} if is_torch_available() else {} ) lowercase__ : Dict = False lowercase__ : Dict = False lowercase__ : int = False lowercase__ : Tuple = False lowercase__ : Optional[Any] = False def __magic_name__ ( self ) -> Dict: __magic_name__ : Optional[Any] = FocalNetModelTester(self ) __magic_name__ : int = ConfigTester(self , config_class=lowerCAmelCase__ , embed_dim=37 , has_text_modality=lowerCAmelCase__ ) def __magic_name__ ( self ) -> List[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __magic_name__ ( self ) -> List[str]: return def __magic_name__ ( self ) -> Tuple: __magic_name__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def __magic_name__ ( self ) -> Tuple: __magic_name__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCAmelCase__ ) def __magic_name__ ( self ) -> List[str]: __magic_name__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCAmelCase__ ) def __magic_name__ ( self ) -> List[Any]: __magic_name__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) @unittest.skip(reason="""FocalNet does not use inputs_embeds""" ) def __magic_name__ ( self ) -> List[str]: pass @unittest.skip(reason="""FocalNet does not use feedforward chunking""" ) def __magic_name__ ( self ) -> List[Any]: pass def __magic_name__ ( self ) -> List[Any]: __magic_name__ ,__magic_name__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: __magic_name__ : Dict = model_class(lowerCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __magic_name__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase__ , nn.Linear ) ) def __magic_name__ ( self ) -> Tuple: __magic_name__ ,__magic_name__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: __magic_name__ : str = model_class(lowerCAmelCase__ ) __magic_name__ : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __magic_name__ : Tuple = [*signature.parameters.keys()] __magic_name__ : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: __magic_name__ : List[Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): __magic_name__ : List[str] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) __magic_name__ : Optional[Any] = outputs.hidden_states __magic_name__ : Union[str, Any] = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) # FocalNet has a different seq_length __magic_name__ : List[str] = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __magic_name__ : Optional[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) __magic_name__ : str = outputs.reshaped_hidden_states self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ : Tuple = reshaped_hidden_states[0].shape __magic_name__ : Union[str, Any] = ( reshaped_hidden_states[0].view(lowerCAmelCase__ , lowerCAmelCase__ , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def __magic_name__ ( self ) -> str: __magic_name__ ,__magic_name__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ : Optional[Any] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: __magic_name__ : List[Any] = True self.check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __magic_name__ : Optional[Any] = True self.check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __magic_name__ ( self ) -> str: __magic_name__ ,__magic_name__ : int = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ : Optional[Any] = 3 __magic_name__ : Union[str, Any] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __magic_name__ : Dict = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __magic_name__ : List[Any] = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __magic_name__ : Tuple = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: __magic_name__ : Optional[int] = True self.check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __magic_name__ : str = True self.check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , (padded_height, padded_width) ) @slow def __magic_name__ ( self ) -> Union[str, Any]: for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __magic_name__ : Optional[int] = FocalNetModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def __magic_name__ ( self ) -> Optional[int]: __magic_name__ ,__magic_name__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ : Dict = _config_zero_init(lowerCAmelCase__ ) for model_class in self.all_model_classes: __magic_name__ : Any = model_class(config=lowerCAmelCase__ ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @require_vision @require_torch class snake_case__ ( unittest.TestCase ): @cached_property def __magic_name__ ( self ) -> Optional[int]: # TODO update organization return AutoImageProcessor.from_pretrained("""microsoft/focalnet-tiny""" ) if is_vision_available() else None @slow def __magic_name__ ( self ) -> Optional[Any]: __magic_name__ : int = FocalNetForImageClassification.from_pretrained("""microsoft/focalnet-tiny""" ).to(lowerCAmelCase__ ) __magic_name__ : Optional[Any] = self.default_image_processor __magic_name__ : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) __magic_name__ : Union[str, Any] = image_processor(images=lowerCAmelCase__ , return_tensors="""pt""" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): __magic_name__ : List[Any] = model(**lowerCAmelCase__ ) # verify the logits __magic_name__ : Union[str, Any] = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) __magic_name__ : Dict = torch.tensor([0.2_1_6_6, -0.4_3_6_8, 0.2_1_9_1] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1e-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 2_81 ) @require_torch class snake_case__ ( _lowerCAmelCase , unittest.TestCase ): lowercase__ : Any = (FocalNetBackbone,) if is_torch_available() else () lowercase__ : Optional[int] = FocalNetConfig lowercase__ : Dict = False def __magic_name__ ( self ) -> int: __magic_name__ : Dict = FocalNetModelTester(self )
342
0
def lowerCamelCase__ (_UpperCAmelCase): if not isinstance(_UpperCAmelCase , _UpperCAmelCase): SCREAMING_SNAKE_CASE = F'''Input value of [number={number}] must be an integer''' raise TypeError(_UpperCAmelCase) if number < 0: return False SCREAMING_SNAKE_CASE = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
354
class _snake_case : def __init__( self , a) -> Optional[Any]: SCREAMING_SNAKE_CASE = val SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None def SCREAMING_SNAKE_CASE__ ( self , a) -> str: if self.val: if val < self.val: if self.left is None: SCREAMING_SNAKE_CASE = Node(a) else: self.left.insert(a) elif val > self.val: if self.right is None: SCREAMING_SNAKE_CASE = Node(a) else: self.right.insert(a) else: SCREAMING_SNAKE_CASE = val def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase): # Recursive traversal if root: inorder(root.left , _UpperCAmelCase) res.append(root.val) inorder(root.right , _UpperCAmelCase) def lowerCamelCase__ (_UpperCAmelCase): # Build BST if len(_UpperCAmelCase) == 0: return arr SCREAMING_SNAKE_CASE = Node(arr[0]) for i in range(1 , len(_UpperCAmelCase)): root.insert(arr[i]) # Traverse BST in order. SCREAMING_SNAKE_CASE = [] inorder(_UpperCAmelCase , _UpperCAmelCase) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
327
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _A : str =logging.get_logger(__name__) _A : int ={ '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json''', # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class _lowercase ( _lowercase ): a = """convbert""" def __init__( self: Tuple , UpperCamelCase__: List[str]=30_522 , UpperCamelCase__: List[Any]=768 , UpperCamelCase__: Union[str, Any]=12 , UpperCamelCase__: Optional[int]=12 , UpperCamelCase__: Optional[int]=3_072 , UpperCamelCase__: Tuple="gelu" , UpperCamelCase__: Any=0.1 , UpperCamelCase__: Tuple=0.1 , UpperCamelCase__: Dict=512 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: Dict=0.02 , UpperCamelCase__: Tuple=1e-12 , UpperCamelCase__: Tuple=1 , UpperCamelCase__: Optional[Any]=0 , UpperCamelCase__: Optional[Any]=2 , UpperCamelCase__: str=768 , UpperCamelCase__: Union[str, Any]=2 , UpperCamelCase__: Any=9 , UpperCamelCase__: Optional[Any]=1 , UpperCamelCase__: Any=None , **UpperCamelCase__: List[Any] , ): super().__init__( pad_token_id=UpperCamelCase__ , bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ , ) lowerCamelCase__ : Any = vocab_size lowerCamelCase__ : List[Any] = hidden_size lowerCamelCase__ : Union[str, Any] = num_hidden_layers lowerCamelCase__ : List[Any] = num_attention_heads lowerCamelCase__ : Tuple = intermediate_size lowerCamelCase__ : Optional[Any] = hidden_act lowerCamelCase__ : Optional[int] = hidden_dropout_prob lowerCamelCase__ : str = attention_probs_dropout_prob lowerCamelCase__ : Optional[Any] = max_position_embeddings lowerCamelCase__ : Optional[Any] = type_vocab_size lowerCamelCase__ : List[Any] = initializer_range lowerCamelCase__ : List[str] = layer_norm_eps lowerCamelCase__ : List[str] = embedding_size lowerCamelCase__ : Optional[Any] = head_ratio lowerCamelCase__ : int = conv_kernel_size lowerCamelCase__ : Optional[int] = num_groups lowerCamelCase__ : Optional[int] = classifier_dropout class _lowercase ( _lowercase ): @property def lowerCamelCase_ ( self: Dict ): if self.task == "multiple-choice": lowerCamelCase__ : Tuple = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase__ : Any = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
41
import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency __A = { "E": 1_2.7_0, "T": 9.0_6, "A": 8.1_7, "O": 7.5_1, "I": 6.9_7, "N": 6.7_5, "S": 6.3_3, "H": 6.0_9, "R": 5.9_9, "D": 4.2_5, "L": 4.0_3, "C": 2.7_8, "U": 2.7_6, "M": 2.4_1, "W": 2.3_6, "F": 2.2_3, "G": 2.0_2, "Y": 1.9_7, "P": 1.9_3, "B": 1.2_9, "V": 0.9_8, "K": 0.7_7, "J": 0.1_5, "X": 0.1_5, "Q": 0.1_0, "Z": 0.0_7, } __A = "ETAOINSHRDLCUMWFGYPBVKJXQZ" __A = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" def lowerCamelCase_ ( UpperCamelCase__ : str ) -> dict[str, int]: """simple docstring""" __lowerCamelCase = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def lowerCamelCase_ ( UpperCamelCase__ : tuple ) -> str: """simple docstring""" return x[0] def lowerCamelCase_ ( UpperCamelCase__ : str ) -> str: """simple docstring""" __lowerCamelCase = get_letter_count(UpperCamelCase__ ) __lowerCamelCase = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(UpperCamelCase__ ) __lowerCamelCase = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=UpperCamelCase__ ) __lowerCamelCase = ''.join(freq_to_letter[freq] ) __lowerCamelCase = list(freq_to_letter_str.items() ) freq_pairs.sort(key=UpperCamelCase__ , reverse=UpperCamelCase__ ) __lowerCamelCase = [freq_pair[1] for freq_pair in freq_pairs] return "".join(UpperCamelCase__ ) def lowerCamelCase_ ( UpperCamelCase__ : str ) -> int: """simple docstring""" __lowerCamelCase = get_frequency_order(UpperCamelCase__ ) __lowerCamelCase = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
90
0
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = {'''vocab_file''': '''vocab.txt'''} lowerCAmelCase = { '''vocab_file''': { '''openbmb/cpm-ant-10b''': '''https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt''', }, } lowerCAmelCase = { '''openbmb/cpm-ant-10b''': 1_0_2_4, } def _lowerCamelCase( lowercase__ ) -> List[Any]: '''simple docstring''' __lowercase= collections.OrderedDict() with open(lowercase__ , 'r' , encoding='utf-8' ) as reader: __lowercase= reader.readlines() for index, token in enumerate(lowercase__ ): __lowercase= token.rstrip('\n' ) __lowercase= index return vocab class A ( A_ ): def __init__(self , lowerCAmelCase , lowerCAmelCase="<unk>" , lowerCAmelCase=2_0_0 ): __lowercase= vocab __lowercase= unk_token __lowercase= max_input_chars_per_word def _A (self , lowerCAmelCase ): __lowercase= list(lowerCAmelCase ) if len(lowerCAmelCase ) > self.max_input_chars_per_word: return [self.unk_token] __lowercase= 0 __lowercase= [] while start < len(lowerCAmelCase ): __lowercase= len(lowerCAmelCase ) __lowercase= None while start < end: __lowercase= ''.join(chars[start:end] ) if substr in self.vocab: __lowercase= substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(lowerCAmelCase ) __lowercase= end return sub_tokens class A ( A_ ): UpperCamelCase_ : Optional[int] =VOCAB_FILES_NAMES UpperCamelCase_ : List[str] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ : List[str] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ : int =['''input_ids''', '''attention_mask'''] UpperCamelCase_ : Optional[int] =False def __init__(self , lowerCAmelCase , lowerCAmelCase="<d>" , lowerCAmelCase="</d>" , lowerCAmelCase="<s>" , lowerCAmelCase="</s>" , lowerCAmelCase="<pad>" , lowerCAmelCase="<unk>" , lowerCAmelCase="</n>" , lowerCAmelCase="</_>" , lowerCAmelCase="left" , **lowerCAmelCase , ): requires_backends(self , ['jieba'] ) super().__init__( bod_token=lowerCAmelCase , eod_token=lowerCAmelCase , bos_token=lowerCAmelCase , eos_token=lowerCAmelCase , pad_token=lowerCAmelCase , unk_token=lowerCAmelCase , line_token=lowerCAmelCase , space_token=lowerCAmelCase , padding_side=lowerCAmelCase , **lowerCAmelCase , ) __lowercase= bod_token __lowercase= eod_token __lowercase= load_vocab(lowerCAmelCase ) __lowercase= self.encoder[space_token] __lowercase= self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] __lowercase= collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCAmelCase : x[1] ) ) __lowercase= {v: k for k, v in self.encoder.items()} __lowercase= WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def _A (self ): return self.encoder[self.bod_token] @property def _A (self ): return self.encoder[self.eod_token] @property def _A (self ): return self.encoder["\n"] @property def _A (self ): return len(self.encoder ) def _A (self ): return dict(self.encoder , **self.added_tokens_encoder ) def _A (self , lowerCAmelCase ): __lowercase= [] for x in jieba.cut(lowerCAmelCase , cut_all=lowerCAmelCase ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(lowerCAmelCase ) ) return output_tokens def _A (self , lowerCAmelCase , **lowerCAmelCase ): __lowercase= [i for i in token_ids if i >= 0] __lowercase= [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(lowerCAmelCase , **lowerCAmelCase ) def _A (self , lowerCAmelCase ): return token in self.encoder def _A (self , lowerCAmelCase ): return "".join(lowerCAmelCase ) def _A (self , lowerCAmelCase ): return self.encoder.get(lowerCAmelCase , self.encoder.get(self.unk_token ) ) def _A (self , lowerCAmelCase ): return self.decoder.get(lowerCAmelCase , self.unk_token ) def _A (self , lowerCAmelCase , lowerCAmelCase = None ): if os.path.isdir(lowerCAmelCase ): __lowercase= os.path.join( lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) else: __lowercase= (filename_prefix + '-' if filename_prefix else '') + save_directory __lowercase= 0 if " " in self.encoder: __lowercase= self.encoder[' '] del self.encoder[" "] if "\n" in self.encoder: __lowercase= self.encoder['\n'] del self.encoder["\n"] __lowercase= collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCAmelCase : x[1] ) ) with open(lowerCAmelCase , 'w' , encoding='utf-8' ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' ' Please check that the vocabulary is not corrupted!' ) __lowercase= token_index writer.write(token + '\n' ) index += 1 return (vocab_file,) def _A (self , lowerCAmelCase , lowerCAmelCase = None ): if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def _A (self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase , token_ids_a=lowerCAmelCase , already_has_special_tokens=lowerCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(lowerCAmelCase )) + [1] + ([0] * len(lowerCAmelCase )) return [1] + ([0] * len(lowerCAmelCase ))
350
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowerCAmelCase = '''<<<<<<< This should probably be modified because it mentions: ''' lowerCAmelCase = '''======= >>>>>>> ''' lowerCAmelCase = [ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] lowerCAmelCase = [ # (pattern, replacement) # Order is important here for some replacements (R'''tfds\.core''', R'''datasets'''), (R'''tf\.io\.gfile\.GFile''', R'''open'''), (R'''tf\.([\w\d]+)''', R'''datasets.Value(\'\1\')'''), (R'''tfds\.features\.Text\(\)''', R'''datasets.Value(\'string\')'''), (R'''tfds\.features\.Text\(''', R'''datasets.Value(\'string\'),'''), (R'''features\s*=\s*tfds.features.FeaturesDict\(''', R'''features=datasets.Features('''), (R'''tfds\.features\.FeaturesDict\(''', R'''dict('''), (R'''The TensorFlow Datasets Authors''', R'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (R'''tfds\.''', R'''datasets.'''), (R'''dl_manager\.manual_dir''', R'''self.config.data_dir'''), (R'''self\.builder_config''', R'''self.config'''), ] def _lowerCamelCase( lowercase__ ) -> Optional[int]: '''simple docstring''' return ConvertCommand(args.tfds_path , args.datasets_directory ) class A ( A_ ): @staticmethod def _A (lowerCAmelCase ): __lowercase= parser.add_parser( 'convert' , help='Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.' , ) train_parser.add_argument( '--tfds_path' , type=lowerCAmelCase , required=lowerCAmelCase , help='Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.' , ) train_parser.add_argument( '--datasets_directory' , type=lowerCAmelCase , required=lowerCAmelCase , help='Path to the HuggingFace Datasets folder.' ) train_parser.set_defaults(func=lowerCAmelCase ) def __init__(self , lowerCAmelCase , lowerCAmelCase , *lowerCAmelCase ): __lowercase= get_logger('datasets-cli/converting' ) __lowercase= tfds_path __lowercase= datasets_directory def _A (self ): if os.path.isdir(self._tfds_path ): __lowercase= os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): __lowercase= os.path.dirname(self._tfds_path ) else: raise ValueError('--tfds_path is neither a directory nor a file. Please check path.' ) __lowercase= os.path.abspath(self._datasets_directory ) self._logger.info(f'Converting datasets from {abs_tfds_path} to {abs_datasets_path}' ) __lowercase= [] __lowercase= [] __lowercase= {} if os.path.isdir(self._tfds_path ): __lowercase= os.listdir(lowerCAmelCase ) else: __lowercase= [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f'Looking at file {f_name}' ) __lowercase= os.path.join(lowerCAmelCase , lowerCAmelCase ) __lowercase= os.path.join(lowerCAmelCase , lowerCAmelCase ) if not os.path.isfile(lowerCAmelCase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('Skipping file' ) continue with open(lowerCAmelCase , encoding='utf-8' ) as f: __lowercase= f.readlines() __lowercase= [] __lowercase= False __lowercase= False __lowercase= [] for line in lines: __lowercase= line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: __lowercase= 'import datasets\n' elif "import tensorflow" in out_line: # order is important here __lowercase= '' continue elif "from absl import logging" in out_line: __lowercase= 'from datasets import logging\n' elif "getLogger" in out_line: __lowercase= out_line.replace('getLogger' , 'get_logger' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): __lowercase= True __lowercase= list(filter(lambda lowerCAmelCase : e in out_line , lowerCAmelCase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(lowerCAmelCase ) + '\n' ) out_lines.append(lowerCAmelCase ) out_lines.append(lowerCAmelCase ) continue else: for pattern, replacement in TO_CONVERT: __lowercase= re.sub(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: __lowercase= re.match(r'from\stensorflow_datasets.*import\s([^\.\r\n]+)' , lowerCAmelCase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(',' ) ) __lowercase= 'from . import ' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f'Error converting {out_line.strip()}' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: __lowercase= True out_lines.append(lowerCAmelCase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset __lowercase= f_name.replace('.py' , '' ) __lowercase= os.path.join(lowerCAmelCase , lowerCAmelCase ) __lowercase= os.path.join(lowerCAmelCase , lowerCAmelCase ) os.makedirs(lowerCAmelCase , exist_ok=lowerCAmelCase ) self._logger.info(f'Adding directory {output_dir}' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(lowerCAmelCase ) if needs_manual_update: with_manual_update.append(lowerCAmelCase ) with open(lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.writelines(lowerCAmelCase ) self._logger.info(f'Converted in {output_file}' ) for utils_file in utils_files: try: __lowercase= os.path.basename(lowerCAmelCase ) __lowercase= imports_to_builder_map[f_name.replace('.py' , '' )] self._logger.info(f'Moving {dest_folder} to {utils_file}' ) shutil.copy(lowerCAmelCase , lowerCAmelCase ) except KeyError: self._logger.error(f'Cannot find destination folder for {utils_file}. Please copy manually.' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f'You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.' )
304
0
import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument A : Any = { "/attention/": "/0/SelfAttention/", "/self_attention/": "/0/SelfAttention/", "/encoder_decoder_attention/": "/1/EncDecAttention/", "value": "v", "query": "q", "key": "k", "out": "o", "pre_self_attention_layer_norm": "0/layer_norm", "pre_cross_attention_layer_norm": "1/layer_norm", "pre_attention_layer_norm": "0/layer_norm", # previously 1, but seems wrong "token_embedder": "shared", "encoder_norm": "final_layer_norm", "decoder_norm": "final_layer_norm", "relpos_bias/rel_embedding": "block/0/layer/0/SelfAttention/relative_attention_bias/weight", "router/router_weights/w/": "router/classifier/", "roer/roer_weights/w/": "router/classifier/", "logits_dense": "lm_head", } def lowercase_ ( _A : Tuple ): """simple docstring""" lowerCamelCase__ : str = list(s_dict.keys() ) for key in keys: lowerCamelCase__ : Optional[int] = r".*/layers_(\d+)" lowerCamelCase__ : Tuple = key if re.match(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowerCamelCase__ : Optional[int] = re.sub(r"layers_(\d+)" , r"block/\1/layer" , __SCREAMING_SNAKE_CASE ) lowerCamelCase__ : Dict = r"(encoder|decoder)\/" if re.match(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowerCamelCase__ : str = re.match(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ).groups() if groups[0] == "encoder": lowerCamelCase__ : Tuple = re.sub(r"/mlp/" , r"/1/mlp/" , __SCREAMING_SNAKE_CASE ) lowerCamelCase__ : str = re.sub(r"/pre_mlp_layer_norm/" , r"/1/layer_norm/" , __SCREAMING_SNAKE_CASE ) elif groups[0] == "decoder": lowerCamelCase__ : Optional[int] = re.sub(r"/mlp/" , r"/2/mlp/" , __SCREAMING_SNAKE_CASE ) lowerCamelCase__ : Union[str, Any] = re.sub(r"/pre_mlp_layer_norm/" , r"/2/layer_norm/" , __SCREAMING_SNAKE_CASE ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: lowerCamelCase__ : int = new_key.replace(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) print(F"{key} -> {new_key}" ) lowerCamelCase__ : Optional[Any] = s_dict.pop(__SCREAMING_SNAKE_CASE ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: lowerCamelCase__ : List[str] = s_dict[ "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: lowerCamelCase__ : Optional[int] = s_dict[ "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: lowerCamelCase__ : Optional[Any] = s_dict[key].shape[0] lowerCamelCase__ : Dict = s_dict[key] for idx in range(__SCREAMING_SNAKE_CASE ): lowerCamelCase__ : int = expert_weihts[idx] print(F"{key} -> {key.replace('expert/' , 'nested fstring' )}" ) s_dict.pop(__SCREAMING_SNAKE_CASE ) return s_dict A : Optional[int] = { "NUM_ENCODER_LAYERS": "num_layers", "NUM_DECODER_LAYERS": "num_decoder_layers", "NUM_HEADS": "num_heads", "HEAD_DIM": "d_kv", "EMBED_DIM": "d_model", "MLP_DIM": "d_ff", "NUM_SELECTED_EXPERTS": "num_selected_experts", "NUM_ENCODER_SPARSE_LAYERS": "num_sparse_encoder_layers", "NUM_DECODER_SPARSE_LAYERS": "num_sparse_decoder_layers", "dense.MlpBlock.activations": "feed_forward_proj", } def lowercase_ ( _A : List[Any] , _A : int ): """simple docstring""" import regex as re with open(__SCREAMING_SNAKE_CASE , "r" ) as f: lowerCamelCase__ : Union[str, Any] = f.read() lowerCamelCase__ : Dict = re.findall(r"(.*) = ([0-9.]*)" , __SCREAMING_SNAKE_CASE ) lowerCamelCase__ : Tuple = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": lowerCamelCase__ : Optional[Any] = float(__SCREAMING_SNAKE_CASE ) if "." in value else int(__SCREAMING_SNAKE_CASE ) lowerCamelCase__ : List[Any] = re.findall(r"(.*activations) = \(\'(.*)\',\)" , __SCREAMING_SNAKE_CASE )[0] lowerCamelCase__ : List[str] = str(activation[1] ) lowerCamelCase__ : Tuple = num_experts lowerCamelCase__ : Any = SwitchTransformersConfig(**__SCREAMING_SNAKE_CASE ) return config def lowercase_ ( _A : List[Any] , _A : Union[str, Any] , _A : Tuple=None , _A : Optional[int]="./" , _A : Optional[Any]=8 ): """simple docstring""" print(F"Loading flax weights from : {flax_checkpoint_path}" ) lowerCamelCase__ : Any = checkpoints.load_tax_checkpoint(__SCREAMING_SNAKE_CASE ) if gin_file is not None: lowerCamelCase__ : int = convert_gin_to_config(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else: lowerCamelCase__ : Tuple = SwitchTransformersConfig.from_pretrained(__SCREAMING_SNAKE_CASE ) lowerCamelCase__ : List[str] = SwitchTransformersForConditionalGeneration(__SCREAMING_SNAKE_CASE ) lowerCamelCase__ : str = flax_params["target"] lowerCamelCase__ : List[str] = flatten_dict(__SCREAMING_SNAKE_CASE , sep="/" ) lowerCamelCase__ : Dict = rename_keys(__SCREAMING_SNAKE_CASE ) lowerCamelCase__ : List[str] = unflatten_dict(__SCREAMING_SNAKE_CASE , sep="/" ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) print(F"Save PyTorch model to {pytorch_dump_path}" ) pt_model.save_pretrained(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": A : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( "--switch_t5x_checkpoint_path", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the" " model architecture. If not provided, a `gin_file` has to be provided." ), ) parser.add_argument( "--gin_file", default=None, type=str, required=False, help="Path to the gin config file. If not provided, a `config_file` has to be passed ", ) parser.add_argument( "--config_name", default=None, type=str, required=False, help="Config name of SwitchTransformers model." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output pytorch model." ) parser.add_argument("--num_experts", default=8, type=int, required=False, help="Number of experts") A : Union[str, Any] = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
184
import inspect from typing import Callable, List, Optional, Union import torch from transformers import ( CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, WhisperForConditionalGeneration, WhisperProcessor, ) from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.utils import logging UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class A_ ( __lowerCamelCase ): '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): super().__init__() if safety_checker is None: logger.warning( F'''You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure''' ' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered' ' results in services or applications open to the public. Both the diffusers team and Hugging Face' ' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling' ' it only for use-cases that involve analyzing network behavior or auditing its results. For more' ' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .' ) self.register_modules( speech_model=snake_case , speech_processor=snake_case , vae=snake_case , text_encoder=snake_case , tokenizer=snake_case , unet=snake_case , scheduler=snake_case , feature_extractor=snake_case , ) def SCREAMING_SNAKE_CASE__ ( self , snake_case = "auto" ): if slice_size == "auto": lowercase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): self.enable_attention_slicing(snake_case ) @torch.no_grad() def __call__( self , snake_case , snake_case=1_6000 , snake_case = 512 , snake_case = 512 , snake_case = 50 , snake_case = 7.5 , snake_case = None , snake_case = 1 , snake_case = 0.0 , snake_case = None , snake_case = None , snake_case = "pil" , snake_case = True , snake_case = None , snake_case = 1 , **snake_case , ): lowercase = self.speech_processor.feature_extractor( snake_case , return_tensors='pt' , sampling_rate=snake_case ).input_features.to(self.device ) lowercase = self.speech_model.generate(snake_case , max_length=48_0000 ) lowercase = self.speech_processor.tokenizer.batch_decode(snake_case , skip_special_tokens=snake_case , normalize=snake_case )[ 0 ] if isinstance(snake_case , snake_case ): lowercase = 1 elif isinstance(snake_case , snake_case ): lowercase = len(snake_case ) else: raise ValueError(F'''`prompt` has to be of type `str` or `list` but is {type(snake_case )}''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(snake_case , snake_case ) or callback_steps <= 0) ): raise ValueError( F'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' F''' {type(snake_case )}.''' ) # get prompt text embeddings lowercase = self.tokenizer( snake_case , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) lowercase = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: lowercase = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' F''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) lowercase = text_input_ids[:, : self.tokenizer.model_max_length] lowercase = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method lowercase , lowercase , lowercase = text_embeddings.shape lowercase = text_embeddings.repeat(1 , snake_case , 1 ) lowercase = text_embeddings.view(bs_embed * num_images_per_prompt , snake_case , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. lowercase = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: lowercase = 42 if negative_prompt is None: lowercase = [''] * batch_size elif type(snake_case ) is not type(snake_case ): raise TypeError( F'''`negative_prompt` should be the same type to `prompt`, but got {type(snake_case )} !=''' F''' {type(snake_case )}.''' ) elif isinstance(snake_case , snake_case ): lowercase = [negative_prompt] elif batch_size != len(snake_case ): raise ValueError( F'''`negative_prompt`: {negative_prompt} has batch size {len(snake_case )}, but `prompt`:''' F''' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches''' ' the batch size of `prompt`.' ) else: lowercase = negative_prompt lowercase = text_input_ids.shape[-1] lowercase = self.tokenizer( snake_case , padding='max_length' , max_length=snake_case , truncation=snake_case , return_tensors='pt' , ) lowercase = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method lowercase = uncond_embeddings.shape[1] lowercase = uncond_embeddings.repeat(1 , snake_case , 1 ) lowercase = uncond_embeddings.view(batch_size * num_images_per_prompt , snake_case , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowercase = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. lowercase = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) lowercase = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps lowercase = torch.randn(snake_case , generator=snake_case , device='cpu' , dtype=snake_case ).to( self.device ) else: lowercase = torch.randn(snake_case , generator=snake_case , device=self.device , dtype=snake_case ) else: if latents.shape != latents_shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) lowercase = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(snake_case ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand lowercase = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler lowercase = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowercase = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowercase = {} if accepts_eta: lowercase = eta for i, t in enumerate(self.progress_bar(snake_case ) ): # expand the latents if we are doing classifier free guidance lowercase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase = self.scheduler.scale_model_input(snake_case , snake_case ) # predict the noise residual lowercase = self.unet(snake_case , snake_case , encoder_hidden_states=snake_case ).sample # perform guidance if do_classifier_free_guidance: lowercase , lowercase = noise_pred.chunk(2 ) lowercase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 lowercase = self.scheduler.step(snake_case , snake_case , snake_case , **snake_case ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(snake_case , snake_case , snake_case ) lowercase = 1 / 0.18_215 * latents lowercase = self.vae.decode(snake_case ).sample lowercase = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowercase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase = self.numpy_to_pil(snake_case ) if not return_dict: return image return StableDiffusionPipelineOutput(images=snake_case , nsfw_content_detected=snake_case )
195
0
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ : def __init__( self : str , _lowerCamelCase : Any , _lowerCamelCase : Union[str, Any]=12 , _lowerCamelCase : str=7 , _lowerCamelCase : List[str]=True , _lowerCamelCase : Optional[Any]=True , _lowerCamelCase : int=True , _lowerCamelCase : Dict=99 , _lowerCamelCase : List[Any]=32 , _lowerCamelCase : Any=32 , _lowerCamelCase : int=2 , _lowerCamelCase : Union[str, Any]=4 , _lowerCamelCase : Union[str, Any]=37 , _lowerCamelCase : int=0.1 , _lowerCamelCase : Optional[Any]=0.1 , _lowerCamelCase : Optional[Any]=512 , _lowerCamelCase : List[str]=0.0_2 , _lowerCamelCase : Optional[Any]=0 , _lowerCamelCase : Union[str, Any]=None , ): _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_input_mask _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = projection_dim _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = dropout _snake_case = attention_dropout _snake_case = max_position_embeddings _snake_case = initializer_range _snake_case = scope _snake_case = bos_token_id def lowercase ( self : List[Any] ): _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = None if self.use_input_mask: _snake_case = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: _snake_case = input_mask.numpy() _snake_case , _snake_case = input_mask.shape _snake_case = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_lowerCamelCase ): _snake_case = 1 _snake_case = 0 _snake_case = self.get_config() return config, input_ids, tf.convert_to_tensor(_lowerCamelCase ) def lowercase ( self : str ): return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def lowercase ( self : int , _lowerCamelCase : str , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Union[str, Any] ): _snake_case = TFBlipTextModel(config=_lowerCamelCase ) _snake_case = model(_lowerCamelCase , attention_mask=_lowerCamelCase , training=_lowerCamelCase ) _snake_case = model(_lowerCamelCase , training=_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowercase ( self : Optional[Any] ): _snake_case = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case = config_and_inputs _snake_case = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowerCAmelCase__ ( A_ , unittest.TestCase ): __a = (TFBlipTextModel,) if is_tf_available() else () __a = False __a = False __a = False def lowercase ( self : List[Any] ): _snake_case = BlipTextModelTester(self ) _snake_case = ConfigTester(self , config_class=_lowerCamelCase , hidden_size=37 ) def lowercase ( self : Dict ): self.config_tester.run_common_tests() def lowercase ( self : Union[str, Any] ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def lowercase ( self : Dict ): pass def lowercase ( self : Any ): pass @unittest.skip(reason='''Blip does not use inputs_embeds''' ) def lowercase ( self : Optional[int] ): pass @unittest.skip(reason='''BlipTextModel has no base class and is not available in MODEL_MAPPING''' ) def lowercase ( self : List[Any] ): pass @unittest.skip(reason='''BlipTextModel has no base class and is not available in MODEL_MAPPING''' ) def lowercase ( self : Optional[int] ): pass @slow def lowercase ( self : Optional[int] ): for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = TFBlipTextModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def lowercase ( self : int , _lowerCamelCase : str=True ): super().test_pt_tf_model_equivalence(allow_missing_keys=_lowerCamelCase )
355
"""simple docstring""" def _UpperCAmelCase ( __lowerCamelCase : float , __lowerCamelCase : float ) -> float: if mass < 0: raise ValueError('''The mass of a body cannot be negative''' ) return 0.5 * mass * abs(__lowerCamelCase ) * abs(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
40
0
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging _lowerCAmelCase : Union[str, Any] = logging.get_logger(__name__) class _UpperCamelCase ( lowerCAmelCase ): UpperCAmelCase_ = """linear""" UpperCAmelCase_ = """cosine""" UpperCAmelCase_ = """cosine_with_restarts""" UpperCAmelCase_ = """polynomial""" UpperCAmelCase_ = """constant""" UpperCAmelCase_ = """constant_with_warmup""" UpperCAmelCase_ = """piecewise_constant""" def lowerCAmelCase ( _lowerCAmelCase : Optimizer , _lowerCAmelCase : int = -1 ): """simple docstring""" return LambdaLR(_lowerCAmelCase , lambda _lowerCAmelCase : 1 , last_epoch=_lowerCAmelCase ) def lowerCAmelCase ( _lowerCAmelCase : Optimizer , _lowerCAmelCase : int , _lowerCAmelCase : int = -1 ): """simple docstring""" def lr_lambda(_lowerCAmelCase : int ): if current_step < num_warmup_steps: return float(_lowerCAmelCase ) / float(max(1.0 , _lowerCAmelCase ) ) return 1.0 return LambdaLR(_lowerCAmelCase , _lowerCAmelCase , last_epoch=_lowerCAmelCase ) def lowerCAmelCase ( _lowerCAmelCase : Optimizer , _lowerCAmelCase : str , _lowerCAmelCase : int = -1 ): """simple docstring""" UpperCAmelCase__ = {} UpperCAmelCase__ = step_rules.split("," ) for rule_str in rule_list[:-1]: UpperCAmelCase__ , UpperCAmelCase__ = rule_str.split(":" ) UpperCAmelCase__ = int(_lowerCAmelCase ) UpperCAmelCase__ = float(_lowerCAmelCase ) UpperCAmelCase__ = value UpperCAmelCase__ = float(rule_list[-1] ) def create_rules_function(_lowerCAmelCase : Any , _lowerCAmelCase : Tuple ): def rule_func(_lowerCAmelCase : int ) -> float: UpperCAmelCase__ = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(_lowerCAmelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func UpperCAmelCase__ = create_rules_function(_lowerCAmelCase , _lowerCAmelCase ) return LambdaLR(_lowerCAmelCase , _lowerCAmelCase , last_epoch=_lowerCAmelCase ) def lowerCAmelCase ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple=-1 ): """simple docstring""" def lr_lambda(_lowerCAmelCase : int ): if current_step < num_warmup_steps: return float(_lowerCAmelCase ) / float(max(1 , _lowerCAmelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase ( _lowerCAmelCase : Optimizer , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 0.5 , _lowerCAmelCase : int = -1 ): """simple docstring""" def lr_lambda(_lowerCAmelCase : List[str] ): if current_step < num_warmup_steps: return float(_lowerCAmelCase ) / float(max(1 , _lowerCAmelCase ) ) UpperCAmelCase__ = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(_lowerCAmelCase ) * 2.0 * progress )) ) return LambdaLR(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase ( _lowerCAmelCase : Optimizer , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : int = 1 , _lowerCAmelCase : int = -1 ): """simple docstring""" def lr_lambda(_lowerCAmelCase : int ): if current_step < num_warmup_steps: return float(_lowerCAmelCase ) / float(max(1 , _lowerCAmelCase ) ) UpperCAmelCase__ = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(_lowerCAmelCase ) * progress) % 1.0) )) ) return LambdaLR(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any]=1E-7 , _lowerCAmelCase : Optional[int]=1.0 , _lowerCAmelCase : Tuple=-1 ): """simple docstring""" UpperCAmelCase__ = optimizer.defaults["lr"] if not (lr_init > lr_end): raise ValueError(F'''lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})''' ) def lr_lambda(_lowerCAmelCase : int ): if current_step < num_warmup_steps: return float(_lowerCAmelCase ) / float(max(1 , _lowerCAmelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: UpperCAmelCase__ = lr_init - lr_end UpperCAmelCase__ = num_training_steps - num_warmup_steps UpperCAmelCase__ = 1 - (current_step - num_warmup_steps) / decay_steps UpperCAmelCase__ = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) _lowerCAmelCase : Tuple = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def lowerCAmelCase ( _lowerCAmelCase : Union[str, SchedulerType] , _lowerCAmelCase : Optimizer , _lowerCAmelCase : Optional[str] = None , _lowerCAmelCase : Optional[int] = None , _lowerCAmelCase : Optional[int] = None , _lowerCAmelCase : int = 1 , _lowerCAmelCase : float = 1.0 , _lowerCAmelCase : int = -1 , ): """simple docstring""" UpperCAmelCase__ = SchedulerType(_lowerCAmelCase ) UpperCAmelCase__ = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(_lowerCAmelCase , last_epoch=_lowerCAmelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(_lowerCAmelCase , step_rules=_lowerCAmelCase , last_epoch=_lowerCAmelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F'''{name} requires `num_warmup_steps`, please provide that argument.''' ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(_lowerCAmelCase , num_warmup_steps=_lowerCAmelCase , last_epoch=_lowerCAmelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F'''{name} requires `num_training_steps`, please provide that argument.''' ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( _lowerCAmelCase , num_warmup_steps=_lowerCAmelCase , num_training_steps=_lowerCAmelCase , num_cycles=_lowerCAmelCase , last_epoch=_lowerCAmelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( _lowerCAmelCase , num_warmup_steps=_lowerCAmelCase , num_training_steps=_lowerCAmelCase , power=_lowerCAmelCase , last_epoch=_lowerCAmelCase , ) return schedule_func( _lowerCAmelCase , num_warmup_steps=_lowerCAmelCase , num_training_steps=_lowerCAmelCase , last_epoch=_lowerCAmelCase )
169
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : List[Any] = logging.get_logger(__name__) _lowerCAmelCase : List[str] = { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json" ), "google/realm-orqa-nq-openqa": "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json", "google/realm-orqa-nq-reader": "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json", "google/realm-orqa-wq-openqa": "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json", "google/realm-orqa-wq-reader": "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json", # See all REALM models at https://huggingface.co/models?filter=realm } class _UpperCamelCase ( lowerCAmelCase ): UpperCAmelCase_ = """realm""" def __init__( self :str , lowerCamelCase :List[Any]=3_0522 , lowerCamelCase :Optional[int]=768 , lowerCamelCase :Any=128 , lowerCamelCase :Tuple=12 , lowerCamelCase :str=12 , lowerCamelCase :List[str]=8 , lowerCamelCase :List[str]=3072 , lowerCamelCase :List[str]="gelu_new" , lowerCamelCase :int=0.1 , lowerCamelCase :Optional[Any]=0.1 , lowerCamelCase :int=512 , lowerCamelCase :Union[str, Any]=2 , lowerCamelCase :str=0.02 , lowerCamelCase :Tuple=1e-12 , lowerCamelCase :Dict=256 , lowerCamelCase :int=10 , lowerCamelCase :List[str]=1e-3 , lowerCamelCase :str=5 , lowerCamelCase :Optional[int]=320 , lowerCamelCase :Union[str, Any]=1335_3718 , lowerCamelCase :str=5000 , lowerCamelCase :str=1 , lowerCamelCase :List[Any]=0 , lowerCamelCase :Tuple=2 , **lowerCamelCase :Optional[int] , ) -> Optional[Any]: super().__init__(pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , **lowerCamelCase ) # Common config UpperCAmelCase__ = vocab_size UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = hidden_size UpperCAmelCase__ = retriever_proj_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = num_candidates UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = initializer_range UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = layer_norm_eps # Reader config UpperCAmelCase__ = span_hidden_size UpperCAmelCase__ = max_span_width UpperCAmelCase__ = reader_layer_norm_eps UpperCAmelCase__ = reader_beam_size UpperCAmelCase__ = reader_seq_len # Retrieval config UpperCAmelCase__ = num_block_records UpperCAmelCase__ = searcher_beam_size
169
1
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class A__ ( unittest.TestCase ): """simple docstring""" def a_ ( self ): snake_case = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) snake_case = Vector() def a_ ( self ): snake_case = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(__snake_case ) , '''(0,0,0,0,0,1)''' ) def a_ ( self ): snake_case = Vector([1, 2, 3, 4] ) self.assertEqual(len(__snake_case ) , 4 ) def a_ ( self ): snake_case = Vector([1, 2] ) snake_case = Vector([1, 2, 3, 4, 5] ) snake_case = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) snake_case = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def a_ ( self ): snake_case = Vector([1, 2, 3] ) snake_case = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def a_ ( self ): snake_case = Vector([1, 2, 3] ) snake_case = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def a_ ( self ): snake_case = Vector([1, 2, 3] ) snake_case = Vector([2, -1, 4] ) # for test of dot product snake_case = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def a_ ( self ): self.assertEqual(str(zero_vector(1_0 ) ).count('''0''' ) , 1_0 ) def a_ ( self ): self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def a_ ( self ): snake_case = Vector([1, 2, 3] ) snake_case = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , __snake_case , __snake_case ) ) , '''(3,4,7)''' ) def a_ ( self ): snake_case = Vector([1, 0, 0, 0, 0, 0] ) snake_case = x.copy() self.assertEqual(str(__snake_case ) , str(__snake_case ) ) def a_ ( self ): snake_case = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(__snake_case ) , '''(0,1,0)''' ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(__snake_case ) ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) snake_case = [[-3, -1_4, -1_0], [-5, -1_0, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(__snake_case , __snake_case ) ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) snake_case = [[-3, 1_4, -1_0], [5, -1_0, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(__snake_case , __snake_case ) ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) snake_case = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(__snake_case ) ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) snake_case = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 1_0]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def a_ ( self ): snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) snake_case = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 1_0]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def a_ ( self ): self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
360
import re from filelock import FileLock try: import nltk _SCREAMING_SNAKE_CASE : Union[str, Any] = True except (ImportError, ModuleNotFoundError): _SCREAMING_SNAKE_CASE : Optional[Any] = False if NLTK_AVAILABLE: with FileLock(".lock") as lock: nltk.download("punkt", quiet=True) def UpperCAmelCase__ (UpperCamelCase_ ): """simple docstring""" re.sub('''<n>''' ,'''''' ,UpperCamelCase_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(UpperCamelCase_ ) )
213
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImgaImgPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import floats_tensor, load_image, load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class A ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase = ShapEImgaImgPipeline lowerCamelCase = ["""image"""] lowerCamelCase = ["""image"""] lowerCamelCase = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] lowerCamelCase = False @property def snake_case__ ( self : Union[str, Any] )-> Union[str, Any]: '''simple docstring''' return 3_2 @property def snake_case__ ( self : Dict )-> Any: '''simple docstring''' return 3_2 @property def snake_case__ ( self : Dict )-> List[str]: '''simple docstring''' return self.time_input_dim * 4 @property def snake_case__ ( self : Union[str, Any] )-> Union[str, Any]: '''simple docstring''' return 8 @property def snake_case__ ( self : Union[str, Any] )-> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) A__ = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size,image_size=6_4,projection_dim=self.text_embedder_hidden_size,intermediate_size=3_7,num_attention_heads=4,num_channels=3,num_hidden_layers=5,patch_size=1,) A__ = CLIPVisionModel(lowerCamelCase__ ) return model @property def snake_case__ ( self : Tuple )-> List[str]: '''simple docstring''' A__ = CLIPImageProcessor( crop_size=2_2_4,do_center_crop=lowerCamelCase__,do_normalize=lowerCamelCase__,do_resize=lowerCamelCase__,image_mean=[0.48_145_466, 0.4_578_275, 0.40_821_073],image_std=[0.26_862_954, 0.26_130_258, 0.27_577_711],resample=3,size=2_2_4,) return image_processor @property def snake_case__ ( self : List[str] )-> Tuple: '''simple docstring''' torch.manual_seed(0 ) A__ = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'embedding_proj_norm_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } A__ = PriorTransformer(**lowerCamelCase__ ) return model @property def snake_case__ ( self : Union[str, Any] )-> Union[str, Any]: '''simple docstring''' torch.manual_seed(0 ) A__ = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } A__ = ShapERenderer(**lowerCamelCase__ ) return model def snake_case__ ( self : List[Any] )-> Tuple: '''simple docstring''' A__ = self.dummy_prior A__ = self.dummy_image_encoder A__ = self.dummy_image_processor A__ = self.dummy_renderer A__ = HeunDiscreteScheduler( beta_schedule='exp',num_train_timesteps=1_0_2_4,prediction_type='sample',use_karras_sigmas=lowerCamelCase__,clip_sample=lowerCamelCase__,clip_sample_range=1.0,) A__ = { 'prior': prior, 'image_encoder': image_encoder, 'image_processor': image_processor, 'renderer': renderer, 'scheduler': scheduler, } return components def snake_case__ ( self : str,lowercase_ : Dict,lowercase_ : int=0 )-> List[Any]: '''simple docstring''' A__ = floats_tensor((1, 3, 6_4, 6_4),rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith('mps' ): A__ = torch.manual_seed(lowerCamelCase__ ) else: A__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) A__ = { 'image': input_image, 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def snake_case__ ( self : Optional[Any] )-> Optional[Any]: '''simple docstring''' A__ = 'cpu' A__ = self.get_dummy_components() A__ = self.pipeline_class(**lowerCamelCase__ ) A__ = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) A__ = pipe(**self.get_dummy_inputs(lowerCamelCase__ ) ) A__ = output.images[0] A__ = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) A__ = np.array( [ 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def snake_case__ ( self : Dict )-> int: '''simple docstring''' self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def snake_case__ ( self : str )-> int: '''simple docstring''' A__ = torch_device == 'cpu' A__ = True self._test_inference_batch_single_identical( batch_size=2,test_max_difference=lowerCamelCase__,relax_max_difference=lowerCamelCase__,) def snake_case__ ( self : Optional[Any] )-> Union[str, Any]: '''simple docstring''' A__ = self.get_dummy_components() A__ = self.pipeline_class(**lowerCamelCase__ ) A__ = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) A__ = 1 A__ = 2 A__ = self.get_dummy_inputs(lowerCamelCase__ ) for key in inputs.keys(): if key in self.batch_params: A__ = batch_size * [inputs[key]] A__ = pipe(**lowerCamelCase__,num_images_per_prompt=lowerCamelCase__ )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class A ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Any )-> int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self : Optional[int] )-> Any: '''simple docstring''' A__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/corgi.png' ) A__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_img2img_out.npy' ) A__ = ShapEImgaImgPipeline.from_pretrained('openai/shap-e-img2img' ) A__ = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) A__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) A__ = pipe( lowerCamelCase__,generator=lowerCamelCase__,guidance_scale=3.0,num_inference_steps=6_4,frame_size=6_4,output_type='np',).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(lowerCamelCase__,lowerCamelCase__ )
7
A_ :Union[str, Any] = { 0: '''0''', 1: '''1''', 2: '''2''', 3: '''3''', 4: '''4''', 5: '''5''', 6: '''6''', 7: '''7''', 8: '''8''', 9: '''9''', 10: '''a''', 11: '''b''', 12: '''c''', 13: '''d''', 14: '''e''', 15: '''f''', } def A ( a_ ) -> str: assert type(a_ ) in (int, float) and decimal == int(a_ ) __UpperCamelCase : Union[str, Any] =int(a_ ) __UpperCamelCase : List[str] ='' __UpperCamelCase : Optional[Any] =False if decimal < 0: __UpperCamelCase : Tuple =True decimal *= -1 while decimal > 0: __UpperCamelCase , __UpperCamelCase : Optional[Any] =divmod(a_ ,16 ) __UpperCamelCase : Tuple =values[remainder] + hexadecimal __UpperCamelCase : Dict ='0x' + hexadecimal if negative: __UpperCamelCase : int ='-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
71
0
'''simple docstring''' import collections import importlib.util import os import re from pathlib import Path A__ : str = '''src/transformers''' # Matches is_xxx_available() A__ : str = re.compile(R'''is\_([a-z_]*)_available()''') # Catches a one-line _import_struct = {xxx} A__ : List[str] = re.compile(R'''^_import_structure\s+=\s+\{([^\}]+)\}''') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] A__ : Any = re.compile(R'''\s+"\S*":\s+\[([^\]]*)\]''') # Catches a line if not is_foo_available A__ : Any = re.compile(R'''^\s*if\s+not\s+is\_[a-z_]*\_available\(\)''') # Catches a line _import_struct["bla"].append("foo") A__ : Union[str, Any] = re.compile(R'''^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)''') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] A__ : List[str] = re.compile(R'''^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]''') # Catches a line with an object between quotes and a comma: "MyModel", A__ : str = re.compile('''^\s+"([^"]+)",''') # Catches a line with objects between brackets only: ["foo", "bar"], A__ : str = re.compile('''^\s+\[([^\]]+)\]''') # Catches a line with from foo import bar, bla, boo A__ : Union[str, Any] = re.compile(R'''\s+from\s+\S*\s+import\s+([^\(\s].*)\n''') # Catches a line with try: A__ : str = re.compile(R'''^\s*try:''') # Catches a line with else: A__ : Optional[int] = re.compile(R'''^\s*else:''') def a_ ( _UpperCAmelCase : Union[str, Any] ) -> int: if _re_test_backend.search(_UpperCAmelCase ) is None: return None __snake_case : Dict = [b[0] for b in _re_backend.findall(_UpperCAmelCase )] backends.sort() return "_and_".join(_UpperCAmelCase ) def a_ ( _UpperCAmelCase : Tuple ) -> Dict: with open(_UpperCAmelCase ,'r' ,encoding='utf-8' ,newline='\n' ) as f: __snake_case : Dict = f.readlines() __snake_case : Optional[int] = 0 while line_index < len(_UpperCAmelCase ) and not lines[line_index].startswith('_import_structure = {' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(_UpperCAmelCase ): return None # First grab the objects without a specific backend in _import_structure __snake_case : Dict = [] while not lines[line_index].startswith('if TYPE_CHECKING' ) and find_backend(lines[line_index] ) is None: __snake_case : Tuple = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(_UpperCAmelCase ): __snake_case : Tuple = _re_one_line_import_struct.search(_UpperCAmelCase ).groups()[0] __snake_case : str = re.findall('\[([^\]]+)\]' ,_UpperCAmelCase ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(', ' )] ) line_index += 1 continue __snake_case : int = _re_import_struct_key_value.search(_UpperCAmelCase ) if single_line_import_search is not None: __snake_case : Dict = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(', ' ) if len(_UpperCAmelCase ) > 0] objects.extend(_UpperCAmelCase ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) line_index += 1 __snake_case : Union[str, Any] = {'none': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('if TYPE_CHECKING' ): # If the line is an if not is_backend_available, we grab all objects associated. __snake_case : Tuple = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: __snake_case : List[str] = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 __snake_case : str = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 4 ): __snake_case : List[str] = lines[line_index] if _re_import_struct_add_one.search(_UpperCAmelCase ) is not None: objects.append(_re_import_struct_add_one.search(_UpperCAmelCase ).groups()[0] ) elif _re_import_struct_add_many.search(_UpperCAmelCase ) is not None: __snake_case : Any = _re_import_struct_add_many.search(_UpperCAmelCase ).groups()[0].split(', ' ) __snake_case : Optional[Any] = [obj[1:-1] for obj in imports if len(_UpperCAmelCase ) > 0] objects.extend(_UpperCAmelCase ) elif _re_between_brackets.search(_UpperCAmelCase ) is not None: __snake_case : List[Any] = _re_between_brackets.search(_UpperCAmelCase ).groups()[0].split(', ' ) __snake_case : List[Any] = [obj[1:-1] for obj in imports if len(_UpperCAmelCase ) > 0] objects.extend(_UpperCAmelCase ) elif _re_quote_object.search(_UpperCAmelCase ) is not None: objects.append(_re_quote_object.search(_UpperCAmelCase ).groups()[0] ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) elif line.startswith(' ' * 12 + '"' ): objects.append(line[13:-3] ) line_index += 1 __snake_case : Tuple = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend __snake_case : str = [] while ( line_index < len(_UpperCAmelCase ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('else' ) ): __snake_case : Any = lines[line_index] __snake_case : Tuple = _re_import.search(_UpperCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 8 ): objects.append(line[8:-2] ) line_index += 1 __snake_case : List[str] = {'none': objects} # Let's continue with backend-specific objects while line_index < len(_UpperCAmelCase ): # If the line is an if is_backend_available, we grab all objects associated. __snake_case : List[Any] = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: __snake_case : Dict = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 __snake_case : Optional[Any] = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 8 ): __snake_case : Optional[int] = lines[line_index] __snake_case : Optional[int] = _re_import.search(_UpperCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 12 ): objects.append(line[12:-2] ) line_index += 1 __snake_case : Optional[Any] = objects else: line_index += 1 return import_dict_objects, type_hint_objects def a_ ( _UpperCAmelCase : Dict ,_UpperCAmelCase : List[Any] ) -> Dict: def find_duplicates(_UpperCAmelCase : Tuple ): return [k for k, v in collections.Counter(_UpperCAmelCase ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] __snake_case : Optional[Any] = [] for key in import_dict_objects.keys(): __snake_case : List[Any] = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(f'''Duplicate _import_structure definitions for: {duplicate_imports}''' ) __snake_case : Optional[int] = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(f'''Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}''' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): __snake_case : Tuple = 'base imports' if key == 'none' else f'''{key} backend''' errors.append(f'''Differences for {name}:''' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(f''' {a} in TYPE_HINT but not in _import_structure.''' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(f''' {a} in _import_structure but not in TYPE_HINT.''' ) return errors def a_ ( ) -> int: __snake_case : Tuple = [] for root, _, files in os.walk(_UpperCAmelCase ): if "__init__.py" in files: __snake_case : Any = os.path.join(_UpperCAmelCase ,'__init__.py' ) __snake_case : Union[str, Any] = parse_init(_UpperCAmelCase ) if objects is not None: __snake_case : Tuple = analyze_results(*_UpperCAmelCase ) if len(_UpperCAmelCase ) > 0: __snake_case : Any = f'''Problem in {fname}, both halves do not define the same objects.\n{errors[0]}''' failures.append('\n'.join(_UpperCAmelCase ) ) if len(_UpperCAmelCase ) > 0: raise ValueError('\n\n'.join(_UpperCAmelCase ) ) def a_ ( ) -> Any: __snake_case : Optional[int] = [] for path, directories, files in os.walk(_UpperCAmelCase ): for folder in directories: # Ignore private modules if folder.startswith('_' ): directories.remove(_UpperCAmelCase ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(_UpperCAmelCase ) / folder).glob('*.py' ) ) ) == 0: continue __snake_case : Dict = str((Path(_UpperCAmelCase ) / folder).relative_to(_UpperCAmelCase ) ) __snake_case : Any = short_path.replace(os.path.sep ,'.' ) submodules.append(_UpperCAmelCase ) for fname in files: if fname == "__init__.py": continue __snake_case : Any = str((Path(_UpperCAmelCase ) / fname).relative_to(_UpperCAmelCase ) ) __snake_case : Any = short_path.replace('.py' ,'' ).replace(os.path.sep ,'.' ) if len(submodule.split('.' ) ) == 1: submodules.append(_UpperCAmelCase ) return submodules A__ : Tuple = [ '''convert_pytorch_checkpoint_to_tf2''', '''modeling_flax_pytorch_utils''', ] def a_ ( ) -> Any: # This is to make sure the transformers module imported is the one in the repo. __snake_case : Any = importlib.util.spec_from_file_location( 'transformers' ,os.path.join(_UpperCAmelCase ,'__init__.py' ) ,submodule_search_locations=[PATH_TO_TRANSFORMERS] ,) __snake_case : Tuple = spec.loader.load_module() __snake_case : Optional[int] = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in transformers._import_structure.keys() ] if len(_UpperCAmelCase ) > 0: __snake_case : Any = '\n'.join(f'''- {module}''' for module in module_not_registered ) raise ValueError( 'The following submodules are not properly registered in the main init of Transformers:\n' f'''{list_of_modules}\n''' 'Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.' ) if __name__ == "__main__": check_all_inits() check_submodules()
0
'''simple docstring''' from math import factorial A__ : dict[str, int] = {str(digit): factorial(digit) for digit in range(1_0)} def a_ ( _UpperCAmelCase : int ) -> int: if not isinstance(_UpperCAmelCase ,_UpperCAmelCase ): raise TypeError('Parameter number must be int' ) if number < 0: raise ValueError('Parameter number must be greater than or equal to 0' ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(_UpperCAmelCase ) ) def a_ ( _UpperCAmelCase : int = 60 ,_UpperCAmelCase : int = 1_00_00_00 ) -> int: if not isinstance(_UpperCAmelCase ,_UpperCAmelCase ) or not isinstance(_UpperCAmelCase ,_UpperCAmelCase ): raise TypeError('Parameters chain_length and number_limit must be int' ) if chain_length <= 0 or number_limit <= 0: raise ValueError( 'Parameters chain_length and number_limit must be greater than 0' ) # the counter for the chains with the exact desired length __snake_case : List[str] = 0 # the cached sizes of the previous chains __snake_case : dict[int, int] = {} for start_chain_element in range(1 ,_UpperCAmelCase ): # The temporary set will contain the elements of the chain __snake_case : Optional[int] = set() __snake_case : List[Any] = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. __snake_case : str = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(_UpperCAmelCase ) chain_set_length += 1 __snake_case : Tuple = digit_factorial_sum(_UpperCAmelCase ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] __snake_case : Optional[Any] = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F"""{solution()}""")
0
1
import math import tensorflow as tf from packaging import version def __magic_name__ ( A : Any ): '''simple docstring''' a = tf.convert_to_tensor(A ) a = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ), x.dtype ) )) return x * cdf def __magic_name__ ( A : List[Any] ): '''simple docstring''' a = tf.convert_to_tensor(A ) a = tf.cast(math.pi, x.dtype ) a = tf.cast(0.04_47_15, x.dtype ) a = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(A, 3 )) )) return x * cdf def __magic_name__ ( A : Optional[int] ): '''simple docstring''' a = tf.convert_to_tensor(A ) return x * tf.tanh(tf.math.softplus(A ) ) def __magic_name__ ( A : Optional[Any] ): '''simple docstring''' a = tf.convert_to_tensor(A ) a = tf.cast(0.04_47_15, x.dtype ) a = tf.cast(0.79_78_84_56_08, x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def __magic_name__ ( A : Dict ): '''simple docstring''' a = tf.convert_to_tensor(A ) a = tf.cast(1.7_02, x.dtype ) return x * tf.math.sigmoid(coeff * x ) def __magic_name__ ( A : int ): '''simple docstring''' return tf.clip_by_value(_gelu(A ), -10, 10 ) def __magic_name__ ( A : Optional[Any], A : int=-1 ): '''simple docstring''' a , a = tf.split(A, 2, axis=A ) return a * tf.math.sigmoid(A ) if version.parse(tf.version.VERSION) >= version.parse('2.4'): def __magic_name__ ( A : List[str] ): '''simple docstring''' return tf.keras.activations.gelu(A, approximate=A ) __lowerCAmelCase : Optional[Any] = tf.keras.activations.gelu __lowerCAmelCase : Optional[int] = approximate_gelu_wrap else: __lowerCAmelCase : Union[str, Any] = _gelu __lowerCAmelCase : Any = _gelu_new __lowerCAmelCase : Dict = { 'gelu': gelu, 'gelu_10': gelu_aa, 'gelu_fast': gelu_fast, 'gelu_new': gelu_new, 'glu': glu, 'mish': mish, 'quick_gelu': quick_gelu, 'relu': tf.keras.activations.relu, 'sigmoid': tf.keras.activations.sigmoid, 'silu': tf.keras.activations.swish, 'swish': tf.keras.activations.swish, 'tanh': tf.keras.activations.tanh, } def __magic_name__ ( A : Optional[int] ): '''simple docstring''' if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
107
def __magic_name__ ( A : str ): '''simple docstring''' a = "" for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def __magic_name__ ( A : str ): '''simple docstring''' a = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key a = remove_duplicates(key.upper() ) a = len(A ) # First fill cipher with key characters a = {alphabet[i]: char for i, char in enumerate(A )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(A ), 26 ): a = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 a = alphabet[i - offset] a = char return cipher_alphabet def __magic_name__ ( A : str, A : dict[str, str] ): '''simple docstring''' return "".join(cipher_map.get(A, A ) for ch in message.upper() ) def __magic_name__ ( A : str, A : dict[str, str] ): '''simple docstring''' a = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(A, A ) for ch in message.upper() ) def __magic_name__ ( ): '''simple docstring''' a = input("Enter message to encode or decode: " ).strip() a = input("Enter keyword: " ).strip() a = input("Encipher or decipher? E/D:" ).strip()[0].lower() try: a = {"e": encipher, "d": decipher}[option] except KeyError: raise KeyError("invalid input option" ) a = create_cipher_map(A ) print(func(A, A ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
107
1
'''simple docstring''' import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging UpperCamelCase__ : List[Any] = logging.get_logger(__name__) class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = ['''input_values''', '''attention_mask'''] def __init__( self , _lowerCamelCase = 1 , _lowerCamelCase = 1_6000 , _lowerCamelCase = 0.0 , _lowerCamelCase = False , _lowerCamelCase = 80 , _lowerCamelCase = 16 , _lowerCamelCase = 64 , _lowerCamelCase = "hann_window" , _lowerCamelCase = 1.0 , _lowerCamelCase = 80 , _lowerCamelCase = 7600 , _lowerCamelCase = 1e-10 , _lowerCamelCase = 2 , _lowerCamelCase = True , **_lowerCamelCase , ) -> List[Any]: super().__init__(feature_size=_lowerCamelCase , sampling_rate=_lowerCamelCase , padding_value=_lowerCamelCase , **_lowerCamelCase ) A_ : List[Any] = do_normalize A_ : Union[str, Any] = return_attention_mask A_ : Tuple = num_mel_bins A_ : List[str] = hop_length A_ : int = win_length A_ : Optional[int] = win_function A_ : List[Any] = frame_signal_scale A_ : str = fmin A_ : Optional[Any] = fmax A_ : Any = mel_floor A_ : Any = reduction_factor A_ : Tuple = win_length * sampling_rate // 1000 A_ : Dict = hop_length * sampling_rate // 1000 A_ : Dict = optimal_fft_length(self.sample_size ) A_ : str = (self.n_fft // 2) + 1 A_ : int = window_function(window_length=self.sample_size , name=self.win_function , periodic=_lowerCamelCase ) A_ : Optional[int] = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm="""slaney""" , mel_scale="""slaney""" , ) if frame_signal_scale != 1.0: warnings.warn( """The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers""" , _lowerCamelCase , ) if reduction_factor != 2.0: warnings.warn( """The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers""" , _lowerCamelCase , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def UpperCAmelCase_ ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 0.0 ) -> List[np.ndarray]: if attention_mask is not None: A_ : Dict = np.array(_lowerCamelCase , np.intaa ) A_ : Dict = [] for vector, length in zip(_lowerCamelCase , attention_mask.sum(-1 ) ): A_ : Optional[int] = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7 ) if length < normed_slice.shape[0]: A_ : Any = padding_value normed_input_values.append(_lowerCamelCase ) else: A_ : List[str] = [(x - x.mean()) / np.sqrt(x.var() + 1e-7 ) for x in input_values] return normed_input_values def UpperCAmelCase_ ( self , _lowerCamelCase , ) -> np.ndarray: A_ : int = spectrogram( _lowerCamelCase , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel="""log10""" , ) return log_mel_spec.T def __call__( self , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = False , _lowerCamelCase = None , _lowerCamelCase = False , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , **_lowerCamelCase , ) -> BatchFeature: if audio is None and audio_target is None: raise ValueError("""You must provide either `audio` or `audio_target` values.""" ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" F" {self.sampling_rate}. Please make sure that the provided audio input was sampled with" F" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) if audio is not None: A_ : Dict = self._process_audio( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase , ) else: A_ : Optional[int] = None if audio_target is not None: A_ : Union[str, Any] = self._process_audio( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase , ) if inputs is None: return inputs_target else: A_ : Optional[int] = inputs_target["""input_values"""] A_ : Tuple = inputs_target.get("""attention_mask""" ) if decoder_attention_mask is not None: A_ : int = decoder_attention_mask return inputs def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = False , _lowerCamelCase = False , _lowerCamelCase = None , _lowerCamelCase = False , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , **_lowerCamelCase , ) -> BatchFeature: A_ : Optional[int] = isinstance(_lowerCamelCase , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) A_ : List[str] = is_batched_numpy or ( isinstance(_lowerCamelCase , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: A_ : Union[str, Any] = [np.asarray(_lowerCamelCase , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(_lowerCamelCase , np.ndarray ): A_ : List[str] = np.asarray(_lowerCamelCase , dtype=np.floataa ) elif isinstance(_lowerCamelCase , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): A_ : Optional[Any] = speech.astype(np.floataa ) # always return batch if not is_batched: A_ : int = [speech] # needed to make pad() work on spectrogram inputs A_ : List[Any] = self.feature_size # convert into correct format for padding if is_target: A_ : Tuple = [self._extract_mel_features(_lowerCamelCase ) for waveform in speech] A_ : Tuple = BatchFeature({"""input_values""": features} ) A_ : Dict = self.num_mel_bins else: A_ : Union[str, Any] = BatchFeature({"""input_values""": speech} ) A_ : Tuple = self.pad( _lowerCamelCase , padding=_lowerCamelCase , max_length=_lowerCamelCase , truncation=_lowerCamelCase , pad_to_multiple_of=_lowerCamelCase , return_attention_mask=_lowerCamelCase , **_lowerCamelCase , ) A_ : Union[str, Any] = feature_size_hack # convert input values to correct format A_ : str = padded_inputs["""input_values"""] if not isinstance(input_values[0] , np.ndarray ): A_ : str = [np.asarray(_lowerCamelCase , dtype=np.floataa ) for array in input_values] elif ( not isinstance(_lowerCamelCase , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): A_ : Tuple = [array.astype(np.floataa ) for array in input_values] elif isinstance(_lowerCamelCase , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): A_ : List[Any] = input_values.astype(np.floataa ) # convert attention_mask to correct format A_ : Any = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: A_ : str = [np.asarray(_lowerCamelCase , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: A_ : Any = ( attention_mask if self._get_padding_strategies(_lowerCamelCase , max_length=_lowerCamelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) A_ : Any = self.zero_mean_unit_var_norm( padded_inputs["""input_values"""] , attention_mask=_lowerCamelCase , padding_value=self.padding_value ) if return_tensors is not None: A_ : Dict = padded_inputs.convert_to_tensors(_lowerCamelCase ) return padded_inputs def UpperCAmelCase_ ( self ) -> Dict[str, Any]: A_ : List[Any] = super().to_dict() # Don't serialize these as they are derived from the other properties. A_ : Optional[int] = ["""window""", """mel_filters""", """sample_size""", """sample_stride""", """n_fft""", """n_freqs"""] for name in names: if name in output: del output[name] return output
164
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> int: """simple docstring""" A_ : int = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): A_ : Tuple = n - k # Calculate C(n,k) for i in range(a_ ): result *= n - i result //= i + 1 return result def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , a_ ) // (node_count + 1) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" if n < 0: raise ValueError("""factorial() not defined for negative values""" ) A_ : Union[str, Any] = 1 for i in range(1 , n + 1 ): result *= i return result def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return catalan_number(a_ ) * factorial(a_ ) if __name__ == "__main__": UpperCamelCase__ : Any = int(input('Enter the number of nodes: ').strip() or 0) if node_count <= 0: raise ValueError('We need some nodes to work with.') print( f'Given {node_count} nodes, there are {binary_tree_count(node_count)} ' f'binary trees and {catalan_number(node_count)} binary search trees.' )
164
1
import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP _UpperCAmelCase : Union[str, Any] = False try: _UpperCAmelCase : Optional[Any] = _is_package_available("""google.colab""") except ModuleNotFoundError: pass @input.register class lowerCAmelCase : def __init__( self : List[Any] , UpperCAmelCase : str = None , UpperCAmelCase : list = [] ) -> Optional[int]: lowerCamelCase__ : Optional[int] = 0 lowerCamelCase__ : Dict = choices lowerCamelCase__ : Union[str, Any] = prompt if sys.platform == "win32": lowerCamelCase__ : List[str] = '*' else: lowerCamelCase__ : Optional[Any] = '➔ ' def A_ ( self : Optional[int] , UpperCAmelCase : Optional[int] , UpperCAmelCase : str = "" ) -> Tuple: if sys.platform != "win32": writeColor(self.choices[index] , 32 , UpperCAmelCase ) else: forceWrite(self.choices[index] , UpperCAmelCase ) def A_ ( self : int , UpperCAmelCase : int ) -> Any: if index == self.position: forceWrite(F""" {self.arrow_char} """ ) self.write_choice(UpperCAmelCase ) else: forceWrite(F""" {self.choices[index]}""" ) reset_cursor() def A_ ( self : Dict , UpperCAmelCase : Direction , UpperCAmelCase : int = 1 ) -> Any: lowerCamelCase__ : int = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(UpperCAmelCase ) move_cursor(UpperCAmelCase , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP['up'] ) def A_ ( self : Optional[Any] ) -> List[str]: self.move_direction(Direction.UP ) @input.mark(KEYMAP['down'] ) def A_ ( self : List[str] ) -> int: self.move_direction(Direction.DOWN ) @input.mark(KEYMAP['newline'] ) def A_ ( self : str ) -> List[Any]: move_cursor(len(self.choices ) - self.position , 'DOWN' ) return self.position @input.mark(KEYMAP['interrupt'] ) def A_ ( self : Tuple ) -> Any: move_cursor(len(self.choices ) - self.position , 'DOWN' ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(UpperCAmelCase )] for number in range(10 )] ) def A_ ( self : Optional[Any] ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = int(chr(self.current_selection ) ) lowerCamelCase__ : Union[str, Any] = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , UpperCAmelCase ) else: return else: return def A_ ( self : Optional[Any] , UpperCAmelCase : int = 0 ) -> str: if self.prompt: linebreak() forceWrite(self.prompt , '\n' ) if in_colab: forceWrite('Please input a choice index (starting from 0), and press enter' , '\n' ) else: forceWrite('Please select a choice using the arrow or number keys, and selecting with enter' , '\n' ) lowerCamelCase__ : int = default_choice for i in range(len(self.choices ) ): self.print_choice(UpperCAmelCase ) forceWrite('\n' ) move_cursor(len(self.choices ) - self.position , 'UP' ) with cursor.hide(): while True: if in_colab: try: lowerCamelCase__ : Any = int(builtins.input() ) except ValueError: lowerCamelCase__ : Optional[int] = default_choice else: lowerCamelCase__ : Union[str, Any] = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , 'UP' ) clear_line() self.write_choice(UpperCAmelCase , '\n' ) return choice
50
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : int = None _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : List[Any] = { """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : List[str] = { """facebook/nllb-large-en-ro""": 10_24, """facebook/nllb-200-distilled-600M""": 10_24, } # fmt: off _UpperCAmelCase : Optional[int] = ["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] UpperCAmelCase__ = NllbTokenizer UpperCAmelCase__ = [] UpperCAmelCase__ = [] def __init__( self : Tuple , UpperCAmelCase : int=None , UpperCAmelCase : Any=None , UpperCAmelCase : str="<s>" , UpperCAmelCase : Optional[Any]="</s>" , UpperCAmelCase : str="</s>" , UpperCAmelCase : Tuple="<s>" , UpperCAmelCase : Optional[Any]="<unk>" , UpperCAmelCase : List[str]="<pad>" , UpperCAmelCase : Union[str, Any]="<mask>" , UpperCAmelCase : Tuple=None , UpperCAmelCase : int=None , UpperCAmelCase : Dict=None , UpperCAmelCase : Any=False , **UpperCAmelCase : Optional[int] , ) -> Tuple: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase__ : List[Any] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else mask_token lowerCamelCase__ : Union[str, Any] = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase , tokenizer_file=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , unk_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , src_lang=UpperCAmelCase , tgt_lang=UpperCAmelCase , additional_special_tokens=UpperCAmelCase , legacy_behaviour=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = vocab_file lowerCamelCase__ : Dict = False if not self.vocab_file else True lowerCamelCase__ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) lowerCamelCase__ : str = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase__ : int = src_lang if src_lang is not None else 'eng_Latn' lowerCamelCase__ : List[Any] = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def A_ ( self : int ) -> str: return self._src_lang @src_lang.setter def A_ ( self : List[Any] , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Dict = [self.sep_token_id] lowerCamelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A_ ( self : int , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Optional[str] , UpperCAmelCase : Optional[str] , **UpperCAmelCase : List[str] ) -> Dict: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowerCamelCase__ : Optional[int] = src_lang lowerCamelCase__ : Optional[int] = self(UpperCAmelCase , add_special_tokens=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.convert_tokens_to_ids(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = tgt_lang_id return inputs def A_ ( self : Dict , UpperCAmelCase : List[str] , UpperCAmelCase : str = "eng_Latn" , UpperCAmelCase : Optional[List[str]] = None , UpperCAmelCase : str = "fra_Latn" , **UpperCAmelCase : Dict , ) -> BatchEncoding: lowerCamelCase__ : Any = src_lang lowerCamelCase__ : int = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Optional[int]: return self.set_src_lang_special_tokens(self.src_lang ) def A_ ( self : Any ) -> Union[str, Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def A_ ( self : str , UpperCAmelCase : Optional[Any] ) -> None: lowerCamelCase__ : int = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : int = [] lowerCamelCase__ : str = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : int = [self.cur_lang_code] lowerCamelCase__ : Tuple = [self.eos_token_id] lowerCamelCase__ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : str = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : int , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Union[str, Any] = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : Dict = [] lowerCamelCase__ : Union[str, Any] = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : Any = [self.cur_lang_code] lowerCamelCase__ : Optional[Any] = [self.eos_token_id] lowerCamelCase__ : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCamelCase__ : int = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.vocab_file , UpperCAmelCase ) return (out_vocab_file,)
50
1
import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoFeatureExtractor, WavaVecaFeatureExtractor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 lowercase : int = get_tests_dir("""fixtures""") class A__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : Tuple = mock.Mock() a__ : Optional[Any] = 500 a__ : Dict = {} a__ : Tuple = HTTPError a__ : Any = {} # Download this model to make sure it's in the cache. a__ : List[Any] = WavaVecaFeatureExtractor.from_pretrained('hf-internal-testing/tiny-random-wav2vec2') # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('requests.Session.request' , return_value=lowercase) as mock_head: a__ : Optional[Any] = WavaVecaFeatureExtractor.from_pretrained('hf-internal-testing/tiny-random-wav2vec2') # This check we did call the fake head request mock_head.assert_called() def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ : Any = WavaVecaFeatureExtractor.from_pretrained( 'https://huggingface.co/hf-internal-testing/tiny-random-wav2vec2/resolve/main/preprocessor_config.json') @is_staging_test class A__ ( unittest.TestCase ): """simple docstring""" @classmethod def __lowercase ( cls) -> Any: '''simple docstring''' a__ : List[str] = TOKEN HfFolder.save_token(lowercase) @classmethod def __lowercase ( cls) -> Tuple: '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-feature-extractor') except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-feature-extractor-org') except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-feature-extractor') except HTTPError: pass def __lowercase ( self) -> str: '''simple docstring''' a__ : List[Any] = WavaVecaFeatureExtractor.from_pretrained(lowercase) feature_extractor.push_to_hub('test-feature-extractor' , use_auth_token=self._token) a__ : Optional[int] = WavaVecaFeatureExtractor.from_pretrained(F'{USER}/test-feature-extractor') for k, v in feature_extractor.__dict__.items(): self.assertEqual(lowercase , getattr(lowercase , lowercase)) # Reset repo delete_repo(token=self._token , repo_id='test-feature-extractor') # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( lowercase , repo_id='test-feature-extractor' , push_to_hub=lowercase , use_auth_token=self._token) a__ : str = WavaVecaFeatureExtractor.from_pretrained(F'{USER}/test-feature-extractor') for k, v in feature_extractor.__dict__.items(): self.assertEqual(lowercase , getattr(lowercase , lowercase)) def __lowercase ( self) -> Any: '''simple docstring''' a__ : Dict = WavaVecaFeatureExtractor.from_pretrained(lowercase) feature_extractor.push_to_hub('valid_org/test-feature-extractor' , use_auth_token=self._token) a__ : Union[str, Any] = WavaVecaFeatureExtractor.from_pretrained('valid_org/test-feature-extractor') for k, v in feature_extractor.__dict__.items(): self.assertEqual(lowercase , getattr(lowercase , lowercase)) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-feature-extractor') # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( lowercase , repo_id='valid_org/test-feature-extractor-org' , push_to_hub=lowercase , use_auth_token=self._token) a__ : List[Any] = WavaVecaFeatureExtractor.from_pretrained('valid_org/test-feature-extractor-org') for k, v in feature_extractor.__dict__.items(): self.assertEqual(lowercase , getattr(lowercase , lowercase)) def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' CustomFeatureExtractor.register_for_auto_class() a__ : Optional[int] = CustomFeatureExtractor.from_pretrained(lowercase) feature_extractor.push_to_hub('test-dynamic-feature-extractor' , use_auth_token=self._token) # This has added the proper auto_map field to the config self.assertDictEqual( feature_extractor.auto_map , {'AutoFeatureExtractor': 'custom_feature_extraction.CustomFeatureExtractor'} , ) a__ : List[Any] = AutoFeatureExtractor.from_pretrained( F'{USER}/test-dynamic-feature-extractor' , trust_remote_code=lowercase) # Can't make an isinstance check because the new_feature_extractor is from the CustomFeatureExtractor class of a dynamic module self.assertEqual(new_feature_extractor.__class__.__name__ , 'CustomFeatureExtractor')
363
from __future__ import annotations from collections.abc import Callable def A_ ( A__ , A__ , A__ , A__ = 100 , ) -> float: a__ : Dict = x_start a__ : Any = fnc(A__ ) a__ : Optional[int] = 0.0 for _ in range(A__ ): # Approximates small segments of curve as linear and solve # for trapezoidal area a__ : Union[str, Any] = (x_end - x_start) / steps + xa a__ : str = fnc(A__ ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step a__ : Optional[Any] = xa a__ : Optional[int] = fxa return area if __name__ == "__main__": def A_ ( A__ ) -> List[str]: return x**3 + x**2 print("""f(x) = x^3 + x^2""") print("""The area between the curve, x = -5, x = 5 and the x axis is:""") lowercase : Union[str, Any] = 1_0 while i <= 1_0_0_0_0_0: print(F"""with {i} steps: {trapezoidal_area(f, -5, 5, i)}""") i *= 1_0
225
0
"""simple docstring""" import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : Union[str, Any] = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', } _UpperCamelCase : Optional[int] = { 'vocab_file': {'ctrl': 'https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-vocab.json'}, 'merges_file': {'ctrl': 'https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-merges.txt'}, } _UpperCamelCase : List[Any] = { 'ctrl': 2_5_6, } _UpperCamelCase : str = { 'Pregnancy': 1_6_8_6_2_9, 'Christianity': 7_6_7_5, 'Explain': 1_0_6_4_2_3, 'Fitness': 6_3_4_4_0, 'Saving': 6_3_1_6_3, 'Ask': 2_7_1_7_1, 'Ass': 9_5_9_8_5, 'Joke': 1_6_3_5_0_9, 'Questions': 4_5_6_2_2, 'Thoughts': 4_9_6_0_5, 'Retail': 5_2_3_4_2, 'Feminism': 1_6_4_3_3_8, 'Writing': 1_1_9_9_2, 'Atheism': 1_9_2_2_6_3, 'Netflix': 4_8_6_1_6, 'Computing': 3_9_6_3_9, 'Opinion': 4_3_2_1_3, 'Alone': 4_4_9_6_7, 'Funny': 5_8_9_1_7, 'Gaming': 4_0_3_5_8, 'Human': 4_0_8_8, 'India': 1_3_3_1, 'Joker': 7_7_1_3_8, 'Diet': 3_6_2_0_6, 'Legal': 1_1_8_5_9, 'Norman': 4_9_3_9, 'Tip': 7_2_6_8_9, 'Weight': 5_2_3_4_3, 'Movies': 4_6_2_7_3, 'Running': 2_3_4_2_5, 'Science': 2_0_9_0, 'Horror': 3_7_7_9_3, 'Confession': 6_0_5_7_2, 'Finance': 1_2_2_5_0, 'Politics': 1_6_3_6_0, 'Scary': 1_9_1_9_8_5, 'Support': 1_2_6_5_4, 'Technologies': 3_2_5_1_6, 'Teenage': 6_6_1_6_0, 'Event': 3_2_7_6_9, 'Learned': 6_7_4_6_0, 'Notion': 1_8_2_7_7_0, 'Wikipedia': 3_7_5_8_3, 'Books': 6_6_6_5, 'Extract': 7_6_0_5_0, 'Confessions': 1_0_2_7_0_1, 'Conspiracy': 7_5_9_3_2, 'Links': 6_3_6_7_4, 'Narcissus': 1_5_0_4_2_5, 'Relationship': 5_4_7_6_6, 'Relationships': 1_3_4_7_9_6, 'Reviews': 4_1_6_7_1, 'News': 4_2_5_6, 'Translation': 2_6_8_2_0, 'multilingual': 1_2_8_4_0_6, } def _SCREAMING_SNAKE_CASE ( __snake_case : str ): '''simple docstring''' lowercase = set() lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase = char lowercase = set(__snake_case ) return pairs class a ( a_ ): UpperCAmelCase_ : Any =VOCAB_FILES_NAMES UpperCAmelCase_ : Any =PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase_ : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase_ : Tuple =CONTROL_CODES def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase="<unk>" , **_lowerCamelCase ): super().__init__(unk_token=_lowerCamelCase , **_lowerCamelCase ) with open(_lowerCamelCase , encoding='utf-8' ) as vocab_handle: lowercase = json.load(_lowerCamelCase ) lowercase = {v: k for k, v in self.encoder.items()} with open(_lowerCamelCase , encoding='utf-8' ) as merges_handle: lowercase = merges_handle.read().split('\n' )[1:-1] lowercase = [tuple(merge.split() ) for merge in merges] lowercase = dict(zip(_lowerCamelCase , range(len(_lowerCamelCase ) ) ) ) lowercase = {} @property def UpperCamelCase_ ( self ): return len(self.encoder ) def UpperCamelCase_ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def UpperCamelCase_ ( self , _lowerCamelCase ): if token in self.cache: return self.cache[token] lowercase = tuple(_lowerCamelCase ) lowercase = tuple(list(word[:-1] ) + [word[-1] + '</w>'] ) lowercase = get_pairs(_lowerCamelCase ) if not pairs: return token while True: lowercase = min(_lowerCamelCase , key=lambda _lowerCamelCase : self.bpe_ranks.get(_lowerCamelCase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase , lowercase = bigram lowercase = [] lowercase = 0 while i < len(_lowerCamelCase ): try: lowercase = word.index(_lowerCamelCase , _lowerCamelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase = j if word[i] == first and i < len(_lowerCamelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase = tuple(_lowerCamelCase ) lowercase = new_word if len(_lowerCamelCase ) == 1: break else: lowercase = get_pairs(_lowerCamelCase ) lowercase = '@@ '.join(_lowerCamelCase ) lowercase = word[:-4] lowercase = word return word def UpperCamelCase_ ( self , _lowerCamelCase ): lowercase = [] lowercase = re.findall(R'\S+\n?' , _lowerCamelCase ) for token in words: split_tokens.extend(list(self.bpe(_lowerCamelCase ).split(' ' ) ) ) return split_tokens def UpperCamelCase_ ( self , _lowerCamelCase ): return self.encoder.get(_lowerCamelCase , self.encoder.get(self.unk_token ) ) def UpperCamelCase_ ( self , _lowerCamelCase ): return self.decoder.get(_lowerCamelCase , self.unk_token ) def UpperCamelCase_ ( self , _lowerCamelCase ): lowercase = ' '.join(_lowerCamelCase ).replace('@@ ' , '' ).strip() return out_string def UpperCamelCase_ ( self , _lowerCamelCase , _lowerCamelCase = None ): if not os.path.isdir(_lowerCamelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return lowercase = os.path.join( _lowerCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase = os.path.join( _lowerCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(_lowerCamelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCamelCase , ensure_ascii=_lowerCamelCase ) + '\n' ) lowercase = 0 with open(_lowerCamelCase , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCamelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' ' Please check that the tokenizer is not corrupted!' ) lowercase = token_index writer.write(' '.join(_lowerCamelCase ) + '\n' ) index += 1 return vocab_file, merge_file # def decode(self, token_ids, skip_special_tokens=False, clean_up_tokenization_spaces=True): # filtered_tokens = ' '.join(self.convert_ids_to_tokens(token_ids, skip_special_tokens=skip_special_tokens)) # tokens_generated_so_far = re.sub('(@@ )', '', string=filtered_tokens) # tokens_generated_so_far = re.sub('(@@ ?$)', '', string=tokens_generated_so_far) # return ''.join(tokens_generated_so_far)
220
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class a : UpperCAmelCase_ : torch.Tensor # [batch_size x 3] UpperCAmelCase_ : torch.Tensor # [batch_size x 3] UpperCAmelCase_ : torch.Tensor # [batch_size x 3] UpperCAmelCase_ : torch.Tensor # [batch_size x 3] UpperCAmelCase_ : int UpperCAmelCase_ : int UpperCAmelCase_ : float UpperCAmelCase_ : float UpperCAmelCase_ : Tuple[int] def UpperCamelCase_ ( self ): assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def UpperCamelCase_ ( self ): return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def UpperCamelCase_ ( self ): return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def UpperCamelCase_ ( self ): lowercase = torch.arange(self.height * self.width ) lowercase = torch.stack( [ pixel_indices % self.width, torch.div(_lowerCamelCase , self.width , rounding_mode='trunc' ), ] , axis=1 , ) return coords @property def UpperCamelCase_ ( self ): lowercase , *lowercase = self.shape lowercase = int(np.prod(_lowerCamelCase ) ) lowercase = self.get_image_coords() lowercase = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) lowercase = self.get_camera_rays(_lowerCamelCase ) lowercase = rays.view(_lowerCamelCase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def UpperCamelCase_ ( self , _lowerCamelCase ): lowercase , *lowercase , lowercase = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] lowercase = coords.view(_lowerCamelCase , -1 , 2 ) lowercase = self.resolution() lowercase = self.fov() lowercase = (flat.float() / (res - 1)) * 2 - 1 lowercase = fracs * torch.tan(fov / 2 ) lowercase = fracs.view(_lowerCamelCase , -1 , 2 ) lowercase = ( self.z.view(_lowerCamelCase , 1 , 3 ) + self.x.view(_lowerCamelCase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(_lowerCamelCase , 1 , 3 ) * fracs[:, :, 1:] ) lowercase = directions / directions.norm(dim=-1 , keepdim=_lowerCamelCase ) lowercase = torch.stack( [ torch.broadcast_to(self.origin.view(_lowerCamelCase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(_lowerCamelCase , *_lowerCamelCase , 2 , 3 ) def UpperCamelCase_ ( self , _lowerCamelCase , _lowerCamelCase ): assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=_lowerCamelCase , height=_lowerCamelCase , x_fov=self.x_fov , y_fov=self.y_fov , ) def _SCREAMING_SNAKE_CASE ( __snake_case : int ): '''simple docstring''' lowercase = [] lowercase = [] lowercase = [] lowercase = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): lowercase = np.array([np.sin(__snake_case ), np.cos(__snake_case ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) lowercase = -z * 4 lowercase = np.array([np.cos(__snake_case ), -np.sin(__snake_case ), 0.0] ) lowercase = np.cross(__snake_case , __snake_case ) origins.append(__snake_case ) xs.append(__snake_case ) ys.append(__snake_case ) zs.append(__snake_case ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(__snake_case , axis=0 ) ).float() , x=torch.from_numpy(np.stack(__snake_case , axis=0 ) ).float() , y=torch.from_numpy(np.stack(__snake_case , axis=0 ) ).float() , z=torch.from_numpy(np.stack(__snake_case , axis=0 ) ).float() , width=__snake_case , height=__snake_case , x_fov=0.7 , y_fov=0.7 , shape=(1, len(__snake_case )) , )
220
1
'''simple docstring''' from io import BytesIO from typing import List, Union import requests from ..utils import add_end_docstrings, is_decord_available, is_torch_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_decord_available(): import numpy as np from decord import VideoReader if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING _lowerCamelCase = logging.get_logger(__name__) @add_end_docstrings(__SCREAMING_SNAKE_CASE) class _snake_case (__SCREAMING_SNAKE_CASE): def __init__( self ,*_snake_case ,**_snake_case ): super().__init__(*_snake_case ,**_snake_case ) requires_backends(self ,"decord" ) self.check_model_type(_snake_case ) def UpperCamelCase__ ( self ,_snake_case=None ,_snake_case=None ,_snake_case=None ): UpperCAmelCase_ : Optional[Any] = {} if frame_sampling_rate is not None: UpperCAmelCase_ : Optional[int] = frame_sampling_rate if num_frames is not None: UpperCAmelCase_ : Union[str, Any] = num_frames UpperCAmelCase_ : int = {} if top_k is not None: UpperCAmelCase_ : Dict = top_k return preprocess_params, {}, postprocess_params def __call__( self ,_snake_case ,**_snake_case ): return super().__call__(_snake_case ,**_snake_case ) def UpperCamelCase__ ( self ,_snake_case ,_snake_case=None ,_snake_case=1 ): if num_frames is None: UpperCAmelCase_ : Any = self.model.config.num_frames if video.startswith("http://" ) or video.startswith("https://" ): UpperCAmelCase_ : Optional[int] = BytesIO(requests.get(_snake_case ).content ) UpperCAmelCase_ : Tuple = VideoReader(_snake_case ) videoreader.seek(0 ) UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : Optional[Any] = num_frames * frame_sampling_rate - 1 UpperCAmelCase_ : Dict = np.linspace(_snake_case ,_snake_case ,num=_snake_case ,dtype=np.intaa ) UpperCAmelCase_ : int = videoreader.get_batch(_snake_case ).asnumpy() UpperCAmelCase_ : int = list(_snake_case ) UpperCAmelCase_ : Any = self.image_processor(_snake_case ,return_tensors=self.framework ) return model_inputs def UpperCamelCase__ ( self ,_snake_case ): UpperCAmelCase_ : int = self.model(**_snake_case ) return model_outputs def UpperCamelCase__ ( self ,_snake_case ,_snake_case=5 ): if top_k > self.model.config.num_labels: UpperCAmelCase_ : List[str] = self.model.config.num_labels if self.framework == "pt": UpperCAmelCase_ : str = model_outputs.logits.softmax(-1 )[0] UpperCAmelCase_ : Optional[Any] = probs.topk(_snake_case ) else: raise ValueError(f'''Unsupported framework: {self.framework}''' ) UpperCAmelCase_ : int = scores.tolist() UpperCAmelCase_ : Tuple = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(_snake_case ,_snake_case )]
365
'''simple docstring''' from ..utils import DummyObject, requires_backends class _snake_case (metaclass=__SCREAMING_SNAKE_CASE): __A : Union[str, Any] =["torch", "torchsde"] def __init__( self ,*_snake_case ,**_snake_case ): requires_backends(self ,["torch", "torchsde"] ) @classmethod def UpperCamelCase__ ( cls ,*_snake_case ,**_snake_case ): requires_backends(cls ,["torch", "torchsde"] ) @classmethod def UpperCamelCase__ ( cls ,*_snake_case ,**_snake_case ): requires_backends(cls ,["torch", "torchsde"] )
67
0
'''simple docstring''' import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] ) -> Any: UpperCAmelCase : int = checkpoint UpperCAmelCase : Any = {} UpperCAmelCase : Optional[Any] = vae_state_dict['''encoder.conv_in.weight'''] UpperCAmelCase : Dict = vae_state_dict['''encoder.conv_in.bias'''] UpperCAmelCase : Tuple = vae_state_dict['''encoder.conv_out.weight'''] UpperCAmelCase : Optional[int] = vae_state_dict['''encoder.conv_out.bias'''] UpperCAmelCase : int = vae_state_dict['''encoder.norm_out.weight'''] UpperCAmelCase : Union[str, Any] = vae_state_dict['''encoder.norm_out.bias'''] UpperCAmelCase : List[Any] = vae_state_dict['''decoder.conv_in.weight'''] UpperCAmelCase : List[Any] = vae_state_dict['''decoder.conv_in.bias'''] UpperCAmelCase : str = vae_state_dict['''decoder.conv_out.weight'''] UpperCAmelCase : Any = vae_state_dict['''decoder.conv_out.bias'''] UpperCAmelCase : List[Any] = vae_state_dict['''decoder.norm_out.weight'''] UpperCAmelCase : Any = vae_state_dict['''decoder.norm_out.bias'''] UpperCAmelCase : int = vae_state_dict['''quant_conv.weight'''] UpperCAmelCase : Union[str, Any] = vae_state_dict['''quant_conv.bias'''] UpperCAmelCase : Any = vae_state_dict['''post_quant_conv.weight'''] UpperCAmelCase : List[Any] = vae_state_dict['''post_quant_conv.bias'''] # Retrieves the keys for the encoder down blocks only UpperCAmelCase : Optional[Any] = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''encoder.down''' in layer} ) UpperCAmelCase : Optional[Any] = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(_lowerCAmelCase ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase : Any = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''decoder.up''' in layer} ) UpperCAmelCase : List[str] = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(_lowerCAmelCase ) } for i in range(_lowerCAmelCase ): UpperCAmelCase : Dict = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase : Optional[int] = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase : List[Any] = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase : Tuple = renew_vae_resnet_paths(_lowerCAmelCase ) UpperCAmelCase : str = {'''old''': f"""down.{i}.block""", '''new''': f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) UpperCAmelCase : List[Any] = [key for key in vae_state_dict if '''encoder.mid.block''' in key] UpperCAmelCase : Union[str, Any] = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase : Dict = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase : Tuple = renew_vae_resnet_paths(_lowerCAmelCase ) UpperCAmelCase : Dict = {'''old''': f"""mid.block_{i}""", '''new''': f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) UpperCAmelCase : Any = [key for key in vae_state_dict if '''encoder.mid.attn''' in key] UpperCAmelCase : List[Any] = renew_vae_attention_paths(_lowerCAmelCase ) UpperCAmelCase : Dict = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) conv_attn_to_linear(_lowerCAmelCase ) for i in range(_lowerCAmelCase ): UpperCAmelCase : List[Any] = num_up_blocks - 1 - i UpperCAmelCase : Dict = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase : List[Any] = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase : List[Any] = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase : List[Any] = renew_vae_resnet_paths(_lowerCAmelCase ) UpperCAmelCase : int = {'''old''': f"""up.{block_id}.block""", '''new''': f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) UpperCAmelCase : Dict = [key for key in vae_state_dict if '''decoder.mid.block''' in key] UpperCAmelCase : Tuple = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase : List[str] = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase : List[str] = renew_vae_resnet_paths(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = {'''old''': f"""mid.block_{i}""", '''new''': f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) UpperCAmelCase : int = [key for key in vae_state_dict if '''decoder.mid.attn''' in key] UpperCAmelCase : Dict = renew_vae_attention_paths(_lowerCAmelCase ) UpperCAmelCase : Tuple = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) conv_attn_to_linear(_lowerCAmelCase ) return new_checkpoint def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str , ) -> Tuple: # Only support V1 UpperCAmelCase : str = requests.get( ''' https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml''' ) UpperCAmelCase : Optional[Any] = io.BytesIO(r.content ) UpperCAmelCase : int = OmegaConf.load(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = 512 UpperCAmelCase : List[str] = '''cuda''' if torch.cuda.is_available() else '''cpu''' if checkpoint_path.endswith('''safetensors''' ): from safetensors import safe_open UpperCAmelCase : int = {} with safe_open(_lowerCAmelCase , framework='''pt''' , device='''cpu''' ) as f: for key in f.keys(): UpperCAmelCase : str = f.get_tensor(_lowerCAmelCase ) else: UpperCAmelCase : Dict = torch.load(_lowerCAmelCase , map_location=_lowerCAmelCase )['''state_dict'''] # Convert the VAE model. UpperCAmelCase : List[Any] = create_vae_diffusers_config(_lowerCAmelCase , image_size=_lowerCAmelCase ) UpperCAmelCase : Tuple = custom_convert_ldm_vae_checkpoint(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Tuple = AutoencoderKL(**_lowerCAmelCase ) vae.load_state_dict(_lowerCAmelCase ) vae.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() parser.add_argument("--vae_pt_path", default=None, type=str, required=True, help="Path to the VAE.pt to convert.") parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the VAE.pt to convert.") UpperCamelCase__: str = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
23
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class _a ( unittest.TestCase): def UpperCAmelCase__( self : Tuple )-> int: lowerCAmelCase__ : List[Any] = tempfile.mkdtemp() # fmt: off lowerCAmelCase__ : Tuple = ['''''', '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on lowerCAmelCase__ : Any = dict(zip(_SCREAMING_SNAKE_CASE , range(len(_SCREAMING_SNAKE_CASE ) ) ) ) lowerCAmelCase__ : Dict = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>''', ''''''] lowerCAmelCase__ : Any = {'''unk_token''': '''<unk>'''} lowerCAmelCase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase__ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(_SCREAMING_SNAKE_CASE ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(_SCREAMING_SNAKE_CASE ) ) lowerCAmelCase__ : Union[str, Any] = { '''do_resize''': True, '''size''': 20, '''do_center_crop''': True, '''crop_size''': 18, '''do_normalize''': True, '''image_mean''': [0.4814_5466, 0.457_8275, 0.4082_1073], '''image_std''': [0.2686_2954, 0.2613_0258, 0.2757_7711], } lowerCAmelCase__ : Tuple = os.path.join(self.tmpdirname , _SCREAMING_SNAKE_CASE ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : int , **_SCREAMING_SNAKE_CASE : Any )-> Any: return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token='''!''' , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Tuple , **_SCREAMING_SNAKE_CASE : List[str] )-> List[str]: return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token='''!''' , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : int , **_SCREAMING_SNAKE_CASE : Optional[Any] )-> Optional[Any]: return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : str )-> Optional[Any]: shutil.rmtree(self.tmpdirname ) def UpperCAmelCase__( self : Dict )-> Any: lowerCAmelCase__ : Tuple = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCAmelCase__ : Optional[Any] = [Image.fromarray(np.moveaxis(_SCREAMING_SNAKE_CASE , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCAmelCase__( self : int )-> List[str]: lowerCAmelCase__ : Optional[Any] = self.get_tokenizer() lowerCAmelCase__ : List[Any] = self.get_rust_tokenizer() lowerCAmelCase__ : Tuple = self.get_image_processor() lowerCAmelCase__ : Tuple = OwlViTProcessor(tokenizer=_SCREAMING_SNAKE_CASE , image_processor=_SCREAMING_SNAKE_CASE ) processor_slow.save_pretrained(self.tmpdirname ) lowerCAmelCase__ : Optional[Any] = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Optional[int] = OwlViTProcessor(tokenizer=_SCREAMING_SNAKE_CASE , image_processor=_SCREAMING_SNAKE_CASE ) processor_fast.save_pretrained(self.tmpdirname ) lowerCAmelCase__ : Optional[int] = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _SCREAMING_SNAKE_CASE ) self.assertIsInstance(processor_fast.tokenizer , _SCREAMING_SNAKE_CASE ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _SCREAMING_SNAKE_CASE ) self.assertIsInstance(processor_fast.image_processor , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Optional[int] )-> Optional[int]: lowerCAmelCase__ : Tuple = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCAmelCase__ : List[Any] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) lowerCAmelCase__ : Union[str, Any] = self.get_image_processor(do_normalize=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Tuple = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=_SCREAMING_SNAKE_CASE ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _SCREAMING_SNAKE_CASE ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Dict )-> Union[str, Any]: lowerCAmelCase__ : List[Any] = self.get_image_processor() lowerCAmelCase__ : str = self.get_tokenizer() lowerCAmelCase__ : Optional[Any] = OwlViTProcessor(tokenizer=_SCREAMING_SNAKE_CASE , image_processor=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = self.prepare_image_inputs() lowerCAmelCase__ : List[Any] = image_processor(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ) lowerCAmelCase__ : Optional[int] = processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__( self : Any )-> Tuple: lowerCAmelCase__ : Any = self.get_image_processor() lowerCAmelCase__ : str = self.get_tokenizer() lowerCAmelCase__ : List[str] = OwlViTProcessor(tokenizer=_SCREAMING_SNAKE_CASE , image_processor=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = '''lower newer''' lowerCAmelCase__ : List[Any] = processor(text=_SCREAMING_SNAKE_CASE , return_tensors='''np''' ) lowerCAmelCase__ : Optional[Any] = tokenizer(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def UpperCAmelCase__( self : Union[str, Any] )-> Tuple: lowerCAmelCase__ : Any = self.get_image_processor() lowerCAmelCase__ : Any = self.get_tokenizer() lowerCAmelCase__ : List[str] = OwlViTProcessor(tokenizer=_SCREAMING_SNAKE_CASE , image_processor=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Optional[int] = '''lower newer''' lowerCAmelCase__ : str = self.prepare_image_inputs() lowerCAmelCase__ : Tuple = processor(text=_SCREAMING_SNAKE_CASE , images=_SCREAMING_SNAKE_CASE ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(_SCREAMING_SNAKE_CASE ): processor() def UpperCAmelCase__( self : Dict )-> Union[str, Any]: lowerCAmelCase__ : List[Any] = '''google/owlvit-base-patch32''' lowerCAmelCase__ : str = OwlViTProcessor.from_pretrained(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Union[str, Any] = ['''cat''', '''nasa badge'''] lowerCAmelCase__ : Optional[int] = processor(text=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Tuple = 16 self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask'''] ) self.assertEqual(inputs['''input_ids'''].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(_SCREAMING_SNAKE_CASE ): processor() def UpperCAmelCase__( self : Any )-> Tuple: lowerCAmelCase__ : str = '''google/owlvit-base-patch32''' lowerCAmelCase__ : int = OwlViTProcessor.from_pretrained(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = [['''cat''', '''nasa badge'''], ['''person''']] lowerCAmelCase__ : Union[str, Any] = processor(text=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Tuple = 16 lowerCAmelCase__ : Optional[int] = len(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = max([len(_SCREAMING_SNAKE_CASE ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask'''] ) self.assertEqual(inputs['''input_ids'''].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(_SCREAMING_SNAKE_CASE ): processor() def UpperCAmelCase__( self : List[str] )-> str: lowerCAmelCase__ : Dict = '''google/owlvit-base-patch32''' lowerCAmelCase__ : Union[str, Any] = OwlViTProcessor.from_pretrained(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Dict = ['''cat''', '''nasa badge'''] lowerCAmelCase__ : Union[str, Any] = processor(text=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Optional[int] = 16 lowerCAmelCase__ : Optional[int] = inputs['''input_ids'''] lowerCAmelCase__ : str = [ [4_9406, 2368, 4_9407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9406, 6841, 1_1301, 4_9407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''attention_mask'''] ) self.assertEqual(inputs['''input_ids'''].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def UpperCAmelCase__( self : Optional[Any] )-> List[str]: lowerCAmelCase__ : Any = self.get_image_processor() lowerCAmelCase__ : Any = self.get_tokenizer() lowerCAmelCase__ : str = OwlViTProcessor(tokenizer=_SCREAMING_SNAKE_CASE , image_processor=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = self.prepare_image_inputs() lowerCAmelCase__ : Union[str, Any] = self.prepare_image_inputs() lowerCAmelCase__ : List[Any] = processor(images=_SCREAMING_SNAKE_CASE , query_images=_SCREAMING_SNAKE_CASE ) self.assertListEqual(list(inputs.keys() ) , ['''query_pixel_values''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(_SCREAMING_SNAKE_CASE ): processor() def UpperCAmelCase__( self : int )-> Dict: lowerCAmelCase__ : int = self.get_image_processor() lowerCAmelCase__ : Any = self.get_tokenizer() lowerCAmelCase__ : Tuple = OwlViTProcessor(tokenizer=_SCREAMING_SNAKE_CASE , image_processor=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCAmelCase__ : Optional[Any] = processor.batch_decode(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = tokenizer.batch_decode(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )
131
0
import json import logging import os import socket import git import numpy as np import torch logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - PID: %(process)d - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO, ) __snake_case = logging.getLogger(__name__) def lowerCAmelCase_ ( __lowerCAmelCase )-> Optional[int]: '''simple docstring''' UpperCAmelCase : Optional[Any] =git.Repo(search_parent_directories=__lowerCAmelCase ) UpperCAmelCase : str ={ '''repo_id''': str(__lowerCAmelCase ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), } with open(os.path.join(__lowerCAmelCase , '''git_log.json''' ) , '''w''' ) as f: json.dump(__lowerCAmelCase , __lowerCAmelCase , indent=4 ) def lowerCAmelCase_ ( __lowerCAmelCase )-> int: '''simple docstring''' if params.n_gpu <= 0: UpperCAmelCase : List[Any] =0 UpperCAmelCase : str =-1 UpperCAmelCase : List[Any] =True UpperCAmelCase : Optional[Any] =False return assert torch.cuda.is_available() logger.info('''Initializing GPUs''' ) if params.n_gpu > 1: assert params.local_rank != -1 UpperCAmelCase : int =int(os.environ['''WORLD_SIZE'''] ) UpperCAmelCase : Dict =int(os.environ['''N_GPU_NODE'''] ) UpperCAmelCase : Dict =int(os.environ['''RANK'''] ) # number of nodes / node ID UpperCAmelCase : Optional[Any] =params.world_size // params.n_gpu_per_node UpperCAmelCase : str =params.global_rank // params.n_gpu_per_node UpperCAmelCase : Any =True assert params.n_nodes == int(os.environ['''N_NODES'''] ) assert params.node_id == int(os.environ['''NODE_RANK'''] ) # local job (single GPU) else: assert params.local_rank == -1 UpperCAmelCase : Union[str, Any] =1 UpperCAmelCase : Union[str, Any] =0 UpperCAmelCase : List[Any] =0 UpperCAmelCase : List[Any] =0 UpperCAmelCase : Union[str, Any] =1 UpperCAmelCase : Dict =1 UpperCAmelCase : Union[str, Any] =False # sanity checks assert params.n_nodes >= 1 assert 0 <= params.node_id < params.n_nodes assert 0 <= params.local_rank <= params.global_rank < params.world_size assert params.world_size == params.n_nodes * params.n_gpu_per_node # define whether this is the master process / if we are in multi-node distributed mode UpperCAmelCase : Tuple =params.node_id == 0 and params.local_rank == 0 UpperCAmelCase : Optional[Any] =params.n_nodes > 1 # summary UpperCAmelCase : Any =f'''--- Global rank: {params.global_rank} - ''' logger.info(PREFIX + '''Number of nodes: %i''' % params.n_nodes ) logger.info(PREFIX + '''Node ID : %i''' % params.node_id ) logger.info(PREFIX + '''Local rank : %i''' % params.local_rank ) logger.info(PREFIX + '''World size : %i''' % params.world_size ) logger.info(PREFIX + '''GPUs per node : %i''' % params.n_gpu_per_node ) logger.info(PREFIX + '''Master : %s''' % str(params.is_master ) ) logger.info(PREFIX + '''Multi-node : %s''' % str(params.multi_node ) ) logger.info(PREFIX + '''Multi-GPU : %s''' % str(params.multi_gpu ) ) logger.info(PREFIX + '''Hostname : %s''' % socket.gethostname() ) # set GPU device torch.cuda.set_device(params.local_rank ) # initialize multi-GPU if params.multi_gpu: logger.info('''Initializing PyTorch distributed''' ) torch.distributed.init_process_group( init_method='''env://''' , backend='''nccl''' , ) def lowerCAmelCase_ ( __lowerCAmelCase )-> Any: '''simple docstring''' np.random.seed(args.seed ) torch.manual_seed(args.seed ) if args.n_gpu > 0: torch.cuda.manual_seed_all(args.seed )
363
import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( lowerCamelCase__ , unittest.TestCase ): __lowerCamelCase : List[Any] = KandinskyVaaControlnetPipeline __lowerCamelCase : int = ["""image_embeds""", """negative_image_embeds""", """hint"""] __lowerCamelCase : Optional[int] = ["""image_embeds""", """negative_image_embeds""", """hint"""] __lowerCamelCase : Optional[Any] = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] __lowerCamelCase : Dict = False @property def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' return 32 @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return 32 @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return self.time_input_dim @property def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' return self.time_input_dim * 4 @property def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' return 100 @property def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase : Any ={ '''in_channels''': 8, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image_hint''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } UpperCAmelCase : List[Any] =UNetaDConditionModel(**snake_case__ ) return model @property def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase : Any =VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : List[str] =self.dummy_unet UpperCAmelCase : Tuple =self.dummy_movq UpperCAmelCase : Union[str, Any] =DDIMScheduler( num_train_timesteps=1000 , beta_schedule='''linear''' , beta_start=0.0_0085 , beta_end=0.012 , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , steps_offset=1 , prediction_type='''epsilon''' , thresholding=snake_case__ , ) UpperCAmelCase : Tuple ={ '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def UpperCAmelCase__ ( self , snake_case__ , snake_case__=0 ) -> Any: '''simple docstring''' UpperCAmelCase : str =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) UpperCAmelCase : Tuple =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( snake_case__ ) # create hint UpperCAmelCase : Tuple =floats_tensor((1, 3, 64, 64) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith('''mps''' ): UpperCAmelCase : Optional[int] =torch.manual_seed(snake_case__ ) else: UpperCAmelCase : int =torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase : List[str] ={ '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''hint''': hint, '''generator''': generator, '''height''': 64, '''width''': 64, '''guidance_scale''': 4.0, '''num_inference_steps''': 2, '''output_type''': '''np''', } return inputs def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : List[Any] ='''cpu''' UpperCAmelCase : List[Any] =self.get_dummy_components() UpperCAmelCase : Tuple =self.pipeline_class(**snake_case__ ) UpperCAmelCase : Tuple =pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase : Optional[int] =pipe(**self.get_dummy_inputs(snake_case__ ) ) UpperCAmelCase : str =output.images UpperCAmelCase : List[str] =pipe( **self.get_dummy_inputs(snake_case__ ) , return_dict=snake_case__ , )[0] UpperCAmelCase : Union[str, Any] =image[0, -3:, -3:, -1] UpperCAmelCase : List[str] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : Union[str, Any] =np.array( [0.695_9826, 0.86_8279, 0.755_8092, 0.6876_9467, 0.8580_5804, 0.6597_7496, 0.4488_5302, 0.595_9111, 0.425_1595] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : Union[str, Any] =load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy''' ) UpperCAmelCase : Tuple =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/hint_image_cat.png''' ) UpperCAmelCase : int =torch.from_numpy(np.array(snake_case__ ) ).float() / 255.0 UpperCAmelCase : List[str] =hint.permute(2 , 0 , 1 ).unsqueeze(0 ) UpperCAmelCase : Dict =KandinskyVaaPriorPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(snake_case__ ) UpperCAmelCase : int =KandinskyVaaControlnetPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-controlnet-depth''' , torch_dtype=torch.floataa ) UpperCAmelCase : str =pipeline.to(snake_case__ ) pipeline.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase : int ='''A robot, 4k photo''' UpperCAmelCase : int =torch.Generator(device='''cuda''' ).manual_seed(0 ) UpperCAmelCase , UpperCAmelCase : List[str] =pipe_prior( snake_case__ , generator=snake_case__ , num_inference_steps=5 , negative_prompt='''''' , ).to_tuple() UpperCAmelCase : List[str] =torch.Generator(device='''cuda''' ).manual_seed(0 ) UpperCAmelCase : Dict =pipeline( image_embeds=snake_case__ , negative_image_embeds=snake_case__ , hint=snake_case__ , generator=snake_case__ , num_inference_steps=100 , output_type='''np''' , ) UpperCAmelCase : List[Any] =output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(snake_case__ , snake_case__ )
78
0
import math def UpperCamelCase__( UpperCamelCase__ : list , UpperCamelCase__ : int )->int: A__ = len(UpperCamelCase__ ) A__ = int(math.floor(math.sqrt(UpperCamelCase__ ) ) ) A__ = 0 while arr[min(UpperCamelCase__ , UpperCamelCase__ ) - 1] < x: A__ = step step += int(math.floor(math.sqrt(UpperCamelCase__ ) ) ) if prev >= n: return -1 while arr[prev] < x: A__ = prev + 1 if prev == min(UpperCamelCase__ , UpperCamelCase__ ): return -1 if arr[prev] == x: return prev return -1 if __name__ == "__main__": a__: Optional[int] = input('Enter numbers separated by a comma:\n').strip() a__: Dict = [int(item) for item in user_input.split(',')] a__: Optional[int] = int(input('Enter the number to be searched:\n')) a__: Optional[Any] = jump_search(arr, x) if res == -1: print('Number not found!') else: print(F"Number {x} is at index {res}")
193
import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors a__: List[str] = logging.getLogger(__name__) class SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = '''sequence-classification''' def __init__( self,__lowerCamelCase ): if type(__lowerCamelCase ) == dict: A__ = Namespace(**__lowerCamelCase ) A__ = glue_output_modes[hparams.task] A__ = glue_tasks_num_labels[hparams.task] super().__init__(__lowerCamelCase,__lowerCamelCase,self.mode ) def UpperCamelCase ( self,**__lowerCamelCase ): return self.model(**__lowerCamelCase ) def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase ): A__ = {'''input_ids''': batch[0], '''attention_mask''': batch[1], '''labels''': batch[3]} if self.config.model_type not in ["distilbert", "bart"]: A__ = batch[2] if self.config.model_type in ['''bert''', '''xlnet''', '''albert'''] else None A__ = self(**__lowerCamelCase ) A__ = outputs[0] A__ = self.trainer.lr_schedulers[0]['''scheduler'''] A__ = {'''loss''': loss, '''rate''': lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def UpperCamelCase ( self ): A__ = self.hparams A__ = processors[args.task]() A__ = processor.get_labels() for mode in ["train", "dev"]: A__ = self._feature_file(__lowerCamelCase ) if os.path.exists(__lowerCamelCase ) and not args.overwrite_cache: logger.info('''Loading features from cached file %s''',__lowerCamelCase ) else: logger.info('''Creating features from dataset file at %s''',args.data_dir ) A__ = ( processor.get_dev_examples(args.data_dir ) if mode == '''dev''' else processor.get_train_examples(args.data_dir ) ) A__ = convert_examples_to_features( __lowerCamelCase,self.tokenizer,max_length=args.max_seq_length,label_list=self.labels,output_mode=args.glue_output_mode,) logger.info('''Saving features into cached file %s''',__lowerCamelCase ) torch.save(__lowerCamelCase,__lowerCamelCase ) def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase,__lowerCamelCase = False ): A__ = '''dev''' if mode == '''test''' else mode A__ = self._feature_file(__lowerCamelCase ) logger.info('''Loading features from cached file %s''',__lowerCamelCase ) A__ = torch.load(__lowerCamelCase ) A__ = torch.tensor([f.input_ids for f in features],dtype=torch.long ) A__ = torch.tensor([f.attention_mask for f in features],dtype=torch.long ) A__ = torch.tensor([f.token_type_ids for f in features],dtype=torch.long ) if self.hparams.glue_output_mode == "classification": A__ = torch.tensor([f.label for f in features],dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": A__ = torch.tensor([f.label for f in features],dtype=torch.float ) return DataLoader( TensorDataset(__lowerCamelCase,__lowerCamelCase,__lowerCamelCase,__lowerCamelCase ),batch_size=__lowerCamelCase,shuffle=__lowerCamelCase,) def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase ): A__ = {'''input_ids''': batch[0], '''attention_mask''': batch[1], '''labels''': batch[3]} if self.config.model_type not in ["distilbert", "bart"]: A__ = batch[2] if self.config.model_type in ['''bert''', '''xlnet''', '''albert'''] else None A__ = self(**__lowerCamelCase ) A__ , A__ = outputs[:2] A__ = logits.detach().cpu().numpy() A__ = inputs['''labels'''].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def UpperCamelCase ( self,__lowerCamelCase ): A__ = torch.stack([x['''val_loss'''] for x in outputs] ).mean().detach().cpu().item() A__ = np.concatenate([x['''pred'''] for x in outputs],axis=0 ) if self.hparams.glue_output_mode == "classification": A__ = np.argmax(__lowerCamelCase,axis=1 ) elif self.hparams.glue_output_mode == "regression": A__ = np.squeeze(__lowerCamelCase ) A__ = np.concatenate([x['''target'''] for x in outputs],axis=0 ) A__ = [[] for _ in range(out_label_ids.shape[0] )] A__ = [[] for _ in range(out_label_ids.shape[0] )] A__ = {**{'''val_loss''': val_loss_mean}, **compute_metrics(self.hparams.task,__lowerCamelCase,__lowerCamelCase )} A__ = dict(results.items() ) A__ = results return ret, preds_list, out_label_list def UpperCamelCase ( self,__lowerCamelCase ): A__ , A__ , A__ = self._eval_end(__lowerCamelCase ) A__ = ret['''log'''] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def UpperCamelCase ( self,__lowerCamelCase ): A__ , A__ , A__ = self._eval_end(__lowerCamelCase ) A__ = ret['''log'''] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def UpperCamelCase ( __lowerCamelCase,__lowerCamelCase ): BaseTransformer.add_model_specific_args(__lowerCamelCase,__lowerCamelCase ) parser.add_argument( '''--max_seq_length''',default=128,type=__lowerCamelCase,help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ),) parser.add_argument( '''--task''',default='''''',type=__lowerCamelCase,required=__lowerCamelCase,help='''The GLUE task to run''',) parser.add_argument( '''--gpus''',default=0,type=__lowerCamelCase,help='''The number of GPUs allocated for this, it is by default 0 meaning none''',) parser.add_argument( '''--overwrite_cache''',action='''store_true''',help='''Overwrite the cached training and evaluation sets''' ) return parser def UpperCamelCase__( )->Any: A__ = argparse.ArgumentParser() add_generic_args(UpperCamelCase__ , os.getcwd() ) A__ = GLUETransformer.add_model_specific_args(UpperCamelCase__ , os.getcwd() ) A__ = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: A__ = os.path.join( '''./results''' , f"{args.task}_{time.strftime('%Y%m%d_%H%M%S' )}" , ) os.makedirs(args.output_dir ) A__ = GLUETransformer(UpperCamelCase__ ) A__ = generic_train(UpperCamelCase__ , UpperCamelCase__ ) # Optionally, predict on dev set and write to output_dir if args.do_predict: A__ = sorted(glob.glob(os.path.join(args.output_dir , '''checkpoint-epoch=*.ckpt''' ) , recursive=UpperCamelCase__ ) ) A__ = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(UpperCamelCase__ ) if __name__ == "__main__": main()
193
1
"""simple docstring""" import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def __a ( _SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Any = [ 'decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( _SCREAMING_SNAKE_CASE ) ->Dict: a__ , a__: Tuple = emb.weight.shape a__: Optional[int] = nn.Linear(lowerCAmelCase__ , lowerCAmelCase__ , bias=lowerCAmelCase__ ) a__: Optional[Any] = emb.weight.data return lin_layer def __a ( _SCREAMING_SNAKE_CASE ) ->List[str]: a__: List[str] = torch.load(lowerCAmelCase__ , map_location='cpu' ) a__: int = Namespace(**checkpoint['cfg']['model'] ) a__: List[str] = checkpoint['model'] remove_ignore_keys_(lowerCAmelCase__ ) a__: Dict = state_dict['decoder.embed_tokens.weight'].shape[0] a__: Union[str, Any] = {key.replace('decoder' , 'model' ): val for key, val in state_dict.items()} a__: List[Any] = XGLMConfig( vocab_size=lowerCAmelCase__ , max_position_embeddings=args.max_target_positions , num_layers=args.decoder_layers , attention_heads=args.decoder_attention_heads , ffn_dim=args.decoder_ffn_embed_dim , d_model=args.decoder_embed_dim , layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='gelu' , scale_embedding=not args.no_scale_embedding , tie_word_embeddings=args.share_decoder_input_output_embed , ) a__: Union[str, Any] = XGLMForCausalLM(lowerCAmelCase__ ) a__: Tuple = model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) print(lowerCAmelCase__ ) a__: str = make_linear_from_emb(model.model.embed_tokens ) return model if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('fairseq_path', type=str, help='path to a model.pt on local filesystem.') parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') lowercase__ = parser.parse_args() lowercase__ = convert_fairseq_xglm_checkpoint_from_disk(args.fairseq_path) model.save_pretrained(args.pytorch_dump_folder_path)
359
"""simple docstring""" import argparse import math import traceback import dateutil.parser as date_parser import requests def __a ( _SCREAMING_SNAKE_CASE ) ->Tuple: a__: Tuple = {} a__: Tuple = job['started_at'] a__: int = job['completed_at'] a__: Any = date_parser.parse(_SCREAMING_SNAKE_CASE ) a__: Tuple = date_parser.parse(_SCREAMING_SNAKE_CASE ) a__: str = round((end_datetime - start_datetime).total_seconds() / 60.0 ) a__: Any = start a__: Dict = end a__: Optional[int] = duration_in_min return job_info def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: a__: Tuple = None if token is not None: a__: List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: int = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: str = {} try: job_time.update({job['name']: extract_time_from_single_job(_SCREAMING_SNAKE_CASE ) for job in result['jobs']} ) a__: Dict = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: str = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() job_time.update({job['name']: extract_time_from_single_job(_SCREAMING_SNAKE_CASE ) for job in result['jobs']} ) return job_time except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') lowercase__ = parser.parse_args() lowercase__ = get_job_time(args.workflow_run_id) lowercase__ = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(f"{k}: {v['duration']}")
203
0
"""simple docstring""" import argparse import os import re _UpperCAmelCase = """src/transformers/models/auto""" # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict _UpperCAmelCase = re.compile(r"""[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict""") # re pattern that matches identifiers in mappings _UpperCAmelCase = re.compile(r"""\s*\(\s*\"(\S[^\"]+)\"""") def __magic_name__ ( lowercase , lowercase = False ): with open(lowercase , """r""" , encoding="""utf-8""" ) as f: SCREAMING_SNAKE_CASE_: Any =f.read() SCREAMING_SNAKE_CASE_: int =content.split("""\n""" ) SCREAMING_SNAKE_CASE_: Optional[Any] =[] SCREAMING_SNAKE_CASE_: Optional[Any] =0 while line_idx < len(lowercase ): if _re_intro_mapping.search(lines[line_idx] ) is not None: SCREAMING_SNAKE_CASE_: Dict =len(re.search(R"""^(\s*)\S""" , lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(""" """ * indent + """(""" ): new_lines.append(lines[line_idx] ) line_idx += 1 SCREAMING_SNAKE_CASE_: Any =[] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": SCREAMING_SNAKE_CASE_: int =line_idx while not lines[line_idx].startswith(""" """ * indent + """)""" ): line_idx += 1 blocks.append("""\n""".join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers SCREAMING_SNAKE_CASE_: Optional[int] =sorted(lowercase , key=lambda lowercase : _re_identifier.search(lowercase ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(lowercase , """w""" , encoding="""utf-8""" ) as f: f.write("""\n""".join(lowercase ) ) elif "\n".join(lowercase ) != content: return True def __magic_name__ ( lowercase = False ): SCREAMING_SNAKE_CASE_: Any =[os.path.join(lowercase , lowercase ) for f in os.listdir(lowercase ) if f.endswith(""".py""" )] SCREAMING_SNAKE_CASE_: Dict =[sort_auto_mapping(lowercase , overwrite=lowercase ) for fname in fnames] if not overwrite and any(lowercase ): SCREAMING_SNAKE_CASE_: List[str] =[f for f, d in zip(lowercase , lowercase ) if d] raise ValueError( f'''The following files have auto mappings that need sorting: {", ".join(lowercase )}. Run `make style` to fix''' """ this.""" ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("""--check_only""", action="""store_true""", help="""Whether to only check or fix style.""") _UpperCAmelCase = parser.parse_args() sort_all_auto_mappings(not args.check_only)
173
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __magic_name__ ( lowercase=None ): if subparsers is not None: SCREAMING_SNAKE_CASE_: Dict =subparsers.add_parser("""test""" ) else: SCREAMING_SNAKE_CASE_: Optional[Any] =argparse.ArgumentParser("""Accelerate test command""" ) parser.add_argument( """--config_file""" , default=lowercase , help=( """The path to use to store the config file. Will default to a file named default_config.yaml in the cache """ """location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have """ """such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed """ """with 'huggingface'.""" ) , ) if subparsers is not None: parser.set_defaults(func=lowercase ) return parser def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: Dict =os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["""test_utils""", """scripts""", """test_script.py"""] ) if args.config_file is None: SCREAMING_SNAKE_CASE_: Any =script_name else: SCREAMING_SNAKE_CASE_: int =f'''--config_file={args.config_file} {script_name}''' SCREAMING_SNAKE_CASE_: Any =["""accelerate-launch"""] + test_args.split() SCREAMING_SNAKE_CASE_: Dict =execute_subprocess_async(lowercase , env=os.environ.copy() ) if result.returncode == 0: print("""Test is a success! You are ready for your distributed training!""" ) def __magic_name__ ( ): SCREAMING_SNAKE_CASE_: Optional[Any] =test_command_parser() SCREAMING_SNAKE_CASE_: List[str] =parser.parse_args() test_command(lowercase ) if __name__ == "__main__": main()
173
1
'''simple docstring''' from __future__ import annotations from typing import Generic, TypeVar _a : int = TypeVar("""T""") class _UpperCAmelCase ( Generic[T] ): def __init__( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = data __lowerCAmelCase = self __lowerCAmelCase = 0 class _UpperCAmelCase ( Generic[T] ): def __init__( self ): '''simple docstring''' __lowerCAmelCase = {} def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = DisjointSetTreeNode(__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = self.map[data] if elem_ref != elem_ref.parent: __lowerCAmelCase = self.find_set(elem_ref.parent.data ) return elem_ref.parent def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if nodea.rank > nodea.rank: __lowerCAmelCase = nodea else: __lowerCAmelCase = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' self.link(self.find_set(__SCREAMING_SNAKE_CASE ),self.find_set(__SCREAMING_SNAKE_CASE ) ) class _UpperCAmelCase ( Generic[T] ): def __init__( self ): '''simple docstring''' __lowerCAmelCase = {} def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if node not in self.connections: __lowerCAmelCase = {} def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' self.add_node(__SCREAMING_SNAKE_CASE ) self.add_node(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = weight __lowerCAmelCase = weight def lowerCamelCase__ ( self ): '''simple docstring''' __lowerCAmelCase = [] __lowerCAmelCase = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda __SCREAMING_SNAKE_CASE : x[2] ) # creating the disjoint set __lowerCAmelCase = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(__SCREAMING_SNAKE_CASE ) # MST generation __lowerCAmelCase = 0 __lowerCAmelCase = 0 __lowerCAmelCase = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = edges[index] index += 1 __lowerCAmelCase = disjoint_set.find_set(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = disjoint_set.find_set(__SCREAMING_SNAKE_CASE ) if parent_u != parent_v: num_edges += 1 graph.add_edge(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) disjoint_set.union(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) return graph
351
'''simple docstring''' from typing import Any, Dict, Optional import torch import torch.nn.functional as F from torch import nn from ..utils import maybe_allow_in_graph from .activations import get_activation from .attention_processor import Attention from .embeddings import CombinedTimestepLabelEmbeddings @maybe_allow_in_graph class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE=0.0,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = "geglu",__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = False,__SCREAMING_SNAKE_CASE = False,__SCREAMING_SNAKE_CASE = False,__SCREAMING_SNAKE_CASE = False,__SCREAMING_SNAKE_CASE = True,__SCREAMING_SNAKE_CASE = "layer_norm",__SCREAMING_SNAKE_CASE = False,): '''simple docstring''' super().__init__() __lowerCAmelCase = only_cross_attention __lowerCAmelCase = (num_embeds_ada_norm is not None) and norm_type == """ada_norm_zero""" __lowerCAmelCase = (num_embeds_ada_norm is not None) and norm_type == """ada_norm""" if norm_type in ("ada_norm", "ada_norm_zero") and num_embeds_ada_norm is None: raise ValueError( f'`norm_type` is set to {norm_type}, but `num_embeds_ada_norm` is not defined. Please make sure to' f' define `num_embeds_ada_norm` if setting `norm_type` to {norm_type}.' ) # Define 3 blocks. Each block has its own normalization layer. # 1. Self-Attn if self.use_ada_layer_norm: __lowerCAmelCase = AdaLayerNorm(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) elif self.use_ada_layer_norm_zero: __lowerCAmelCase = AdaLayerNormZero(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) else: __lowerCAmelCase = nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = Attention( query_dim=__SCREAMING_SNAKE_CASE,heads=__SCREAMING_SNAKE_CASE,dim_head=__SCREAMING_SNAKE_CASE,dropout=__SCREAMING_SNAKE_CASE,bias=__SCREAMING_SNAKE_CASE,cross_attention_dim=cross_attention_dim if only_cross_attention else None,upcast_attention=__SCREAMING_SNAKE_CASE,) # 2. Cross-Attn if cross_attention_dim is not None or double_self_attention: # We currently only use AdaLayerNormZero for self attention where there will only be one attention block. # I.e. the number of returned modulation chunks from AdaLayerZero would not make sense if returned during # the second cross attention block. __lowerCAmelCase = ( AdaLayerNorm(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) if self.use_ada_layer_norm else nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE ) ) __lowerCAmelCase = Attention( query_dim=__SCREAMING_SNAKE_CASE,cross_attention_dim=cross_attention_dim if not double_self_attention else None,heads=__SCREAMING_SNAKE_CASE,dim_head=__SCREAMING_SNAKE_CASE,dropout=__SCREAMING_SNAKE_CASE,bias=__SCREAMING_SNAKE_CASE,upcast_attention=__SCREAMING_SNAKE_CASE,) # is self-attn if encoder_hidden_states is none else: __lowerCAmelCase = None __lowerCAmelCase = None # 3. Feed-forward __lowerCAmelCase = nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = FeedForward(__SCREAMING_SNAKE_CASE,dropout=__SCREAMING_SNAKE_CASE,activation_fn=__SCREAMING_SNAKE_CASE,final_dropout=__SCREAMING_SNAKE_CASE ) # let chunk size default to None __lowerCAmelCase = None __lowerCAmelCase = 0 def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = chunk_size __lowerCAmelCase = dim def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = None,): '''simple docstring''' if self.use_ada_layer_norm: __lowerCAmelCase = self.norma(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) elif self.use_ada_layer_norm_zero: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self.norma( __SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,hidden_dtype=hidden_states.dtype ) else: __lowerCAmelCase = self.norma(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = cross_attention_kwargs if cross_attention_kwargs is not None else {} __lowerCAmelCase = self.attna( __SCREAMING_SNAKE_CASE,encoder_hidden_states=encoder_hidden_states if self.only_cross_attention else None,attention_mask=__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE,) if self.use_ada_layer_norm_zero: __lowerCAmelCase = gate_msa.unsqueeze(1 ) * attn_output __lowerCAmelCase = attn_output + hidden_states # 2. Cross-Attention if self.attna is not None: __lowerCAmelCase = ( self.norma(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) if self.use_ada_layer_norm else self.norma(__SCREAMING_SNAKE_CASE ) ) __lowerCAmelCase = self.attna( __SCREAMING_SNAKE_CASE,encoder_hidden_states=__SCREAMING_SNAKE_CASE,attention_mask=__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE,) __lowerCAmelCase = attn_output + hidden_states # 3. Feed-forward __lowerCAmelCase = self.norma(__SCREAMING_SNAKE_CASE ) if self.use_ada_layer_norm_zero: __lowerCAmelCase = norm_hidden_states * (1 + scale_mlp[:, None]) + shift_mlp[:, None] if self._chunk_size is not None: # "feed_forward_chunk_size" can be used to save memory if norm_hidden_states.shape[self._chunk_dim] % self._chunk_size != 0: raise ValueError( f'`hidden_states` dimension to be chunked: {norm_hidden_states.shape[self._chunk_dim]} has to be divisible by chunk size: {self._chunk_size}. Make sure to set an appropriate `chunk_size` when calling `unet.enable_forward_chunking`.' ) __lowerCAmelCase = norm_hidden_states.shape[self._chunk_dim] // self._chunk_size __lowerCAmelCase = torch.cat( [self.ff(__SCREAMING_SNAKE_CASE ) for hid_slice in norm_hidden_states.chunk(__SCREAMING_SNAKE_CASE,dim=self._chunk_dim )],dim=self._chunk_dim,) else: __lowerCAmelCase = self.ff(__SCREAMING_SNAKE_CASE ) if self.use_ada_layer_norm_zero: __lowerCAmelCase = gate_mlp.unsqueeze(1 ) * ff_output __lowerCAmelCase = ff_output + hidden_states return hidden_states class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = 4,__SCREAMING_SNAKE_CASE = 0.0,__SCREAMING_SNAKE_CASE = "geglu",__SCREAMING_SNAKE_CASE = False,): '''simple docstring''' super().__init__() __lowerCAmelCase = int(dim * mult ) __lowerCAmelCase = dim_out if dim_out is not None else dim if activation_fn == "gelu": __lowerCAmelCase = GELU(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) if activation_fn == "gelu-approximate": __lowerCAmelCase = GELU(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,approximate="""tanh""" ) elif activation_fn == "geglu": __lowerCAmelCase = GEGLU(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) elif activation_fn == "geglu-approximate": __lowerCAmelCase = ApproximateGELU(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.ModuleList([] ) # project in self.net.append(__SCREAMING_SNAKE_CASE ) # project dropout self.net.append(nn.Dropout(__SCREAMING_SNAKE_CASE ) ) # project out self.net.append(nn.Linear(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) ) # FF as used in Vision Transformer, MLP-Mixer, etc. have a final dropout if final_dropout: self.net.append(nn.Dropout(__SCREAMING_SNAKE_CASE ) ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' for module in self.net: __lowerCAmelCase = module(__SCREAMING_SNAKE_CASE ) return hidden_states class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = "none" ): '''simple docstring''' super().__init__() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = approximate def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if gate.device.type != "mps": return F.gelu(__SCREAMING_SNAKE_CASE,approximate=self.approximate ) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa ),approximate=self.approximate ).to(dtype=gate.dtype ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = self.proj(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = self.gelu(__SCREAMING_SNAKE_CASE ) return hidden_states class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,dim_out * 2 ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if gate.device.type != "mps": return F.gelu(__SCREAMING_SNAKE_CASE ) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa ) ).to(dtype=gate.dtype ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase , __lowerCAmelCase = self.proj(__SCREAMING_SNAKE_CASE ).chunk(2,dim=-1 ) return hidden_states * self.gelu(__SCREAMING_SNAKE_CASE ) class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = self.proj(__SCREAMING_SNAKE_CASE ) return x * torch.sigmoid(1.702 * x ) class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__() __lowerCAmelCase = nn.Embedding(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.SiLU() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,embedding_dim * 2 ) __lowerCAmelCase = nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = self.linear(self.silu(self.emb(__SCREAMING_SNAKE_CASE ) ) ) __lowerCAmelCase , __lowerCAmelCase = torch.chunk(__SCREAMING_SNAKE_CASE,2 ) __lowerCAmelCase = self.norm(__SCREAMING_SNAKE_CASE ) * (1 + scale) + shift return x class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__() __lowerCAmelCase = CombinedTimestepLabelEmbeddings(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.SiLU() __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,6 * embedding_dim,bias=__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.LayerNorm(__SCREAMING_SNAKE_CASE,elementwise_affine=__SCREAMING_SNAKE_CASE,eps=1e-6 ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE=None ): '''simple docstring''' __lowerCAmelCase = self.linear(self.silu(self.emb(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,hidden_dtype=__SCREAMING_SNAKE_CASE ) ) ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = emb.chunk(6,dim=1 ) __lowerCAmelCase = self.norm(__SCREAMING_SNAKE_CASE ) * (1 + scale_msa[:, None]) + shift_msa[:, None] return x, gate_msa, shift_mlp, scale_mlp, gate_mlp class _UpperCAmelCase ( nn.Module ): def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = 1e-5 ): '''simple docstring''' super().__init__() __lowerCAmelCase = num_groups __lowerCAmelCase = eps if act_fn is None: __lowerCAmelCase = None else: __lowerCAmelCase = get_activation(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = nn.Linear(__SCREAMING_SNAKE_CASE,out_dim * 2 ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if self.act: __lowerCAmelCase = self.act(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = self.linear(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase = emb[:, :, None, None] __lowerCAmelCase , __lowerCAmelCase = emb.chunk(2,dim=1 ) __lowerCAmelCase = F.group_norm(__SCREAMING_SNAKE_CASE,self.num_groups,eps=self.eps ) __lowerCAmelCase = x * (1 + scale) + shift return x
46
0
"""simple docstring""" import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _snake_case : Union[str, Any] = pytest.mark.integration @require_faiss class _UpperCAmelCase ( __snake_case ): def lowerCamelCase ( self :Union[str, Any] ): A = Dataset.from_dict({"filename": ["my_name-train" + "_" + str(A_ ) for x in np.arange(30 ).tolist()]} ) return dset def lowerCamelCase ( self :Optional[Any] ): import faiss A = self._create_dummy_dataset() A = dset.map( lambda __UpperCamelCase , __UpperCamelCase : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=A_ , keep_in_memory=A_ ) A = dset.add_faiss_index("vecs" , batch_size=1_00 , metric_type=faiss.METRIC_INNER_PRODUCT ) A = dset.get_nearest_examples("vecs" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) dset.drop_index("vecs" ) def lowerCamelCase ( self :Tuple ): import faiss A = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="vecs" , batch_size=1_00 , metric_type=faiss.METRIC_INNER_PRODUCT , ) A = dset.get_nearest_examples("vecs" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) def lowerCamelCase ( self :List[str] ): import faiss A = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="vecs" , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=A_ ) as tmp_file: dset.save_faiss_index("vecs" , tmp_file.name ) dset.load_faiss_index("vecs2" , tmp_file.name ) os.unlink(tmp_file.name ) A = dset.get_nearest_examples("vecs2" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) def lowerCamelCase ( self :Any ): A = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="vecs" ) dset.drop_index("vecs" ) self.assertRaises(A_ , partial(dset.get_nearest_examples , "vecs2" , np.ones(5 , dtype=np.floataa ) ) ) def lowerCamelCase ( self :str ): from elasticsearch import Elasticsearch A = self._create_dummy_dataset() with patch("elasticsearch.Elasticsearch.search" ) as mocked_search, patch( "elasticsearch.client.IndicesClient.create" ) as mocked_index_create, patch("elasticsearch.helpers.streaming_bulk" ) as mocked_bulk: A = {"acknowledged": True} mocked_bulk.return_value([(True, None)] * 30 ) A = {"hits": {"hits": [{"_score": 1, "_id": 29}]}} A = Elasticsearch() dset.add_elasticsearch_index("filename" , es_client=A_ ) A = dset.get_nearest_examples("filename" , "my_name-train_29" ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) @require_faiss class _UpperCAmelCase ( __snake_case ): def lowerCamelCase ( self :Optional[int] ): import faiss A = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query A = np.zeros(5 , dtype=np.floataa ) A = 1 A = index.search(A_ ) self.assertRaises(A_ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries A = np.eye(5 , dtype=np.floataa )[::-1] A = index.search_batch(A_ ) self.assertRaises(A_ , index.search_batch , queries[0] ) A = [scores[0] for scores in total_scores] A = [indices[0] for indices in total_indices] self.assertGreater(np.min(A_ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , A_ ) def lowerCamelCase ( self :List[Any] ): import faiss A = FaissIndex(string_factory="Flat" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) A = FaissIndex(string_factory="LSH" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(A_ ): A = FaissIndex(string_factory="Flat" , custom_index=faiss.IndexFlat(5 ) ) def lowerCamelCase ( self :List[str] ): import faiss A = faiss.IndexFlat(5 ) A = FaissIndex(custom_index=A_ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def lowerCamelCase ( self :List[Any] ): import faiss A = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=A_ ) as tmp_file: index.save(tmp_file.name ) A = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) A = np.zeros(5 , dtype=np.floataa ) A = 1 A = index.search(A_ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def A__ ( UpperCamelCase ): import faiss A = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) A = "index.faiss" A = F"mock://{index_name}" index.save(_lowerCAmelCase , storage_options=mockfs.storage_options ) A = FaissIndex.load(_lowerCAmelCase , storage_options=mockfs.storage_options ) A = np.zeros(5 , dtype=np.floataa ) A = 1 A = index.search(_lowerCAmelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class _UpperCAmelCase ( __snake_case ): def lowerCamelCase ( self :Union[str, Any] ): from elasticsearch import Elasticsearch with patch("elasticsearch.Elasticsearch.search" ) as mocked_search, patch( "elasticsearch.client.IndicesClient.create" ) as mocked_index_create, patch("elasticsearch.helpers.streaming_bulk" ) as mocked_bulk: A = Elasticsearch() A = {"acknowledged": True} A = ElasticSearchIndex(es_client=A_ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(["foo", "bar", "foobar"] ) # single query A = "foo" A = {"hits": {"hits": [{"_score": 1, "_id": 0}]}} A = index.search(A_ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout A = "foo" A = {"hits": {"hits": [{"_score": 1, "_id": 0}]}} A = index.search(A_ , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries A = ["foo", "bar", "foobar"] A = {"hits": {"hits": [{"_score": 1, "_id": 1}]}} A = index.search_batch(A_ ) A = [scores[0] for scores in total_scores] A = [indices[0] for indices in total_indices] self.assertGreater(np.min(A_ ) , 0 ) self.assertListEqual([1, 1, 1] , A_ ) # batched queries with timeout A = ["foo", "bar", "foobar"] A = {"hits": {"hits": [{"_score": 1, "_id": 1}]}} A = index.search_batch(A_ , request_timeout=30 ) A = [scores[0] for scores in total_scores] A = [indices[0] for indices in total_indices] self.assertGreater(np.min(A_ ) , 0 ) self.assertListEqual([1, 1, 1] , A_ )
292
import pickle import numpy as np from matplotlib import pyplot as plt class A__ : def __init__( self , A_ , A_ , A_ , A_ , A_ , A_=0.2 , A_=0.2 ): '''simple docstring''' UpperCamelCase : int = bp_numa UpperCamelCase : int = bp_numa UpperCamelCase : List[Any] = bp_numa UpperCamelCase : Optional[int] = conva_get[:2] UpperCamelCase : Optional[Any] = conva_get[2] UpperCamelCase : Dict = size_pa UpperCamelCase : Union[str, Any] = rate_w UpperCamelCase : Dict = rate_t UpperCamelCase : Union[str, Any] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] UpperCamelCase : Any = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCamelCase : List[Any] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCamelCase : Optional[Any] = -2 * np.random.rand(self.conva[1] ) + 1 UpperCamelCase : Any = -2 * np.random.rand(self.num_bpa ) + 1 UpperCamelCase : int = -2 * np.random.rand(self.num_bpa ) + 1 def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(A_ , "wb" ) as f: pickle.dump(A_ , A_ ) print(F"""Model saved: {save_path}""" ) @classmethod def __UpperCamelCase( cls , A_ ): '''simple docstring''' with open(A_ , "rb" ) as f: UpperCamelCase : Optional[Any] = pickle.load(A_ ) # noqa: S301 UpperCamelCase : List[Any] = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) UpperCamelCase : Union[str, Any] = model_dic.get("size_pooling1" ) UpperCamelCase : List[Any] = model_dic.get("num_bp1" ) UpperCamelCase : Dict = model_dic.get("num_bp2" ) UpperCamelCase : Dict = model_dic.get("num_bp3" ) UpperCamelCase : Dict = model_dic.get("rate_weight" ) UpperCamelCase : str = model_dic.get("rate_thre" ) # create model instance UpperCamelCase : Any = CNN(A_ , A_ , A_ , A_ , A_ , A_ , A_ ) # modify model parameter UpperCamelCase : str = model_dic.get("w_conv1" ) UpperCamelCase : Optional[Any] = model_dic.get("wkj" ) UpperCamelCase : int = model_dic.get("vji" ) UpperCamelCase : Any = model_dic.get("thre_conv1" ) UpperCamelCase : Optional[int] = model_dic.get("thre_bp2" ) UpperCamelCase : Union[str, Any] = model_dic.get("thre_bp3" ) return conv_ins def __UpperCamelCase( self , A_ ): '''simple docstring''' return 1 / (1 + np.exp(-1 * x )) def __UpperCamelCase( self , A_ ): '''simple docstring''' return round(A_ , 3 ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = convs[0] UpperCamelCase : Optional[Any] = convs[1] UpperCamelCase : Optional[Any] = np.shape(A_ )[0] # get the data slice of original image data, data_focus UpperCamelCase : List[str] = [] for i_focus in range(0 , size_data - size_conv + 1 , A_ ): for j_focus in range(0 , size_data - size_conv + 1 , A_ ): UpperCamelCase : Union[str, Any] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(A_ ) # calculate the feature map of every single kernel, and saved as list of matrix UpperCamelCase : int = [] UpperCamelCase : Optional[Any] = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(A_ ): UpperCamelCase : str = [] for i_focus in range(len(A_ ) ): UpperCamelCase : List[Any] = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(A_ ) ) UpperCamelCase : Optional[int] = np.asmatrix(A_ ).reshape( A_ , A_ ) data_featuremap.append(A_ ) # expanding the data slice to One dimenssion UpperCamelCase : List[Any] = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(A_ ) ) UpperCamelCase : Tuple = np.asarray(A_ ) return focus_list, data_featuremap def __UpperCamelCase( self , A_ , A_ , A_="average_pool" ): '''simple docstring''' UpperCamelCase : Any = len(featuremaps[0] ) UpperCamelCase : str = int(size_map / size_pooling ) UpperCamelCase : Optional[int] = [] for i_map in range(len(A_ ) ): UpperCamelCase : Tuple = featuremaps[i_map] UpperCamelCase : Any = [] for i_focus in range(0 , A_ , A_ ): for j_focus in range(0 , A_ , A_ ): UpperCamelCase : int = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(A_ ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(A_ ) ) UpperCamelCase : Optional[Any] = np.asmatrix(A_ ).reshape(A_ , A_ ) featuremap_pooled.append(A_ ) return featuremap_pooled def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = [] for i in range(len(A_ ) ): UpperCamelCase : List[Any] = np.shape(data[i] ) UpperCamelCase : str = data[i].reshape(1 , shapes[0] * shapes[1] ) UpperCamelCase : Optional[int] = data_listed.getA().tolist()[0] data_expanded.extend(A_ ) UpperCamelCase : Any = np.asarray(A_ ) return data_expanded def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = np.asarray(A_ ) UpperCamelCase : List[Any] = np.shape(A_ ) UpperCamelCase : Any = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : int = [] UpperCamelCase : Optional[int] = 0 for i_map in range(A_ ): UpperCamelCase : int = np.ones((size_map, size_map) ) for i in range(0 , A_ , A_ ): for j in range(0 , A_ , A_ ): UpperCamelCase : str = pd_pool[ i_pool ] UpperCamelCase : str = i_pool + 1 UpperCamelCase : str = np.multiply( A_ , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(A_ ) return pd_all def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_=bool ): '''simple docstring''' print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(A_ )) ) print((" - - Shape: Teach_Data ", np.shape(A_ )) ) UpperCamelCase : List[str] = 0 UpperCamelCase : Union[str, Any] = [] UpperCamelCase : int = 1_0000 while rp < n_repeat and mse >= error_accuracy: UpperCamelCase : Tuple = 0 print(F"""-------------Learning Time {rp}--------------""" ) for p in range(len(A_ ) ): # print('------------Learning Image: %d--------------'%p) UpperCamelCase : Any = np.asmatrix(datas_train[p] ) UpperCamelCase : List[str] = np.asarray(datas_teach[p] ) UpperCamelCase , UpperCamelCase : Dict = self.convolute( A_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase : Tuple = self.pooling(A_ , self.size_poolinga ) UpperCamelCase : int = np.shape(A_ ) UpperCamelCase : List[str] = self._expand(A_ ) UpperCamelCase : Optional[int] = data_bp_input UpperCamelCase : str = np.dot(A_ , self.vji.T ) - self.thre_bpa UpperCamelCase : Optional[int] = self.sig(A_ ) UpperCamelCase : List[Any] = np.dot(A_ , self.wkj.T ) - self.thre_bpa UpperCamelCase : Dict = self.sig(A_ ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- UpperCamelCase : List[Any] = np.multiply( (data_teach - bp_outa) , np.multiply(A_ , (1 - bp_outa) ) ) UpperCamelCase : str = np.multiply( np.dot(A_ , self.wkj ) , np.multiply(A_ , (1 - bp_outa) ) ) UpperCamelCase : Any = np.dot(A_ , self.vji ) UpperCamelCase : Dict = pd_i_all / (self.size_poolinga * self.size_poolinga) UpperCamelCase : List[Any] = pd_conva_pooled.T.getA().tolist() UpperCamelCase : List[Any] = self._calculate_gradient_from_pool( A_ , A_ , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): UpperCamelCase : List[Any] = self._expand_mat(pd_conva_all[k_conv] ) UpperCamelCase : List[Any] = self.rate_weight * np.dot(A_ , A_ ) UpperCamelCase : str = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) UpperCamelCase : Dict = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer UpperCamelCase : Optional[Any] = self.wkj + pd_k_all.T * bp_outa * self.rate_weight UpperCamelCase : List[Any] = self.vji + pd_j_all.T * bp_outa * self.rate_weight UpperCamelCase : Optional[Any] = self.thre_bpa - pd_k_all * self.rate_thre UpperCamelCase : List[str] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image UpperCamelCase : List[Any] = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) UpperCamelCase : Any = rp + 1 UpperCamelCase : Union[str, Any] = error_count / patterns all_mse.append(A_ ) def draw_error(): UpperCamelCase : Tuple = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(A_ , "+-" ) plt.plot(A_ , "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(A_ , alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, F""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(A_ )) ) for p in range(len(A_ ) ): UpperCamelCase : int = np.asmatrix(datas_test[p] ) UpperCamelCase , UpperCamelCase : Any = self.convolute( A_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase : List[str] = self.pooling(A_ , self.size_poolinga ) UpperCamelCase : Dict = self._expand(A_ ) UpperCamelCase : List[Any] = data_bp_input UpperCamelCase : Any = bp_outa * self.vji.T - self.thre_bpa UpperCamelCase : List[Any] = self.sig(A_ ) UpperCamelCase : int = bp_outa * self.wkj.T - self.thre_bpa UpperCamelCase : Optional[int] = self.sig(A_ ) produce_out.extend(bp_outa.getA().tolist() ) UpperCamelCase : List[str] = [list(map(self.do_round , A_ ) ) for each in produce_out] return np.asarray(A_ ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = np.asmatrix(A_ ) UpperCamelCase , UpperCamelCase : List[Any] = self.convolute( A_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase : str = self.pooling(A_ , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
52
0
import math import sys def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" if number != int(__SCREAMING_SNAKE_CASE ): raise ValueError('''the value of input must be a natural number''' ) if number < 0: raise ValueError('''the value of input must not be a negative number''' ) if number == 0: return 1 snake_case_ : int = [-1] * (number + 1) snake_case_ : Optional[Any] = 0 for i in range(1 , number + 1 ): snake_case_ : List[str] = sys.maxsize snake_case_ : int = int(math.sqrt(__SCREAMING_SNAKE_CASE ) ) for j in range(1 , root + 1 ): snake_case_ : Optional[Any] = 1 + answers[i - (j**2)] snake_case_ : List[Any] = min(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case_ : Optional[Any] = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
370
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { '''microsoft/biogpt''': '''https://huggingface.co/microsoft/biogpt/resolve/main/config.json''', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class __lowerCAmelCase ( _a ): lowerCamelCase_ : Any = '''biogpt''' def __init__(self , __magic_name__=4_2384 , __magic_name__=1024 , __magic_name__=24 , __magic_name__=16 , __magic_name__=4096 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=1024 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=True , __magic_name__=True , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , **__magic_name__ , ) -> List[str]: '''simple docstring''' snake_case_ : List[str] = vocab_size snake_case_ : Dict = max_position_embeddings snake_case_ : Optional[int] = hidden_size snake_case_ : List[Any] = num_hidden_layers snake_case_ : List[str] = num_attention_heads snake_case_ : int = intermediate_size snake_case_ : List[Any] = hidden_act snake_case_ : List[Any] = hidden_dropout_prob snake_case_ : Optional[int] = attention_probs_dropout_prob snake_case_ : Optional[int] = initializer_range snake_case_ : Optional[int] = layer_norm_eps snake_case_ : str = scale_embedding snake_case_ : Optional[Any] = use_cache snake_case_ : Optional[Any] = layerdrop snake_case_ : Optional[Any] = activation_dropout super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ )
279
0
'''simple docstring''' import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def UpperCamelCase_ ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any]=False ) -> Any: """simple docstring""" _UpperCAmelCase : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""module.blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""module.blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append( (F"""module.blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""module.blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""module.blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""module.blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""module.blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""module.blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""module.blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""module.blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("module.cls_token", "vit.embeddings.cls_token"), ("module.patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("module.patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("module.pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("module.norm.weight", "layernorm.weight"), ("module.norm.bias", "layernorm.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCAmelCase : str = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def UpperCamelCase_ ( _UpperCAmelCase : int , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict=False ) -> Optional[Any]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCAmelCase : int = "" else: _UpperCAmelCase : Any = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCAmelCase : List[Any] = state_dict.pop(F"""module.blocks.{i}.attn.qkv.weight""" ) _UpperCAmelCase : Optional[int] = state_dict.pop(F"""module.blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _UpperCAmelCase : List[Any] = in_proj_weight[ : config.hidden_size, : ] _UpperCAmelCase : List[str] = in_proj_bias[: config.hidden_size] _UpperCAmelCase : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCAmelCase : str = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCAmelCase : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] _UpperCAmelCase : List[str] = in_proj_bias[-config.hidden_size :] def UpperCamelCase_ ( _UpperCAmelCase : List[str] ) -> Optional[int]: """simple docstring""" _UpperCAmelCase : str = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def UpperCamelCase_ ( _UpperCAmelCase : List[Any] ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase : List[str] = [ "module.fc.fc1.weight", "module.fc.fc1.bias", "module.fc.bn1.weight", "module.fc.bn1.bias", "module.fc.bn1.running_mean", "module.fc.bn1.running_var", "module.fc.bn1.num_batches_tracked", "module.fc.fc2.weight", "module.fc.fc2.bias", "module.fc.bn2.weight", "module.fc.bn2.bias", "module.fc.bn2.running_mean", "module.fc.bn2.running_var", "module.fc.bn2.num_batches_tracked", "module.fc.fc3.weight", "module.fc.fc3.bias", ] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def UpperCamelCase_ ( _UpperCAmelCase : Dict , _UpperCAmelCase : Any , _UpperCAmelCase : Tuple ) -> List[str]: """simple docstring""" _UpperCAmelCase : List[Any] = dct.pop(_UpperCAmelCase ) _UpperCAmelCase : int = val def UpperCamelCase_ ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict ) -> str: """simple docstring""" _UpperCAmelCase : List[Any] = ViTMSNConfig() _UpperCAmelCase : Optional[int] = 1_000 _UpperCAmelCase : str = "datasets/huggingface/label-files" _UpperCAmelCase : List[str] = "imagenet-1k-id2label.json" _UpperCAmelCase : Optional[Any] = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase ) , "r" ) ) _UpperCAmelCase : List[Any] = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} _UpperCAmelCase : Tuple = idalabel _UpperCAmelCase : Dict = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: _UpperCAmelCase : Dict = 384 _UpperCAmelCase : Any = 1_536 _UpperCAmelCase : str = 6 elif "l16" in checkpoint_url: _UpperCAmelCase : Union[str, Any] = 1_024 _UpperCAmelCase : Any = 4_096 _UpperCAmelCase : List[Any] = 24 _UpperCAmelCase : Tuple = 16 _UpperCAmelCase : Optional[Any] = 0.1 elif "b4" in checkpoint_url: _UpperCAmelCase : Optional[Any] = 4 elif "l7" in checkpoint_url: _UpperCAmelCase : List[str] = 7 _UpperCAmelCase : Optional[Any] = 1_024 _UpperCAmelCase : List[str] = 4_096 _UpperCAmelCase : Optional[int] = 24 _UpperCAmelCase : Union[str, Any] = 16 _UpperCAmelCase : Optional[int] = 0.1 _UpperCAmelCase : Tuple = ViTMSNModel(_UpperCAmelCase ) _UpperCAmelCase : int = torch.hub.load_state_dict_from_url(_UpperCAmelCase , map_location="cpu" )["target_encoder"] _UpperCAmelCase : List[Any] = ViTImageProcessor(size=config.image_size ) remove_projection_head(_UpperCAmelCase ) _UpperCAmelCase : str = create_rename_keys(_UpperCAmelCase , base_model=_UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) read_in_q_k_v(_UpperCAmelCase , _UpperCAmelCase , base_model=_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() _UpperCAmelCase : Any = "http://images.cocodataset.org/val2017/000000039769.jpg" _UpperCAmelCase : List[Any] = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) _UpperCAmelCase : Any = ViTImageProcessor( size=config.image_size , image_mean=_UpperCAmelCase , image_std=_UpperCAmelCase ) _UpperCAmelCase : Dict = image_processor(images=_UpperCAmelCase , return_tensors="pt" ) # forward pass torch.manual_seed(2 ) _UpperCAmelCase : Tuple = model(**_UpperCAmelCase ) _UpperCAmelCase : int = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: _UpperCAmelCase : Optional[int] = torch.tensor([[-1.0_9_1_5, -1.4_8_7_6, -1.1_8_0_9]] ) elif "b16" in checkpoint_url: _UpperCAmelCase : List[str] = torch.tensor([[1_4.2_8_8_9, -1_8.9_0_4_5, 1_1.7_2_8_1]] ) elif "l16" in checkpoint_url: _UpperCAmelCase : Optional[int] = torch.tensor([[4_1.5_0_2_8, -2_2.8_6_8_1, 4_5.6_4_7_5]] ) elif "b4" in checkpoint_url: _UpperCAmelCase : List[str] = torch.tensor([[-4.3_8_6_8, 5.2_9_3_2, -0.4_1_3_7]] ) else: _UpperCAmelCase : str = torch.tensor([[-0.1_7_9_2, -0.6_4_6_5, 2.4_2_6_3]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , _UpperCAmelCase , atol=1e-4 ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar""", type=str, help="""URL of the checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) __SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
31
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _lowerCamelCase =get_logger(__name__) class A__ : def __init__( self , __magic_name__ = None ): lowerCamelCase : Dict = ( os.path.join(__magic_name__ , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) lowerCamelCase : List[str] = Extractor def UpperCamelCase__ ( self , __magic_name__ ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" lowerCamelCase : int = os.path.abspath(__magic_name__ ) return os.path.join(self.extract_dir , hash_url_to_filename(__magic_name__ ) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): return force_extract or ( not os.path.isfile(__magic_name__ ) and not (os.path.isdir(__magic_name__ ) and os.listdir(__magic_name__ )) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False ): lowerCamelCase : Union[str, Any] = self.extractor.infer_extractor_format(__magic_name__ ) if not extractor_format: return input_path lowerCamelCase : int = self._get_output_path(__magic_name__ ) if self._do_extract(__magic_name__ , __magic_name__ ): self.extractor.extract(__magic_name__ , __magic_name__ , __magic_name__ ) return output_path class A__ ( __SCREAMING_SNAKE_CASE): @classmethod @abstractmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): ... @staticmethod @abstractmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): ... class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[bytes] = [] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with open(__magic_name__ , """rb""" ) as f: return f.read(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if not magic_number: lowerCamelCase : Optional[Any] = max(len(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) try: lowerCamelCase : Tuple = cls.read_magic_number(__magic_name__ , __magic_name__ ) except OSError: return False return any(magic_number.startswith(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) class A__ ( __SCREAMING_SNAKE_CASE): @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): return tarfile.is_tarfile(__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): def resolved(__magic_name__ ) -> str: return os.path.realpath(os.path.abspath(__magic_name__ ) ) def badpath(__magic_name__ , __magic_name__ ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__magic_name__ , __magic_name__ ) ).startswith(__magic_name__ ) def badlink(__magic_name__ , __magic_name__ ) -> bool: # Links are interpreted relative to the directory containing the link lowerCamelCase : List[str] = resolved(os.path.join(__magic_name__ , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__magic_name__ ) lowerCamelCase : Optional[Any] = resolved(__magic_name__ ) for finfo in members: if badpath(finfo.name , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Dict = tarfile.open(__magic_name__ ) tar_file.extractall(__magic_name__ , members=TarExtractor.safemembers(__magic_name__ , __magic_name__ ) ) tar_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = [B"""\x1F\x8B"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with gzip.open(__magic_name__ , """rb""" ) as gzip_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[int] = [ B"""PK\x03\x04""", B"""PK\x05\x06""", # empty archive B"""PK\x07\x08""", # spanned archive ] @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if super().is_extractable(__magic_name__ , magic_number=__magic_name__ ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__magic_name__ , """rb""" ) as fp: lowerCamelCase : List[str] = _EndRecData(__magic_name__ ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: lowerCamelCase : List[Any] = fp.read(__magic_name__ ) # CD is where we expect it to be if len(__magic_name__ ) == sizeCentralDir: lowerCamelCase : str = struct.unpack(__magic_name__ , __magic_name__ ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with zipfile.ZipFile(__magic_name__ , """r""" ) as zip_file: zip_file.extractall(__magic_name__ ) zip_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[str] = [B"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with lzma.open(__magic_name__ ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""Rar!\x1a\x07\x00""", B"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Union[str, Any] = rarfile.RarFile(__magic_name__ ) rf.extractall(__magic_name__ ) rf.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = [B"""\x28\xb5\x2F\xFD"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd lowerCamelCase : int = zstd.ZstdDecompressor() with open(__magic_name__ , """rb""" ) as ifh, open(__magic_name__ , """wb""" ) as ofh: dctx.copy_stream(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""\x42\x5A\x68"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with bza.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with pyazr.SevenZipFile(__magic_name__ , """r""" ) as archive: archive.extractall(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x04\x22\x4D\x18"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) _UpperCAmelCase : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def UpperCamelCase__ ( cls ): return max( len(__magic_name__ ) for extractor in cls.extractors.values() if issubclass(__magic_name__ , __magic_name__ ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): try: return MagicNumberBaseExtractor.read_magic_number(__magic_name__ , magic_number_length=__magic_name__ ) except OSError: return b"" @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = False ): warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = cls.infer_extractor_format(__magic_name__ ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): # <Added version="2.4.0"/> lowerCamelCase : Dict = cls._get_magic_number_max_length() lowerCamelCase : Optional[Any] = cls._read_magic_number(__magic_name__ , __magic_name__ ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__magic_name__ , magic_number=__magic_name__ ): return extractor_format @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = "deprecated" , ): os.makedirs(os.path.dirname(__magic_name__ ) , exist_ok=__magic_name__ ) # Prevent parallel extractions lowerCamelCase : Tuple = str(Path(__magic_name__ ).with_suffix(""".lock""" ) ) with FileLock(__magic_name__ ): shutil.rmtree(__magic_name__ , ignore_errors=__magic_name__ ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__magic_name__ , __magic_name__ ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = extractor if extractor != """deprecated""" else extractor_format else: lowerCamelCase : Optional[int] = cls.extractors[extractor_format] return extractor.extract(__magic_name__ , __magic_name__ ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=__magic_name__ , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__magic_name__ ): return extractor.extract(__magic_name__ , __magic_name__ )
287
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: lowercase_ = None lowercase_ = logging.get_logger(__name__) lowercase_ = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} lowercase_ = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", }, "tokenizer_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json", }, } lowercase_ = { "albert-base-v1": 5_1_2, "albert-large-v1": 5_1_2, "albert-xlarge-v1": 5_1_2, "albert-xxlarge-v1": 5_1_2, "albert-base-v2": 5_1_2, "albert-large-v2": 5_1_2, "albert-xlarge-v2": 5_1_2, "albert-xxlarge-v2": 5_1_2, } lowercase_ = "▁" class __lowerCAmelCase ( __a ): '''simple docstring''' __UpperCAmelCase : Any = VOCAB_FILES_NAMES __UpperCAmelCase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : List[str] = AlbertTokenizer def __init__( self , _a=None , _a=None , _a=True , _a=True , _a=False , _a="[CLS]" , _a="[SEP]" , _a="<unk>" , _a="[SEP]" , _a="<pad>" , _a="[CLS]" , _a="[MASK]" , **_a , ): __a = ( AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ , normalized=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token ) super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , do_lower_case=UpperCamelCase__ , remove_space=UpperCamelCase__ , keep_accents=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , **UpperCamelCase__ , ) __a = do_lower_case __a = remove_space __a = keep_accents __a = vocab_file __a = False if not self.vocab_file else True def __UpperCAmelCase ( self , _a , _a = None ): __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __UpperCAmelCase ( self , _a , _a = None ): __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self , _a , _a = None ): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(UpperCamelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __a = os.path.join( UpperCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) return (out_vocab_file,)
368
"""simple docstring""" import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowercase ( lowerCAmelCase__ : Dict ) -> Optional[int]: __a , __a = image.size __a , __a = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 __a = image.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) __a = np.array(lowerCAmelCase__ ).astype(np.floataa ) / 2_55.0 __a = image[None].transpose(0 , 3 , 1 , 2 ) __a = torch.from_numpy(lowerCAmelCase__ ) return 2.0 * image - 1.0 class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , _a , _a , _a , ): super().__init__() self.register_modules(vqvae=_a , unet=_a , scheduler=_a ) @torch.no_grad() def __call__( self , _a = None , _a = 1 , _a = 100 , _a = 0.0 , _a = None , _a = "pil" , _a = True , ): if isinstance(_a , PIL.Image.Image ): __a = 1 elif isinstance(_a , torch.Tensor ): __a = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(_a )}''' ) if isinstance(_a , PIL.Image.Image ): __a = preprocess(_a ) __a , __a = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __a = (batch_size, self.unet.config.in_channels // 2, height, width) __a = next(self.unet.parameters() ).dtype __a = randn_tensor(_a , generator=_a , device=self.device , dtype=_a ) __a = image.to(device=self.device , dtype=_a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(_a , device=self.device ) __a = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __a = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __a = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __a = {} if accepts_eta: __a = eta for t in self.progress_bar(_a ): # concat latents and low resolution image in the channel dimension. __a = torch.cat([latents, image] , dim=1 ) __a = self.scheduler.scale_model_input(_a , _a ) # predict the noise residual __a = self.unet(_a , _a ).sample # compute the previous noisy sample x_t -> x_t-1 __a = self.scheduler.step(_a , _a , _a , **_a ).prev_sample # decode the image latents with the VQVAE __a = self.vqvae.decode(_a ).sample __a = torch.clamp(_a , -1.0 , 1.0 ) __a = image / 2 + 0.5 __a = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __a = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
11
0
"""simple docstring""" import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: __UpperCamelCase = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=3 , lowerCAmelCase__=18 , lowerCAmelCase__=30 , lowerCAmelCase__=400 , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=None , ) -> Any: SCREAMING_SNAKE_CASE = size if size is not None else {'height': 20, 'width': 20} SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = image_size SCREAMING_SNAKE_CASE = min_resolution SCREAMING_SNAKE_CASE = max_resolution SCREAMING_SNAKE_CASE = size SCREAMING_SNAKE_CASE = do_normalize SCREAMING_SNAKE_CASE = do_convert_rgb SCREAMING_SNAKE_CASE = [512, 1_024, 2_048, 4_096] SCREAMING_SNAKE_CASE = patch_size if patch_size is not None else {'height': 16, 'width': 16} def __A ( self ) -> str: return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def __A ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE = 'https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg' SCREAMING_SNAKE_CASE = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ).convert('RGB' ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class lowerCAmelCase ( lowerCamelCase_ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = PixaStructImageProcessor if is_vision_available() else None def __A ( self ) -> Any: SCREAMING_SNAKE_CASE = PixaStructImageProcessingTester(self ) @property def __A ( self ) -> Tuple: return self.image_processor_tester.prepare_image_processor_dict() def __A ( self ) -> Union[str, Any]: SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_normalize' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_convert_rgb' ) ) def __A ( self ) -> Dict: SCREAMING_SNAKE_CASE = self.image_processor_tester.prepare_dummy_image() SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) SCREAMING_SNAKE_CASE = 2_048 SCREAMING_SNAKE_CASE = image_processor(lowerCAmelCase__ , return_tensors='pt' , max_patches=lowerCAmelCase__ ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() , torch.tensor(0.06_06 ) , atol=1e-3 , rtol=1e-3 ) ) def __A ( self ) -> Union[str, Any]: # Initialize image_processor SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input SCREAMING_SNAKE_CASE = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched SCREAMING_SNAKE_CASE = image_processor( lowerCAmelCase__ , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __A ( self ) -> Any: # Initialize image_processor SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 SCREAMING_SNAKE_CASE = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(lowerCAmelCase__ ): SCREAMING_SNAKE_CASE = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches SCREAMING_SNAKE_CASE = 'Hello' SCREAMING_SNAKE_CASE = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=lowerCAmelCase__ , header_text=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched SCREAMING_SNAKE_CASE = image_processor( lowerCAmelCase__ , return_tensors='pt' , max_patches=lowerCAmelCase__ , header_text=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __A ( self ) -> Tuple: # Initialize image_processor SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , numpify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , np.ndarray ) SCREAMING_SNAKE_CASE = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input SCREAMING_SNAKE_CASE = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched SCREAMING_SNAKE_CASE = image_processor( lowerCAmelCase__ , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __A ( self ) -> Any: # Initialize image_processor SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , torchify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input SCREAMING_SNAKE_CASE = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched SCREAMING_SNAKE_CASE = image_processor( lowerCAmelCase__ , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class lowerCAmelCase ( lowerCamelCase_ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[Any] = PixaStructImageProcessor if is_vision_available() else None def __A ( self ) -> Dict: SCREAMING_SNAKE_CASE = PixaStructImageProcessingTester(self , num_channels=4 ) SCREAMING_SNAKE_CASE = 3 @property def __A ( self ) -> List[str]: return self.image_processor_tester.prepare_image_processor_dict() def __A ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_normalize' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_convert_rgb' ) ) def __A ( self ) -> int: # Initialize image_processor SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input SCREAMING_SNAKE_CASE = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched SCREAMING_SNAKE_CASE = image_processor( lowerCAmelCase__ , return_tensors='pt' , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , )
113
"""simple docstring""" from __future__ import annotations from collections.abc import Callable def lowercase (SCREAMING_SNAKE_CASE_ : Callable[[int | float], int | float] , SCREAMING_SNAKE_CASE_ : int | float , SCREAMING_SNAKE_CASE_ : int | float , SCREAMING_SNAKE_CASE_ : int = 1_00 , ) -> float: SCREAMING_SNAKE_CASE = x_start SCREAMING_SNAKE_CASE = fnc(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = 0.0 for _ in range(SCREAMING_SNAKE_CASE_ ): # Approximates small segments of curve as linear and solve # for trapezoidal area SCREAMING_SNAKE_CASE = (x_end - x_start) / steps + xa SCREAMING_SNAKE_CASE = fnc(SCREAMING_SNAKE_CASE_ ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step SCREAMING_SNAKE_CASE = xa SCREAMING_SNAKE_CASE = fxa return area if __name__ == "__main__": def lowercase (SCREAMING_SNAKE_CASE_ : Optional[int] ) -> Dict: return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') __UpperCamelCase = 10 while i <= 100000: print(f'''with {i} steps: {trapezoidal_area(f, -5, 5, i)}''') i *= 10
113
1
'''simple docstring''' import qiskit def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> qiskit.result.counts.Counts: _a : Union[str, Any] = qiskit.Aer.get_backend('aer_simulator' ) # Create a Quantum Circuit acting on the q register _a : Optional[int] = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator _a : str = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(f"""Total count for various states are: {single_qubit_measure(1, 1)}""")
107
'''simple docstring''' import heapq def __lowerCamelCase ( lowerCAmelCase_ ) -> set[int]: _a : list[list] = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(lowerCAmelCase_ , [-1 * len(lowerCAmelCase_ ), (key, value)] ) # chosen_vertices = set of chosen vertices _a : str = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices _a : Optional[int] = heapq.heappop(lowerCAmelCase_ )[1][0] chosen_vertices.add(lowerCAmelCase_ ) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: _a : Dict = elem[1][1].index(lowerCAmelCase_ ) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(lowerCAmelCase_ ) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() __lowerCAmelCase = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(f"""Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}""")
107
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase = { '''configuration_groupvit''': [ '''GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GroupViTConfig''', '''GroupViTOnnxConfig''', '''GroupViTTextConfig''', '''GroupViTVisionConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GroupViTModel''', '''GroupViTPreTrainedModel''', '''GroupViTTextModel''', '''GroupViTVisionModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFGroupViTModel''', '''TFGroupViTPreTrainedModel''', '''TFGroupViTTextModel''', '''TFGroupViTVisionModel''', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
87
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase=False , __lowercase=False , __lowercase=False ) -> Optional[Any]: A: str = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""transformer.blocks.{i}.norm1.weight""", F"""vilt.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""transformer.blocks.{i}.norm1.bias""", F"""vilt.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append( (F"""transformer.blocks.{i}.attn.proj.weight""", F"""vilt.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append( (F"""transformer.blocks.{i}.attn.proj.bias""", F"""vilt.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""transformer.blocks.{i}.norm2.weight""", F"""vilt.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""transformer.blocks.{i}.norm2.bias""", F"""vilt.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append( (F"""transformer.blocks.{i}.mlp.fc1.weight""", F"""vilt.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""transformer.blocks.{i}.mlp.fc1.bias""", F"""vilt.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""transformer.blocks.{i}.mlp.fc2.weight""", F"""vilt.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""transformer.blocks.{i}.mlp.fc2.bias""", F"""vilt.encoder.layer.{i}.output.dense.bias""") ) # embeddings rename_keys.extend( [ # text embeddings ('''text_embeddings.word_embeddings.weight''', '''vilt.embeddings.text_embeddings.word_embeddings.weight'''), ( '''text_embeddings.position_embeddings.weight''', '''vilt.embeddings.text_embeddings.position_embeddings.weight''', ), ('''text_embeddings.position_ids''', '''vilt.embeddings.text_embeddings.position_ids'''), ( '''text_embeddings.token_type_embeddings.weight''', '''vilt.embeddings.text_embeddings.token_type_embeddings.weight''', ), ('''text_embeddings.LayerNorm.weight''', '''vilt.embeddings.text_embeddings.LayerNorm.weight'''), ('''text_embeddings.LayerNorm.bias''', '''vilt.embeddings.text_embeddings.LayerNorm.bias'''), # patch embeddings ('''transformer.cls_token''', '''vilt.embeddings.cls_token'''), ('''transformer.patch_embed.proj.weight''', '''vilt.embeddings.patch_embeddings.projection.weight'''), ('''transformer.patch_embed.proj.bias''', '''vilt.embeddings.patch_embeddings.projection.bias'''), ('''transformer.pos_embed''', '''vilt.embeddings.position_embeddings'''), # token type embeddings ('''token_type_embeddings.weight''', '''vilt.embeddings.token_type_embeddings.weight'''), ] ) # final layernorm + pooler rename_keys.extend( [ ('''transformer.norm.weight''', '''vilt.layernorm.weight'''), ('''transformer.norm.bias''', '''vilt.layernorm.bias'''), ('''pooler.dense.weight''', '''vilt.pooler.dense.weight'''), ('''pooler.dense.bias''', '''vilt.pooler.dense.bias'''), ] ) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ('''vqa_classifier.0.weight''', '''classifier.0.weight'''), ('''vqa_classifier.0.bias''', '''classifier.0.bias'''), ('''vqa_classifier.1.weight''', '''classifier.1.weight'''), ('''vqa_classifier.1.bias''', '''classifier.1.bias'''), ('''vqa_classifier.3.weight''', '''classifier.3.weight'''), ('''vqa_classifier.3.bias''', '''classifier.3.bias'''), ] ) elif nlvr_model: # classification head rename_keys.extend( [ ('''nlvr2_classifier.0.weight''', '''classifier.0.weight'''), ('''nlvr2_classifier.0.bias''', '''classifier.0.bias'''), ('''nlvr2_classifier.1.weight''', '''classifier.1.weight'''), ('''nlvr2_classifier.1.bias''', '''classifier.1.bias'''), ('''nlvr2_classifier.3.weight''', '''classifier.3.weight'''), ('''nlvr2_classifier.3.bias''', '''classifier.3.bias'''), ] ) else: pass return rename_keys def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> Any: for i in range(config.num_hidden_layers ): A: Tuple = '''vilt.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A: List[str] = state_dict.pop(F"""transformer.blocks.{i}.attn.qkv.weight""" ) A: Optional[Any] = state_dict.pop(F"""transformer.blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict A: Dict = in_proj_weight[ : config.hidden_size, : ] A: int = in_proj_bias[: config.hidden_size] A: Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A: int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A: Optional[int] = in_proj_weight[ -config.hidden_size :, : ] A: Optional[Any] = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE( __lowercase ) -> int: A: Optional[int] = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__lowercase , __lowercase ) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase ) -> int: A: List[Any] = dct.pop(__lowercase ) A: int = val @torch.no_grad() def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> str: A: Optional[Any] = ViltConfig(image_size=3_8_4 , patch_size=3_2 , tie_word_embeddings=__lowercase ) A: Tuple = False A: str = False A: List[Any] = False A: Optional[int] = False if "vqa" in checkpoint_url: A: Union[str, Any] = True A: Union[str, Any] = 3_1_2_9 A: List[Any] = '''huggingface/label-files''' A: Any = '''vqa2-id2label.json''' A: Optional[Any] = json.load(open(hf_hub_download(__lowercase , __lowercase , repo_type='''dataset''' ) , '''r''' ) ) A: Union[str, Any] = {int(__lowercase ): v for k, v in idalabel.items()} A: Any = idalabel A: Optional[Any] = {v: k for k, v in idalabel.items()} A: List[str] = ViltForQuestionAnswering(__lowercase ) elif "nlvr" in checkpoint_url: A: Dict = True A: str = 2 A: Union[str, Any] = {0: '''False''', 1: '''True'''} A: Any = {v: k for k, v in config.idalabel.items()} A: Optional[Any] = 3 A: Any = ViltForImagesAndTextClassification(__lowercase ) elif "irtr" in checkpoint_url: A: Tuple = True A: Optional[Any] = ViltForImageAndTextRetrieval(__lowercase ) elif "mlm_itm" in checkpoint_url: A: Tuple = True A: Optional[int] = ViltForMaskedLM(__lowercase ) else: raise ValueError('''Unknown model type''' ) # load state_dict of original model, remove and rename some keys A: int = torch.hub.load_state_dict_from_url(__lowercase , map_location='''cpu''' )['''state_dict'''] A: List[str] = create_rename_keys(__lowercase , __lowercase , __lowercase , __lowercase ) for src, dest in rename_keys: rename_key(__lowercase , __lowercase , __lowercase ) read_in_q_k_v(__lowercase , __lowercase ) if mlm_model or irtr_model: A: str = ['''itm_score.fc.weight''', '''itm_score.fc.bias'''] for k in ignore_keys: state_dict.pop(__lowercase , __lowercase ) # load state dict into HuggingFace model model.eval() if mlm_model: A , A: Union[str, Any] = model.load_state_dict(__lowercase , strict=__lowercase ) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(__lowercase ) # Define processor A: Optional[Any] = ViltImageProcessor(size=3_8_4 ) A: Dict = BertTokenizer.from_pretrained('''bert-base-uncased''' ) A: Optional[int] = ViltProcessor(__lowercase , __lowercase ) # Forward pass on example inputs (image + text) if nlvr_model: A: str = Image.open(requests.get('''https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg''' , stream=__lowercase ).raw ) A: List[str] = Image.open(requests.get('''https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg''' , stream=__lowercase ).raw ) A: Any = ( '''The left image contains twice the number of dogs as the right image, and at least two dogs in total are''' ''' standing.''' ) A: List[Any] = processor(__lowercase , __lowercase , return_tensors='''pt''' ) A: List[Any] = processor(__lowercase , __lowercase , return_tensors='''pt''' ) A: List[str] = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: A: Any = Image.open(requests.get('''http://images.cocodataset.org/val2017/000000039769.jpg''' , stream=__lowercase ).raw ) if mlm_model: A: Optional[int] = '''a bunch of [MASK] laying on a [MASK].''' else: A: Optional[int] = '''How many cats are there?''' A: Union[str, Any] = processor(__lowercase , __lowercase , return_tensors='''pt''' ) A: Any = model(**__lowercase ) # Verify outputs if mlm_model: A: Any = torch.Size([1, 1_1, 3_0_5_2_2] ) A: Tuple = torch.tensor([-1_2.5_0_6_1, -1_2.5_1_2_3, -1_2.5_1_7_4] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , __lowercase , atol=1E-4 ) # verify masked token prediction equals "cats" A: List[str] = outputs.logits[0, 4, :].argmax(-1 ).item() assert tokenizer.decode([predicted_id] ) == "cats" elif vqa_model: A: Any = torch.Size([1, 3_1_2_9] ) A: Optional[int] = torch.tensor([-1_5.9_4_9_5, -1_8.1_4_7_2, -1_0.3_0_4_1] ) assert torch.allclose(outputs.logits[0, :3] , __lowercase , atol=1E-4 ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , __lowercase , atol=1E-4 ) # verify vqa prediction equals "2" A: Dict = outputs.logits.argmax(-1 ).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: A: Union[str, Any] = torch.Size([1, 2] ) A: Optional[Any] = torch.tensor([-2.8_7_2_1, 2.1_2_9_1] ) assert torch.allclose(outputs.logits[0, :3] , __lowercase , atol=1E-4 ) assert outputs.logits.shape == expected_shape Path(__lowercase ).mkdir(exist_ok=__lowercase ) print(F"""Saving model and processor to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowercase ) processor.save_pretrained(__lowercase ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) UpperCamelCase = parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
319
0
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __UpperCAmelCase ( _UpperCamelCase ): '''simple docstring''' __lowercase : Union[str, Any] = ['image_processor', 'tokenizer'] __lowercase : int = 'CLIPImageProcessor' __lowercase : Union[str, Any] = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE ) -> List[Any]: A_ = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _SCREAMING_SNAKE_CASE , ) A_ = kwargs.pop('''feature_extractor''' ) A_ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __call__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE ) -> Union[str, Any]: if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: A_ = self.tokenizer(_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if images is not None: A_ = self.image_processor(_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if text is not None and images is not None: A_ = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_SCREAMING_SNAKE_CASE ) , tensor_type=_SCREAMING_SNAKE_CASE ) def __A ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Any: return self.tokenizer.batch_decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def __A ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> List[Any]: return self.tokenizer.decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) @property def __A ( self ) -> Optional[Any]: A_ = self.tokenizer.model_input_names A_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
367
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def _UpperCAmelCase ( ) -> Dict: A_ = ArgumentParser('''Accelerate CLI tool''', usage='''accelerate <command> [<args>]''', allow_abbrev=_UpperCamelCase ) A_ = parser.add_subparsers(help='''accelerate command helpers''' ) # Register commands get_config_parser(subparsers=_UpperCamelCase ) env_command_parser(subparsers=_UpperCamelCase ) launch_command_parser(subparsers=_UpperCamelCase ) tpu_command_parser(subparsers=_UpperCamelCase ) test_command_parser(subparsers=_UpperCamelCase ) # Let's go A_ = parser.parse_args() if not hasattr(_UpperCamelCase, '''func''' ): parser.print_help() exit(1 ) # Run args.func(_UpperCamelCase ) if __name__ == "__main__": main()
18
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class UpperCAmelCase : """simple docstring""" _UpperCAmelCase :Tuple = XGLMConfig _UpperCAmelCase :List[str] = {} _UpperCAmelCase :Tuple = "gelu" def __init__( self , _UpperCAmelCase , _UpperCAmelCase=14 , _UpperCAmelCase=7 , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=99 , _UpperCAmelCase=32 , _UpperCAmelCase=2 , _UpperCAmelCase=4 , _UpperCAmelCase=37 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=512 , _UpperCAmelCase=0.02 , ): lowercase__: str = parent lowercase__: str = batch_size lowercase__: Any = seq_length lowercase__: Any = is_training lowercase__: List[Any] = use_input_mask lowercase__: Optional[Any] = use_labels lowercase__: int = vocab_size lowercase__: int = d_model lowercase__: str = num_hidden_layers lowercase__: Dict = num_attention_heads lowercase__: Dict = ffn_dim lowercase__: Union[str, Any] = activation_function lowercase__: Union[str, Any] = activation_dropout lowercase__: Union[str, Any] = attention_dropout lowercase__: Optional[Any] = max_position_embeddings lowercase__: Dict = initializer_range lowercase__: str = None lowercase__: Any = 0 lowercase__: List[Any] = 2 lowercase__: Optional[int] = 1 def _snake_case ( self ): return XGLMConfig.from_pretrained('''facebook/xglm-564M''' ) def _snake_case ( self ): lowercase__: str = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) lowercase__: Optional[int] = None if self.use_input_mask: lowercase__: str = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__: List[Any] = self.get_config() lowercase__: Any = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def _snake_case ( self ): return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=_UpperCAmelCase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=_UpperCAmelCase , ) def _snake_case ( self ): lowercase__: List[Any] = self.prepare_config_and_inputs() ( ( lowercase__ ), ( lowercase__ ), ( lowercase__ ), ( lowercase__ ), ): Optional[int] = config_and_inputs lowercase__: Tuple = { '''input_ids''': input_ids, '''head_mask''': head_mask, } return config, inputs_dict @require_tf class UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): """simple docstring""" _UpperCAmelCase :Optional[int] = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () _UpperCAmelCase :Dict = (TFXGLMForCausalLM,) if is_tf_available() else () _UpperCAmelCase :Tuple = ( {"feature-extraction": TFXGLMModel, "text-generation": TFXGLMForCausalLM} if is_tf_available() else {} ) _UpperCAmelCase :List[str] = False _UpperCAmelCase :str = False _UpperCAmelCase :Union[str, Any] = False def _snake_case ( self ): lowercase__: Union[str, Any] = TFXGLMModelTester(self ) lowercase__: Any = ConfigTester(self , config_class=_UpperCAmelCase , n_embd=37 ) def _snake_case ( self ): self.config_tester.run_common_tests() @slow def _snake_case ( self ): for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__: Optional[int] = TFXGLMModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) @unittest.skip(reason='''Currently, model embeddings are going to undergo a major refactor.''' ) def _snake_case ( self ): super().test_resize_token_embeddings() @require_tf class UpperCAmelCase (unittest.TestCase ): """simple docstring""" @slow def _snake_case ( self , _UpperCAmelCase=True ): lowercase__: Union[str, Any] = TFXGLMForCausalLM.from_pretrained('''facebook/xglm-564M''' ) lowercase__: Any = tf.convert_to_tensor([[2, 268, 9865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off lowercase__: List[Any] = [2, 268, 9865, 67, 11, 1988, 57252, 9865, 5, 984, 67, 1988, 213838, 1658, 53, 70446, 33, 6657, 278, 1581] # fmt: on lowercase__: List[str] = model.generate(_UpperCAmelCase , do_sample=_UpperCAmelCase , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , _UpperCAmelCase ) @slow def _snake_case ( self ): lowercase__: Optional[Any] = XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) lowercase__: str = TFXGLMForCausalLM.from_pretrained('''facebook/xglm-564M''' ) tf.random.set_seed(0 ) lowercase__: str = tokenizer('''Today is a nice day and''' , return_tensors='''tf''' ) lowercase__: Tuple = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(''':/CPU:0''' ): lowercase__: Optional[Any] = model.generate(_UpperCAmelCase , do_sample=_UpperCAmelCase , seed=[7, 0] ) lowercase__: Optional[int] = tokenizer.decode(output_ids[0] , skip_special_tokens=_UpperCAmelCase ) lowercase__: Any = ( '''Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due''' ) self.assertEqual(_UpperCAmelCase , _UpperCAmelCase ) @slow def _snake_case ( self ): lowercase__: str = TFXGLMForCausalLM.from_pretrained('''facebook/xglm-564M''' ) lowercase__: int = XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) lowercase__: Any = '''left''' # use different length sentences to test batching lowercase__: List[Any] = [ '''This is an extremelly long sentence that only exists to test the ability of the model to cope with ''' '''left-padding, such as in batched generation. The output for the sequence below should be the same ''' '''regardless of whether left padding is applied or not. When''', '''Hello, my dog is a little''', ] lowercase__: Union[str, Any] = tokenizer(_UpperCAmelCase , return_tensors='''tf''' , padding=_UpperCAmelCase ) lowercase__: Union[str, Any] = inputs['''input_ids'''] lowercase__: Dict = model.generate(input_ids=_UpperCAmelCase , attention_mask=inputs['''attention_mask'''] , max_new_tokens=12 ) lowercase__: int = tokenizer(sentences[0] , return_tensors='''tf''' ).input_ids lowercase__: Tuple = model.generate(input_ids=_UpperCAmelCase , max_new_tokens=12 ) lowercase__: Any = tokenizer(sentences[1] , return_tensors='''tf''' ).input_ids lowercase__: List[Any] = model.generate(input_ids=_UpperCAmelCase , max_new_tokens=12 ) lowercase__: Dict = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase ) lowercase__: Union[str, Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=_UpperCAmelCase ) lowercase__: int = tokenizer.decode(output_padded[0] , skip_special_tokens=_UpperCAmelCase ) lowercase__: Dict = [ '''This is an extremelly long sentence that only exists to test the ability of the model to cope with ''' '''left-padding, such as in batched generation. The output for the sequence below should be the same ''' '''regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ''' '''a single''', '''Hello, my dog is a little bit of a shy one, but he is very friendly''', ] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , [non_padded_sentence, padded_sentence] )
177
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A = { "configuration_lilt": ["LILT_PRETRAINED_CONFIG_ARCHIVE_MAP", "LiltConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "LILT_PRETRAINED_MODEL_ARCHIVE_LIST", "LiltForQuestionAnswering", "LiltForSequenceClassification", "LiltForTokenClassification", "LiltModel", "LiltPreTrainedModel", ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
177
1
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL __a = logging.get_logger(__name__) def lowerCamelCase__ ( _lowercase , _lowercase , _lowercase , _lowercase ): '''simple docstring''' def constraint_to_multiple_of(_lowercase , _lowercase , _lowercase=0 , _lowercase=None ): UpperCAmelCase_ : Any = round(val / multiple ) * multiple if max_val is not None and x > max_val: UpperCAmelCase_ : Optional[int] = math.floor(val / multiple ) * multiple if x < min_val: UpperCAmelCase_ : List[str] = math.ceil(val / multiple ) * multiple return x UpperCAmelCase_ : Any = (output_size, output_size) if isinstance(_lowercase , _lowercase ) else output_size UpperCAmelCase_, UpperCAmelCase_ : Optional[int] = get_image_size(_lowercase ) UpperCAmelCase_, UpperCAmelCase_ : str = output_size # determine new height and width UpperCAmelCase_ : Union[str, Any] = output_height / input_height UpperCAmelCase_ : List[Any] = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width UpperCAmelCase_ : Optional[int] = scale_width else: # fit height UpperCAmelCase_ : int = scale_height UpperCAmelCase_ : List[Any] = constraint_to_multiple_of(scale_height * input_height , multiple=_lowercase ) UpperCAmelCase_ : int = constraint_to_multiple_of(scale_width * input_width , multiple=_lowercase ) return (new_height, new_width) class __a( _a ): """simple docstring""" lowerCAmelCase = ['''pixel_values'''] def __init__( self ,_SCREAMING_SNAKE_CASE = True ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = PILImageResampling.BILINEAR ,_SCREAMING_SNAKE_CASE = False ,_SCREAMING_SNAKE_CASE = 1 ,_SCREAMING_SNAKE_CASE = True ,_SCREAMING_SNAKE_CASE = 1 / 255 ,_SCREAMING_SNAKE_CASE = True ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,**_SCREAMING_SNAKE_CASE ,) -> None: super().__init__(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = size if size is not None else {'''height''': 384, '''width''': 384} UpperCAmelCase_ : Dict = get_size_dict(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = do_resize UpperCAmelCase_ : str = size UpperCAmelCase_ : Dict = keep_aspect_ratio UpperCAmelCase_ : Dict = ensure_multiple_of UpperCAmelCase_ : int = resample UpperCAmelCase_ : Optional[int] = do_rescale UpperCAmelCase_ : Union[str, Any] = rescale_factor UpperCAmelCase_ : Optional[Any] = do_normalize UpperCAmelCase_ : str = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_ : Dict = image_std if image_std is not None else IMAGENET_STANDARD_STD def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = False ,_SCREAMING_SNAKE_CASE = 1 ,_SCREAMING_SNAKE_CASE = PILImageResampling.BICUBIC ,_SCREAMING_SNAKE_CASE = None ,**_SCREAMING_SNAKE_CASE ,) -> np.ndarray: UpperCAmelCase_ : Optional[Any] = get_size_dict(_SCREAMING_SNAKE_CASE ) if "height" not in size or "width" not in size: raise ValueError(f'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) UpperCAmelCase_ : List[Any] = get_resize_output_image_size( _SCREAMING_SNAKE_CASE ,output_size=(size['''height'''], size['''width''']) ,keep_aspect_ratio=_SCREAMING_SNAKE_CASE ,multiple=_SCREAMING_SNAKE_CASE ,) return resize(_SCREAMING_SNAKE_CASE ,size=_SCREAMING_SNAKE_CASE ,resample=_SCREAMING_SNAKE_CASE ,data_format=_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ,**_SCREAMING_SNAKE_CASE ,) -> Any: return rescale(_SCREAMING_SNAKE_CASE ,scale=_SCREAMING_SNAKE_CASE ,data_format=_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ,**_SCREAMING_SNAKE_CASE ,) -> np.ndarray: return normalize(_SCREAMING_SNAKE_CASE ,mean=_SCREAMING_SNAKE_CASE ,std=_SCREAMING_SNAKE_CASE ,data_format=_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = ChannelDimension.FIRST ,**_SCREAMING_SNAKE_CASE ,) -> PIL.Image.Image: UpperCAmelCase_ : Optional[int] = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_ : Tuple = size if size is not None else self.size UpperCAmelCase_ : Union[str, Any] = get_size_dict(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio UpperCAmelCase_ : str = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of UpperCAmelCase_ : Optional[int] = resample if resample is not None else self.resample UpperCAmelCase_ : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_ : Any = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_ : List[str] = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_ : Union[str, Any] = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_ : Optional[Any] = image_std if image_std is not None else self.image_std UpperCAmelCase_ : str = make_list_of_images(_SCREAMING_SNAKE_CASE ) if not valid_images(_SCREAMING_SNAKE_CASE ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. UpperCAmelCase_ : List[Any] = [to_numpy_array(_SCREAMING_SNAKE_CASE ) for image in images] if do_resize: UpperCAmelCase_ : Tuple = [self.resize(image=_SCREAMING_SNAKE_CASE ,size=_SCREAMING_SNAKE_CASE ,resample=_SCREAMING_SNAKE_CASE ) for image in images] if do_rescale: UpperCAmelCase_ : List[Any] = [self.rescale(image=_SCREAMING_SNAKE_CASE ,scale=_SCREAMING_SNAKE_CASE ) for image in images] if do_normalize: UpperCAmelCase_ : int = [self.normalize(image=_SCREAMING_SNAKE_CASE ,mean=_SCREAMING_SNAKE_CASE ,std=_SCREAMING_SNAKE_CASE ) for image in images] UpperCAmelCase_ : Tuple = [to_channel_dimension_format(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) for image in images] UpperCAmelCase_ : Dict = {'''pixel_values''': images} return BatchFeature(data=_SCREAMING_SNAKE_CASE ,tensor_type=_SCREAMING_SNAKE_CASE ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ) -> Dict: UpperCAmelCase_ : Union[str, Any] = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(_SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : List[Any] = target_sizes.numpy() UpperCAmelCase_ : List[str] = [] for idx in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : Optional[int] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) ,size=target_sizes[idx] ,mode='''bilinear''' ,align_corners=_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Union[str, Any] = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ : Optional[Any] = logits.argmax(dim=1 ) UpperCAmelCase_ : str = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
235
import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def lowerCamelCase__ ( _lowercase ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = SwinConfig() UpperCAmelCase_ : Dict = swin_name.split('''_''' ) UpperCAmelCase_ : List[Any] = name_split[1] UpperCAmelCase_ : Optional[Any] = int(name_split[4] ) UpperCAmelCase_ : Union[str, Any] = int(name_split[3][-1] ) if model_size == "tiny": UpperCAmelCase_ : Tuple = 96 UpperCAmelCase_ : Optional[int] = (2, 2, 6, 2) UpperCAmelCase_ : str = (3, 6, 12, 24) elif model_size == "small": UpperCAmelCase_ : Dict = 96 UpperCAmelCase_ : str = (2, 2, 18, 2) UpperCAmelCase_ : List[str] = (3, 6, 12, 24) elif model_size == "base": UpperCAmelCase_ : Tuple = 128 UpperCAmelCase_ : Optional[int] = (2, 2, 18, 2) UpperCAmelCase_ : Optional[int] = (4, 8, 16, 32) else: UpperCAmelCase_ : List[str] = 192 UpperCAmelCase_ : str = (2, 2, 18, 2) UpperCAmelCase_ : Union[str, Any] = (6, 12, 24, 48) if "in22k" in swin_name: UpperCAmelCase_ : int = 21841 else: UpperCAmelCase_ : Tuple = 1000 UpperCAmelCase_ : Any = '''huggingface/label-files''' UpperCAmelCase_ : int = '''imagenet-1k-id2label.json''' UpperCAmelCase_ : str = json.load(open(hf_hub_download(_lowercase , _lowercase , repo_type='''dataset''' ) , '''r''' ) ) UpperCAmelCase_ : Optional[int] = {int(_lowercase ): v for k, v in idalabel.items()} UpperCAmelCase_ : List[Any] = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : Tuple = img_size UpperCAmelCase_ : List[Any] = num_classes UpperCAmelCase_ : str = embed_dim UpperCAmelCase_ : Optional[int] = depths UpperCAmelCase_ : Dict = num_heads UpperCAmelCase_ : Optional[int] = window_size return config def lowerCamelCase__ ( _lowercase ): '''simple docstring''' if "patch_embed.proj" in name: UpperCAmelCase_ : str = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: UpperCAmelCase_ : List[Any] = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if "layers" in name: UpperCAmelCase_ : Tuple = '''encoder.''' + name if "attn.proj" in name: UpperCAmelCase_ : Tuple = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: UpperCAmelCase_ : int = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: UpperCAmelCase_ : Union[str, Any] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: UpperCAmelCase_ : List[Any] = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: UpperCAmelCase_ : Tuple = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: UpperCAmelCase_ : Optional[int] = name.replace('''mlp.fc2''' , '''output.dense''' ) if name == "norm.weight": UpperCAmelCase_ : Optional[Any] = '''layernorm.weight''' if name == "norm.bias": UpperCAmelCase_ : List[str] = '''layernorm.bias''' if "head" in name: UpperCAmelCase_ : Union[str, Any] = name.replace('''head''' , '''classifier''' ) else: UpperCAmelCase_ : Union[str, Any] = '''swin.''' + name return name def lowerCamelCase__ ( _lowercase , _lowercase ): '''simple docstring''' for key in orig_state_dict.copy().keys(): UpperCAmelCase_ : str = orig_state_dict.pop(_lowercase ) if "mask" in key: continue elif "qkv" in key: UpperCAmelCase_ : str = key.split('''.''' ) UpperCAmelCase_ : str = int(key_split[1] ) UpperCAmelCase_ : Optional[Any] = int(key_split[3] ) UpperCAmelCase_ : Optional[int] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: UpperCAmelCase_ : Any = val[:dim, :] UpperCAmelCase_ : Optional[int] = val[ dim : dim * 2, : ] UpperCAmelCase_ : List[str] = val[-dim:, :] else: UpperCAmelCase_ : Any = val[ :dim ] UpperCAmelCase_ : Optional[int] = val[ dim : dim * 2 ] UpperCAmelCase_ : Union[str, Any] = val[ -dim: ] else: UpperCAmelCase_ : Optional[Any] = val return orig_state_dict def lowerCamelCase__ ( _lowercase , _lowercase ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = timm.create_model(_lowercase , pretrained=_lowercase ) timm_model.eval() UpperCAmelCase_ : Any = get_swin_config(_lowercase ) UpperCAmelCase_ : Union[str, Any] = SwinForImageClassification(_lowercase ) model.eval() UpperCAmelCase_ : Tuple = convert_state_dict(timm_model.state_dict() , _lowercase ) model.load_state_dict(_lowercase ) UpperCAmelCase_ : List[Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' UpperCAmelCase_ : int = AutoImageProcessor.from_pretrained('''microsoft/{}'''.format(swin_name.replace('''_''' , '''-''' ) ) ) UpperCAmelCase_ : Optional[Any] = Image.open(requests.get(_lowercase , stream=_lowercase ).raw ) UpperCAmelCase_ : Union[str, Any] = image_processor(images=_lowercase , return_tensors='''pt''' ) UpperCAmelCase_ : Optional[Any] = timm_model(inputs['''pixel_values'''] ) UpperCAmelCase_ : Dict = model(**_lowercase ).logits assert torch.allclose(_lowercase , _lowercase , atol=1E-3 ) print(f'''Saving model {swin_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_lowercase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(_lowercase ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swin_name', default='swin_tiny_patch4_window7_224', type=str, help='Name of the Swin timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) __a = parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
235
1
'''simple docstring''' from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def UpperCamelCase_ ( _UpperCAmelCase : int ) -> bool: """simple docstring""" _UpperCAmelCase : int = int(number**0.5 ) return number == sq * sq def UpperCamelCase_ ( _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> tuple[int, int]: """simple docstring""" _UpperCAmelCase : int = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den _UpperCAmelCase : int = x_den * y_den * z_den _UpperCAmelCase : int = gcd(_UpperCAmelCase , _UpperCAmelCase ) top //= hcf bottom //= hcf return top, bottom def UpperCamelCase_ ( _UpperCAmelCase : int = 35 ) -> int: """simple docstring""" _UpperCAmelCase : set = set() _UpperCAmelCase : int _UpperCAmelCase : Fraction = Fraction(0 ) _UpperCAmelCase : tuple[int, int] for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 _UpperCAmelCase : Dict = x_num * y_den + x_den * y_num _UpperCAmelCase : Any = x_den * y_den _UpperCAmelCase : List[str] = gcd(_UpperCAmelCase , _UpperCAmelCase ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _UpperCAmelCase : Optional[Any] = add_three( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) unique_s.add(_UpperCAmelCase ) # n=2 _UpperCAmelCase : int = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) _UpperCAmelCase : int = x_den * x_den * y_den * y_den if is_sq(_UpperCAmelCase ) and is_sq(_UpperCAmelCase ): _UpperCAmelCase : List[str] = int(sqrt(_UpperCAmelCase ) ) _UpperCAmelCase : List[str] = int(sqrt(_UpperCAmelCase ) ) _UpperCAmelCase : Tuple = gcd(_UpperCAmelCase , _UpperCAmelCase ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _UpperCAmelCase : List[Any] = add_three( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) unique_s.add(_UpperCAmelCase ) # n=-1 _UpperCAmelCase : Dict = x_num * y_num _UpperCAmelCase : str = x_den * y_num + x_num * y_den _UpperCAmelCase : Dict = gcd(_UpperCAmelCase , _UpperCAmelCase ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _UpperCAmelCase : Optional[Any] = add_three( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) unique_s.add(_UpperCAmelCase ) # n=2 _UpperCAmelCase : Any = x_num * x_num * y_num * y_num _UpperCAmelCase : List[Any] = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(_UpperCAmelCase ) and is_sq(_UpperCAmelCase ): _UpperCAmelCase : int = int(sqrt(_UpperCAmelCase ) ) _UpperCAmelCase : Dict = int(sqrt(_UpperCAmelCase ) ) _UpperCAmelCase : Dict = gcd(_UpperCAmelCase , _UpperCAmelCase ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _UpperCAmelCase : str = add_three( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) unique_s.add(_UpperCAmelCase ) for num, den in unique_s: total += Fraction(_UpperCAmelCase , _UpperCAmelCase ) return total.denominator + total.numerator if __name__ == "__main__": print(F'{solution() = }')
31
'''simple docstring''' __SCREAMING_SNAKE_CASE : Dict = 8.3_1_4_4_6_2 # Unit - J mol-1 K-1 def UpperCamelCase_ ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: """simple docstring""" if moles < 0 or kelvin < 0 or volume < 0: raise ValueError("Invalid inputs. Enter positive value." ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / volume def UpperCamelCase_ ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: """simple docstring""" if moles < 0 or kelvin < 0 or pressure < 0: raise ValueError("Invalid inputs. Enter positive value." ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / pressure if __name__ == "__main__": from doctest import testmod testmod()
31
1
'''simple docstring''' import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class a__: @staticmethod def lowercase_ ( *__snake_case : List[Any] , **__snake_case : Tuple ): pass def lowerCamelCase__ ( _A ): return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. lowerCAmelCase: List[Any] = ( 'https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png' ) @is_pipeline_test @require_torch @require_vision class a__( unittest.TestCase ): lowercase__ = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def lowercase_ ( self : Optional[Any] , __snake_case : List[str] , __snake_case : List[str] , __snake_case : Any ): a : Dict = pipeline( 'document-question-answering' , model=__snake_case , tokenizer=__snake_case , image_processor=__snake_case ) a : Tuple = INVOICE_URL a : str = list(zip(*apply_tesseract(load_image(__snake_case ) , __snake_case , '' ) ) ) a : int = 'What is the placebo?' a : Tuple = [ { 'image': load_image(__snake_case ), 'question': question, }, { 'image': image, 'question': question, }, { 'image': image, 'question': question, 'word_boxes': word_boxes, }, ] return dqa_pipeline, examples def lowercase_ ( self : Optional[int] , __snake_case : Any , __snake_case : str ): a : Any = dqa_pipeline(__snake_case , top_k=2 ) self.assertEqual( __snake_case , [ [ {'score': ANY(__snake_case ), 'answer': ANY(__snake_case ), 'start': ANY(__snake_case ), 'end': ANY(__snake_case )}, {'score': ANY(__snake_case ), 'answer': ANY(__snake_case ), 'start': ANY(__snake_case ), 'end': ANY(__snake_case )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def lowercase_ ( self : Optional[int] ): a : List[Any] = pipeline('document-question-answering' , model='hf-internal-testing/tiny-random-layoutlmv2' ) a : Tuple = INVOICE_URL a : Optional[int] = 'How many cats are there?' a : List[Any] = [ {'score': 0.0001, 'answer': 'oy 2312/2019', 'start': 38, 'end': 39}, {'score': 0.0001, 'answer': 'oy 2312/2019 DUE', 'start': 38, 'end': 40}, ] a : List[Any] = dqa_pipeline(image=__snake_case , question=__snake_case , top_k=2 ) self.assertEqual(nested_simplify(__snake_case , decimals=4 ) , __snake_case ) a : List[str] = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual(nested_simplify(__snake_case , decimals=4 ) , __snake_case ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably a : Tuple = './tests/fixtures/tests_samples/COCO/000000039769.png' a : int = dqa_pipeline(image=__snake_case , question=__snake_case , top_k=2 ) self.assertEqual(__snake_case , [] ) # We can optionnally pass directly the words and bounding boxes a : List[Any] = './tests/fixtures/tests_samples/COCO/000000039769.png' a : Tuple = [] a : List[str] = [] a : Dict = dqa_pipeline(image=__snake_case , question=__snake_case , words=__snake_case , boxes=__snake_case , top_k=2 ) self.assertEqual(__snake_case , [] ) @slow @require_torch @require_detectrona @require_pytesseract def lowercase_ ( self : int ): a : List[str] = pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , ) a : Dict = INVOICE_URL a : List[Any] = 'What is the invoice number?' a : List[str] = dqa_pipeline(image=__snake_case , question=__snake_case , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.9944, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0009, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) a : List[str] = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.9944, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0009, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) a : Union[str, Any] = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ [ {'score': 0.9944, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0009, 'answer': 'us-001', 'start': 16, 'end': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def lowercase_ ( self : Optional[Any] ): a : Optional[int] = pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , max_seq_len=50 , ) a : str = INVOICE_URL a : List[Any] = 'What is the invoice number?' a : Union[str, Any] = dqa_pipeline(image=__snake_case , question=__snake_case , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.9974, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9948, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) a : Tuple = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.9974, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9948, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) a : str = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ [ {'score': 0.9974, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9948, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def lowercase_ ( self : Optional[int] ): a : List[str] = AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=__snake_case ) a : List[str] = pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=__snake_case , revision='3dc6de3' , ) a : Tuple = INVOICE_URL a : Union[str, Any] = 'What is the invoice number?' a : Optional[int] = dqa_pipeline(image=__snake_case , question=__snake_case , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.4251, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0819, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) a : List[Any] = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.4251, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0819, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) a : List[Any] = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ [ {'score': 0.4251, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0819, 'answer': '1110212019', 'start': 23, 'end': 23}, ] ] * 2 , ) a : Tuple = list(zip(*apply_tesseract(load_image(__snake_case ) , __snake_case , '' ) ) ) # This model should also work if `image` is set to None a : int = dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.4251, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0819, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def lowercase_ ( self : Optional[Any] ): a : str = AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=__snake_case ) a : List[str] = pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=__snake_case , revision='3dc6de3' , max_seq_len=50 , ) a : List[Any] = INVOICE_URL a : List[Any] = 'What is the invoice number?' a : str = dqa_pipeline(image=__snake_case , question=__snake_case , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.9999, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9998, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) a : Dict = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ [ {'score': 0.9999, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9998, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) a : int = list(zip(*apply_tesseract(load_image(__snake_case ) , __snake_case , '' ) ) ) # This model should also work if `image` is set to None a : int = dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(__snake_case , decimals=4 ) , [ {'score': 0.9999, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9998, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) @slow @require_torch def lowercase_ ( self : int ): a : str = pipeline( 'document-question-answering' , model='naver-clova-ix/donut-base-finetuned-docvqa' , tokenizer=AutoTokenizer.from_pretrained('naver-clova-ix/donut-base-finetuned-docvqa' ) , feature_extractor='naver-clova-ix/donut-base-finetuned-docvqa' , ) a : int = INVOICE_URL a : List[str] = 'What is the invoice number?' a : Dict = dqa_pipeline(image=__snake_case , question=__snake_case , top_k=2 ) self.assertEqual(nested_simplify(__snake_case , decimals=4 ) , [{'answer': 'us-001'}] ) @require_tf @unittest.skip('Document question answering not implemented in TF' ) def lowercase_ ( self : int ): pass
96
'''simple docstring''' from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class a__( lowerCamelCase__ ): lowercase__ = 42 @flax_register_to_config class a__( nn.Module , lowerCamelCase__ , lowerCamelCase__ ): lowercase__ = 32 lowercase__ = 4 lowercase__ = 4 lowercase__ = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) lowercase__ = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") lowercase__ = False lowercase__ = (3_20, 6_40, 12_80, 12_80) lowercase__ = 2 lowercase__ = 8 lowercase__ = None lowercase__ = 12_80 lowercase__ = 0.0 lowercase__ = False lowercase__ = jnp.floataa lowercase__ = True lowercase__ = 0 lowercase__ = False def lowercase_ ( self : List[Any] , __snake_case : jax.random.KeyArray ): # init input tensors a : Dict = (1, self.in_channels, self.sample_size, self.sample_size) a : str = jnp.zeros(__snake_case , dtype=jnp.floataa ) a : Union[str, Any] = jnp.ones((1,) , dtype=jnp.intaa ) a : List[Any] = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) a , a : Optional[int] = jax.random.split(__snake_case ) a : Any = {'params': params_rng, 'dropout': dropout_rng} return self.init(__snake_case , __snake_case , __snake_case , __snake_case )["params"] def lowercase_ ( self : Union[str, Any] ): a : int = self.block_out_channels a : Tuple = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( 'At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19.' ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. a : str = self.num_attention_heads or self.attention_head_dim # input a : List[Any] = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time a : List[str] = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) a : Optional[int] = FlaxTimestepEmbedding(__snake_case , dtype=self.dtype ) a : Optional[Any] = self.only_cross_attention if isinstance(__snake_case , __snake_case ): a : int = (only_cross_attention,) * len(self.down_block_types ) if isinstance(__snake_case , __snake_case ): a : str = (num_attention_heads,) * len(self.down_block_types ) # down a : int = [] a : Dict = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): a : List[Any] = output_channel a : Dict = block_out_channels[i] a : Union[str, Any] = i == len(__snake_case ) - 1 if down_block_type == "CrossAttnDownBlock2D": a : Dict = FlaxCrossAttnDownBlockaD( in_channels=__snake_case , out_channels=__snake_case , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: a : List[str] = FlaxDownBlockaD( in_channels=__snake_case , out_channels=__snake_case , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(__snake_case ) a : Dict = down_blocks # mid a : Union[str, Any] = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up a : List[Any] = [] a : Dict = list(reversed(__snake_case ) ) a : List[str] = list(reversed(__snake_case ) ) a : List[str] = list(reversed(__snake_case ) ) a : Optional[int] = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): a : Dict = output_channel a : Dict = reversed_block_out_channels[i] a : Dict = reversed_block_out_channels[min(i + 1 , len(__snake_case ) - 1 )] a : List[Any] = i == len(__snake_case ) - 1 if up_block_type == "CrossAttnUpBlock2D": a : List[Any] = FlaxCrossAttnUpBlockaD( in_channels=__snake_case , out_channels=__snake_case , prev_output_channel=__snake_case , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: a : Dict = FlaxUpBlockaD( in_channels=__snake_case , out_channels=__snake_case , prev_output_channel=__snake_case , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(__snake_case ) a : Dict = output_channel a : int = up_blocks # out a : int = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) a : Any = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Optional[Any] , __snake_case : Dict , __snake_case : List[str] , __snake_case : Tuple , __snake_case : Dict=None , __snake_case : Any=None , __snake_case : bool = True , __snake_case : bool = False , ): # 1. time if not isinstance(__snake_case , jnp.ndarray ): a : Tuple = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(__snake_case , jnp.ndarray ) and len(timesteps.shape ) == 0: a : Tuple = timesteps.astype(dtype=jnp.floataa ) a : Optional[int] = jnp.expand_dims(__snake_case , 0 ) a : Optional[Any] = self.time_proj(__snake_case ) a : Optional[int] = self.time_embedding(__snake_case ) # 2. pre-process a : str = jnp.transpose(__snake_case , (0, 2, 3, 1) ) a : Optional[int] = self.conv_in(__snake_case ) # 3. down a : Optional[Any] = (sample,) for down_block in self.down_blocks: if isinstance(__snake_case , __snake_case ): a , a : Any = down_block(__snake_case , __snake_case , __snake_case , deterministic=not train ) else: a , a : Optional[Any] = down_block(__snake_case , __snake_case , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: a : Tuple = () for down_block_res_sample, down_block_additional_residual in zip( __snake_case , __snake_case ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) a : Tuple = new_down_block_res_samples # 4. mid a : int = self.mid_block(__snake_case , __snake_case , __snake_case , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: a : int = down_block_res_samples[-(self.layers_per_block + 1) :] a : List[Any] = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(__snake_case , __snake_case ): a : Optional[int] = up_block( __snake_case , temb=__snake_case , encoder_hidden_states=__snake_case , res_hidden_states_tuple=__snake_case , deterministic=not train , ) else: a : Optional[Any] = up_block(__snake_case , temb=__snake_case , res_hidden_states_tuple=__snake_case , deterministic=not train ) # 6. post-process a : Any = self.conv_norm_out(__snake_case ) a : List[str] = nn.silu(__snake_case ) a : Optional[int] = self.conv_out(__snake_case ) a : Any = jnp.transpose(__snake_case , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=__snake_case )
96
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _UpperCAmelCase : int = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Any = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys _UpperCAmelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration _UpperCAmelCase = [ # tf -> hf ("""/""", """."""), ("""layer_""", """layers."""), ("""kernel""", """weight"""), ("""beta""", """bias"""), ("""gamma""", """weight"""), ("""pegasus""", """model"""), ] _UpperCAmelCase = [ (""".output.dense""", """.fc2"""), ("""intermediate.LayerNorm""", """final_layer_norm"""), ("""intermediate.dense""", """fc1"""), ] _UpperCAmelCase = ( INIT_COMMON + [ ("""attention.self.LayerNorm""", """self_attn_layer_norm"""), ("""attention.output.dense""", """self_attn.out_proj"""), ("""attention.self""", """self_attn"""), ("""attention.encdec.LayerNorm""", """encoder_attn_layer_norm"""), ("""attention.encdec_output.dense""", """encoder_attn.out_proj"""), ("""attention.encdec""", """encoder_attn"""), ("""key""", """k_proj"""), ("""value""", """v_proj"""), ("""query""", """q_proj"""), ("""decoder.LayerNorm""", """decoder.layernorm_embedding"""), ] + END_COMMON ) _UpperCAmelCase = ( INIT_COMMON + [ ("""embeddings.word_embeddings""", """shared.weight"""), ("""embeddings.position_embeddings""", """embed_positions.weight"""), ("""attention.self.LayerNorm""", """self_attn_layer_norm"""), ("""attention.output.dense""", """self_attn.output"""), ("""attention.self""", """self_attn.self"""), ("""encoder.LayerNorm""", """encoder.layernorm_embedding"""), ] + END_COMMON ) _UpperCAmelCase = [ """encdec/key/bias""", """encdec/query/bias""", """encdec/value/bias""", """self/key/bias""", """self/query/bias""", """self/value/bias""", """encdec_output/dense/bias""", """attention/output/dense/bias""", ] def UpperCamelCase ( __lowercase : Optional[Any] ,__lowercase : Tuple ): '''simple docstring''' for tf_name, hf_name in patterns: A_ : Tuple = k.replace(__lowercase ,__lowercase ) return k def UpperCamelCase ( __lowercase : dict ,__lowercase : dict ): '''simple docstring''' A_ : int = BigBirdPegasusConfig(**__lowercase ) A_ : Any = BigBirdPegasusForConditionalGeneration(__lowercase ) A_ : Union[str, Any] = torch_model.state_dict() A_ : Any = {} # separating decoder weights A_ : Any = {k: tf_weights[k] for k in tf_weights if k.startswith('pegasus/decoder' )} A_ : str = {k: tf_weights[k] for k in tf_weights if not k.startswith('pegasus/decoder' )} for k, v in tqdm(decoder_weights.items() ,'tf -> hf conversion' ): A_ : Optional[int] = [k.endswith(__lowercase ) for ending in KEYS_TO_IGNORE] if any(__lowercase ): continue A_ : Optional[Any] = DECODER_PATTERNS A_ : Tuple = rename_state_dict_key(__lowercase ,__lowercase ) if new_k not in state_dict: raise ValueError(f'''could not find new key {new_k} in state dict. (converted from {k})''' ) if any(True if i in k else False for i in ['dense', 'query', 'key', 'value'] ): A_ : Any = v.T A_ : Any = torch.from_numpy(__lowercase ) assert v.shape == state_dict[new_k].shape, f'''{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}''' for k, v in tqdm(remaining_weights.items() ,'tf -> hf conversion' ): A_ : int = [k.endswith(__lowercase ) for ending in KEYS_TO_IGNORE] if any(__lowercase ): continue A_ : Any = REMAINING_PATTERNS A_ : List[str] = rename_state_dict_key(__lowercase ,__lowercase ) if new_k not in state_dict and k != "pegasus/embeddings/position_embeddings": raise ValueError(f'''could not find new key {new_k} in state dict. (converted from {k})''' ) if any(True if i in k else False for i in ['dense', 'query', 'key', 'value'] ): A_ : int = v.T A_ : Dict = torch.from_numpy(__lowercase ) if k != "pegasus/embeddings/position_embeddings": assert v.shape == state_dict[new_k].shape, f'''{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}''' A_ : Optional[int] = mapping['model.embed_positions.weight'] A_ : Tuple = mapping.pop('model.embed_positions.weight' ) A_ , A_ : Optional[Any] = torch_model.load_state_dict(__lowercase ,strict=__lowercase ) A_ : Optional[int] = [ k for k in missing if k not in [ 'final_logits_bias', 'model.encoder.embed_tokens.weight', 'model.decoder.embed_tokens.weight', 'lm_head.weight', ] ] assert unexpected_missing == [], f'''no matches found for the following torch keys {unexpected_missing}''' assert extra == [], f'''no matches found for the following tf keys {extra}''' return torch_model def UpperCamelCase ( __lowercase : Union[str, Any] ): '''simple docstring''' A_ : str = tf.train.list_variables(__lowercase ) A_ : Union[str, Any] = {} A_ : Optional[Any] = ['global_step'] for name, shape in tqdm(__lowercase ,desc='converting tf checkpoint to dict' ): A_ : Union[str, Any] = any(pat in name for pat in ignore_name ) if skip_key: continue A_ : Tuple = tf.train.load_variable(__lowercase ,__lowercase ) A_ : Dict = array return tf_weights def UpperCamelCase ( __lowercase : str ,__lowercase : str ,__lowercase : dict ): '''simple docstring''' A_ : Optional[Any] = get_tf_weights_as_numpy(__lowercase ) A_ : Dict = convert_bigbird_pegasus(__lowercase ,__lowercase ) torch_model.save_pretrained(__lowercase ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("""--tf_ckpt_path""", type=str, help="""passed to tf.train.list_variables""") parser.add_argument("""--save_dir""", default=None, type=str, help="""Path to the output PyTorch model.""") _UpperCAmelCase = parser.parse_args() _UpperCAmelCase = {} convert_bigbird_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir, config_update=config_update)
140
0
def A ( a_ ,a_ ) -> str: __UpperCamelCase : Union[str, Any] ='' for i in table: res += inp[i - 1] return res def A ( a_ ) -> str: return data[1:] + data[0] def A ( a_ ,a_ ) -> Dict: __UpperCamelCase : int ='' for i in range(len(a_ ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def A ( a_ ,a_ ) -> Tuple: __UpperCamelCase : List[str] =int('0b' + data[0] + data[-1] ,2 ) __UpperCamelCase : List[str] =int('0b' + data[1:3] ,2 ) return bin(s[row][col] )[2:] def A ( a_ ,a_ ,a_ ,a_ ,a_ ) -> Optional[int]: __UpperCamelCase : List[Any] =message[:4] __UpperCamelCase : Tuple =message[4:] __UpperCamelCase : Any =apply_table(a_ ,a_ ) __UpperCamelCase : List[Any] =xor(a_ ,a_ ) __UpperCamelCase : Union[str, Any] =apply_sbox(a_ ,temp[:4] ) # noqa: E741 __UpperCamelCase : str =apply_sbox(a_ ,temp[4:] ) __UpperCamelCase : Any ='0' * (2 - len(a_ )) + l # noqa: E741 __UpperCamelCase : Union[str, Any] ='0' * (2 - len(a_ )) + r __UpperCamelCase : Union[str, Any] =apply_table(l + r ,a_ ) __UpperCamelCase : Optional[Any] =xor(a_ ,a_ ) return temp + right if __name__ == "__main__": A_ :int = input('''Enter 10 bit key: ''') A_ :Tuple = input('''Enter 8 bit message: ''') A_ :int = [6, 3, 7, 4, 8, 5, 10, 9] A_ :int = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] A_ :Union[str, Any] = [2, 4, 3, 1] A_ :Dict = [2, 6, 3, 1, 4, 8, 5, 7] A_ :Any = [4, 1, 3, 5, 7, 2, 8, 6] A_ :int = [4, 1, 2, 3, 2, 3, 4, 1] A_ :int = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] A_ :Dict = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation A_ :int = apply_table(key, paa_table) A_ :int = temp[:5] A_ :List[Any] = temp[5:] A_ :Optional[Any] = left_shift(left) A_ :List[str] = left_shift(right) A_ :List[str] = apply_table(left + right, pa_table) A_ :Dict = left_shift(left) A_ :Tuple = left_shift(right) A_ :Dict = left_shift(left) A_ :Union[str, Any] = left_shift(right) A_ :Optional[Any] = apply_table(left + right, pa_table) # encryption A_ :Optional[Any] = apply_table(message, IP) A_ :Tuple = function(expansion, sa, sa, keya, temp) A_ :List[str] = temp[4:] + temp[:4] A_ :Union[str, Any] = function(expansion, sa, sa, keya, temp) A_ :str = apply_table(temp, IP_inv) print('''Cipher text is:''', CT) # decryption A_ :List[Any] = apply_table(CT, IP) A_ :int = function(expansion, sa, sa, keya, temp) A_ :Optional[int] = temp[4:] + temp[:4] A_ :Optional[int] = function(expansion, sa, sa, keya, temp) A_ :Any = apply_table(temp, IP_inv) print('''Plain text after decypting is:''', PT)
351
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ :Union[str, Any] = {'''configuration_plbart''': ['''PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PLBartConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Optional[Any] = ['''PLBartTokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Optional[int] = [ '''PLBART_PRETRAINED_MODEL_ARCHIVE_LIST''', '''PLBartForCausalLM''', '''PLBartForConditionalGeneration''', '''PLBartForSequenceClassification''', '''PLBartModel''', '''PLBartPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys A_ :Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
245
0
'''simple docstring''' import torch from ..models.speechta import SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaProcessor from ..utils import is_datasets_available from .base import PipelineTool if is_datasets_available(): from datasets import load_dataset class lowerCAmelCase__ ( lowerCamelCase_ ): lowerCAmelCase : List[str] = "microsoft/speecht5_tts" lowerCAmelCase : Any = ( "This is a tool that reads an English text out loud. It takes an input named `text` which should contain the " "text to read (in English) and returns a waveform object containing the sound." ) lowerCAmelCase : Optional[Any] = "text_reader" lowerCAmelCase : Optional[Any] = SpeechTaProcessor lowerCAmelCase : Union[str, Any] = SpeechTaForTextToSpeech lowerCAmelCase : str = SpeechTaHifiGan lowerCAmelCase : Union[str, Any] = ["text"] lowerCAmelCase : List[str] = ["audio"] def lowerCAmelCase__ ( self : str ) ->Tuple: '''simple docstring''' if self.post_processor is None: _UpperCAmelCase : str = '''microsoft/speecht5_hifigan''' super().setup() def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : int , lowerCamelCase__ : Union[str, Any]=None ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Tuple = self.pre_processor(text=__SCREAMING_SNAKE_CASE , return_tensors="pt" , truncation=__SCREAMING_SNAKE_CASE ) if speaker_embeddings is None: if not is_datasets_available(): raise ImportError("Datasets needs to be installed if not passing speaker embeddings." ) _UpperCAmelCase : Optional[Any] = load_dataset("Matthijs/cmu-arctic-xvectors" , split="validation" ) _UpperCAmelCase : List[str] = torch.tensor(embeddings_dataset[73_05]["xvector"] ).unsqueeze(0 ) return {"input_ids": inputs["input_ids"], "speaker_embeddings": speaker_embeddings} def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : List[Any] ) ->Tuple: '''simple docstring''' with torch.no_grad(): return self.model.generate_speech(**__SCREAMING_SNAKE_CASE ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : List[Any] ) ->List[str]: '''simple docstring''' with torch.no_grad(): return self.post_processor(__SCREAMING_SNAKE_CASE ).cpu().detach()
234
'''simple docstring''' import argparse import copy def snake_case_ ( __SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" lowercase_ : List[Any] = {} with open(__SCREAMING_SNAKE_CASE ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: lowercase_ : Union[str, Any] = [] _list.append([line.split()[1], line.split()[2]] ) lowercase_ : str = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: lowercase_ : Optional[int] = [] _list.append([line.split()[0], line.split()[2]] ) lowercase_ : Dict = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def snake_case_ ( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Tuple ): """simple docstring""" with open(__SCREAMING_SNAKE_CASE ) as f: lowercase_ : List[str] = f.read(1 ) lowercase_ : Optional[int] = start_node lowercase_ : Any = [] lowercase_ : List[str] = start_node lowercase_ : Optional[Any] = 0 while visiting not in first_solution: lowercase_ : Any = 10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(__SCREAMING_SNAKE_CASE ) and k[0] not in first_solution: lowercase_ : List[Any] = k[1] lowercase_ : List[Any] = k[0] first_solution.append(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = distance_of_first_solution + int(__SCREAMING_SNAKE_CASE ) lowercase_ : int = best_node first_solution.append(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 lowercase_ : Optional[Any] = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def snake_case_ ( __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" lowercase_ : Tuple = [] for n in solution[1:-1]: lowercase_ : List[str] = solution.index(__SCREAMING_SNAKE_CASE ) for kn in solution[1:-1]: lowercase_ : Any = solution.index(__SCREAMING_SNAKE_CASE ) if n == kn: continue lowercase_ : Dict = copy.deepcopy(__SCREAMING_SNAKE_CASE ) lowercase_ : Dict = kn lowercase_ : List[Any] = n lowercase_ : str = 0 for k in _tmp[:-1]: lowercase_ : Tuple = _tmp[_tmp.index(__SCREAMING_SNAKE_CASE ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: lowercase_ : Optional[Any] = distance + int(i[1] ) _tmp.append(__SCREAMING_SNAKE_CASE ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) lowercase_ : Union[str, Any] = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda __SCREAMING_SNAKE_CASE : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def snake_case_ ( __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Optional[Any] ): """simple docstring""" lowercase_ : Optional[int] = 1 lowercase_ : List[str] = first_solution lowercase_ : Dict = [] lowercase_ : List[str] = distance_of_first_solution lowercase_ : Optional[Any] = solution while count <= iters: lowercase_ : int = find_neighborhood(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowercase_ : Any = 0 lowercase_ : Dict = neighborhood[index_of_best_solution] lowercase_ : Optional[Any] = len(__SCREAMING_SNAKE_CASE ) - 1 lowercase_ : Tuple = False while not found: lowercase_ : Optional[int] = 0 while i < len(__SCREAMING_SNAKE_CASE ): if best_solution[i] != solution[i]: lowercase_ : Tuple = best_solution[i] lowercase_ : Optional[int] = solution[i] break lowercase_ : int = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) lowercase_ : Tuple = True lowercase_ : Optional[int] = best_solution[:-1] lowercase_ : Optional[Any] = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: lowercase_ : Optional[Any] = cost lowercase_ : int = solution else: lowercase_ : Any = index_of_best_solution + 1 lowercase_ : Any = neighborhood[index_of_best_solution] if len(__SCREAMING_SNAKE_CASE ) >= size: tabu_list.pop(0 ) lowercase_ : List[Any] = count + 1 return best_solution_ever, best_cost def snake_case_ ( __SCREAMING_SNAKE_CASE : List[str]=None ): """simple docstring""" lowercase_ : Any = generate_neighbours(args.File ) lowercase_ , lowercase_ : Union[str, Any] = generate_first_solution( args.File , __SCREAMING_SNAKE_CASE ) lowercase_ , lowercase_ : Optional[int] = tabu_search( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , args.Iterations , args.Size , ) print(F'''Best solution: {best_sol}, with total distance: {best_cost}.''' ) if __name__ == "__main__": _lowercase : Any = argparse.ArgumentParser(description="Tabu Search") parser.add_argument( "-f", "--File", type=str, help="Path to the file containing the data", required=True, ) parser.add_argument( "-i", "--Iterations", type=int, help="How many iterations the algorithm should perform", required=True, ) parser.add_argument( "-s", "--Size", type=int, help="Size of the tabu list", required=True ) # Pass the arguments to main method main(parser.parse_args())
93
0
from __future__ import annotations def UpperCamelCase( lowercase_ , lowercase_ = None , lowercase_ = None ) -> None: '''simple docstring''' if start is None: snake_case_ = 0 if end is None: snake_case_ = len(lowercase_ ) - 1 if start >= end: return snake_case_ = (start + end) // 2 slowsort(lowercase_ , lowercase_ , lowercase_ ) slowsort(lowercase_ , mid + 1 , lowercase_ ) if sequence[end] < sequence[mid]: snake_case_ , snake_case_ = sequence[mid], sequence[end] slowsort(lowercase_ , lowercase_ , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
34
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase_ = { '''configuration_groupvit''': [ '''GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GroupViTConfig''', '''GroupViTOnnxConfig''', '''GroupViTTextConfig''', '''GroupViTVisionConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GroupViTModel''', '''GroupViTPreTrainedModel''', '''GroupViTTextModel''', '''GroupViTVisionModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFGroupViTModel''', '''TFGroupViTPreTrainedModel''', '''TFGroupViTTextModel''', '''TFGroupViTVisionModel''', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
34
1
'''simple docstring''' def UpperCAmelCase__ ( UpperCAmelCase__ ) -> list: A_ = int(__A ) if n_element < 1: A_ = ValueError("""a should be a positive number""" ) raise my_error A_ = [1] A_ , A_ , A_ = (0, 0, 0) A_ = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2, hamming_list[j] * 3, hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": __lowerCamelCase = input('''Enter the last number (nth term) of the Hamming Number Series: ''') print('''Formula of Hamming Number Series => 2^i * 3^j * 5^k''') __lowerCamelCase = hamming(int(n)) print('''-----------------------------------------------------''') print(f"""The list with nth numbers is: {hamming_numbers}""") print('''-----------------------------------------------------''')
162
from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput snake_case_ : List[str] = 8 def A (__A : Union[str, Any] , __A : List[Any]=BITS ) -> Tuple: """simple docstring""" UpperCAmelCase_ = x.device UpperCAmelCase_ = (x * 255).int().clamp(0 , 255 ) UpperCAmelCase_ = 2 ** torch.arange(bits - 1 , -1 , -1 , device=__A ) UpperCAmelCase_ = rearrange(__A , '''d -> d 1 1''' ) UpperCAmelCase_ = rearrange(__A , '''b c h w -> b c 1 h w''' ) UpperCAmelCase_ = ((x & mask) != 0).float() UpperCAmelCase_ = rearrange(__A , '''b c d h w -> b (c d) h w''' ) UpperCAmelCase_ = bits * 2 - 1 return bits def A (__A : Dict , __A : Tuple=BITS ) -> List[str]: """simple docstring""" UpperCAmelCase_ = x.device UpperCAmelCase_ = (x > 0).int() UpperCAmelCase_ = 2 ** torch.arange(bits - 1 , -1 , -1 , device=__A , dtype=torch.intaa ) UpperCAmelCase_ = rearrange(__A , '''d -> d 1 1''' ) UpperCAmelCase_ = rearrange(__A , '''b (c d) h w -> b c d h w''' , d=8 ) UpperCAmelCase_ = reduce(x * mask , '''b c d h w -> b c h w''' , '''sum''' ) return (dec / 255).clamp(0.0 , 1.0 ) def A (self : List[Any] , __A : torch.FloatTensor , __A : int , __A : torch.FloatTensor , __A : float = 0.0 , __A : bool = True , __A : Tuple=None , __A : bool = True , ) -> Union[DDIMSchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( '''Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler''' ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) UpperCAmelCase_ = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas UpperCAmelCase_ = self.alphas_cumprod[timestep] UpperCAmelCase_ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod UpperCAmelCase_ = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf UpperCAmelCase_ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" UpperCAmelCase_ = self.bit_scale if self.config.clip_sample: UpperCAmelCase_ = torch.clamp(__A , -scale , __A ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) UpperCAmelCase_ = self._get_variance(__A , __A ) UpperCAmelCase_ = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide UpperCAmelCase_ = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf UpperCAmelCase_ = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf UpperCAmelCase_ = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 UpperCAmelCase_ = model_output.device if torch.is_tensor(__A ) else '''cpu''' UpperCAmelCase_ = torch.randn(model_output.shape , dtype=model_output.dtype , generator=__A ).to(__A ) UpperCAmelCase_ = self._get_variance(__A , __A ) ** 0.5 * eta * noise UpperCAmelCase_ = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=__A , pred_original_sample=__A ) def A (self : Optional[int] , __A : torch.FloatTensor , __A : int , __A : torch.FloatTensor , __A : int="epsilon" , __A : Optional[Any]=None , __A : bool = True , ) -> Union[DDPMSchedulerOutput, Tuple]: """simple docstring""" UpperCAmelCase_ = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: UpperCAmelCase_ , UpperCAmelCase_ = torch.split(__A , sample.shape[1] , dim=1 ) else: UpperCAmelCase_ = None # 1. compute alphas, betas UpperCAmelCase_ = self.alphas_cumprod[t] UpperCAmelCase_ = self.alphas_cumprod[t - 1] if t > 0 else self.one UpperCAmelCase_ = 1 - alpha_prod_t UpperCAmelCase_ = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": UpperCAmelCase_ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": UpperCAmelCase_ = model_output else: raise ValueError(F"""Unsupported prediction_type {prediction_type}.""" ) # 3. Clip "predicted x_0" UpperCAmelCase_ = self.bit_scale if self.config.clip_sample: UpperCAmelCase_ = torch.clamp(__A , -scale , __A ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase_ = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t UpperCAmelCase_ = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase_ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCAmelCase_ = 0 if t > 0: UpperCAmelCase_ = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=__A ).to(model_output.device ) UpperCAmelCase_ = (self._get_variance(__A , predicted_variance=__A ) ** 0.5) * noise UpperCAmelCase_ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=__A , pred_original_sample=__A ) class __snake_case ( a ): def __init__( self : Union[str, Any] , _snake_case : UNetaDConditionModel , _snake_case : Union[DDIMScheduler, DDPMScheduler] , _snake_case : Optional[float] = 1.0 , ): """simple docstring""" super().__init__() UpperCAmelCase_ = bit_scale UpperCAmelCase_ = ( ddim_bit_scheduler_step if isinstance(_snake_case , _snake_case) else ddpm_bit_scheduler_step ) self.register_modules(unet=_snake_case , scheduler=_snake_case) @torch.no_grad() def __call__( self : Union[str, Any] , _snake_case : Optional[int] = 256 , _snake_case : Optional[int] = 256 , _snake_case : Optional[int] = 50 , _snake_case : Optional[torch.Generator] = None , _snake_case : Optional[int] = 1 , _snake_case : Optional[str] = "pil" , _snake_case : bool = True , **_snake_case : Optional[Any] , ): """simple docstring""" UpperCAmelCase_ = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=_snake_case , ) UpperCAmelCase_ = decimal_to_bits(_snake_case) * self.bit_scale UpperCAmelCase_ = latents.to(self.device) self.scheduler.set_timesteps(_snake_case) for t in self.progress_bar(self.scheduler.timesteps): # predict the noise residual UpperCAmelCase_ = self.unet(_snake_case , _snake_case).sample # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ = self.scheduler.step(_snake_case , _snake_case , _snake_case).prev_sample UpperCAmelCase_ = bits_to_decimal(_snake_case) if output_type == "pil": UpperCAmelCase_ = self.numpy_to_pil(_snake_case) if not return_dict: return (image,) return ImagePipelineOutput(images=_snake_case)
51
0
from __future__ import annotations def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Optional[int] ): __UpperCamelCase =2 __UpperCamelCase =[] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(__a ) if n > 1: factors.append(__a ) return factors if __name__ == "__main__": import doctest doctest.testmod()
350
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, is_vision_available, ) _A = {'configuration_vit': ['VIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ViTConfig', 'ViTOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ['ViTFeatureExtractor'] _A = ['ViTImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ 'VIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'ViTForImageClassification', 'ViTForMaskedImageModeling', 'ViTModel', 'ViTPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ 'TFViTForImageClassification', 'TFViTModel', 'TFViTPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ 'FlaxViTForImageClassification', 'FlaxViTModel', 'FlaxViTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_vit import VIT_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTConfig, ViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_vit import ViTFeatureExtractor from .image_processing_vit import ViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit import ( VIT_PRETRAINED_MODEL_ARCHIVE_LIST, ViTForImageClassification, ViTForMaskedImageModeling, ViTModel, ViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit import TFViTForImageClassification, TFViTModel, TFViTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel, FlaxViTPreTrainedModel else: import sys _A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
117
0
def a__ ( snake_case , snake_case , snake_case ): """simple docstring""" def count_of_possible_combinations(snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(snake_case ) def a__ ( snake_case , snake_case , snake_case ): """simple docstring""" def count_of_possible_combinations_with_dp_array( snake_case , snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] __SCREAMING_SNAKE_CASE : Any = sum( count_of_possible_combinations_with_dp_array(target - item , snake_case ) for item in array ) __SCREAMING_SNAKE_CASE : List[str] = answer return answer __SCREAMING_SNAKE_CASE : List[str] = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(snake_case , snake_case ) def a__ ( snake_case , snake_case , snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = [0] * (target + 1) __SCREAMING_SNAKE_CASE : Optional[int] = 1 for i in range(1 , target + 1 ): for j in range(snake_case ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() lowercase_ = 3 lowercase_ = 5 lowercase_ = [1, 2, 5] print(combination_sum_iv(n, array, target))
303
from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline lowercase_ = logging.get_logger(__name__) # pylint: disable=invalid-name class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" def __init__( self : List[str] , _A : Dict , _A : List[Any] ): """simple docstring""" super().__init__() self.register_modules(unet=_A , scheduler=_A ) @torch.no_grad() def __call__( self : List[str] , _A : int = 1 , _A : int = 100 , _A : Optional[Union[torch.Generator, List[torch.Generator]]] = None , _A : Optional[float] = None , _A : bool = True , ): """simple docstring""" if audio_length_in_s is None: __SCREAMING_SNAKE_CASE : Optional[Any] = self.unet.config.sample_size / self.unet.config.sample_rate __SCREAMING_SNAKE_CASE : List[Any] = audio_length_in_s * self.unet.config.sample_rate __SCREAMING_SNAKE_CASE : Any = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( F'''{audio_length_in_s} is too small. Make sure it\'s bigger or equal to''' F''' {3 * down_scale_factor / self.unet.config.sample_rate}.''' ) __SCREAMING_SNAKE_CASE : int = int(_A ) if sample_size % down_scale_factor != 0: __SCREAMING_SNAKE_CASE : Optional[int] = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( F'''{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled''' F''' by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising''' ''' process.''' ) __SCREAMING_SNAKE_CASE : List[Any] = int(_A ) __SCREAMING_SNAKE_CASE : Union[str, Any] = next(iter(self.unet.parameters() ) ).dtype __SCREAMING_SNAKE_CASE : int = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(_A , _A ) and len(_A ) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(_A )}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) __SCREAMING_SNAKE_CASE : Dict = randn_tensor(_A , generator=_A , device=self.device , dtype=_A ) # set step values self.scheduler.set_timesteps(_A , device=audio.device ) __SCREAMING_SNAKE_CASE : Dict = self.scheduler.timesteps.to(_A ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output __SCREAMING_SNAKE_CASE : List[Any] = self.unet(_A , _A ).sample # 2. compute previous image: x_t -> t_t-1 __SCREAMING_SNAKE_CASE : Optional[int] = self.scheduler.step(_A , _A , _A ).prev_sample __SCREAMING_SNAKE_CASE : str = audio.clamp(-1 , 1 ).float().cpu().numpy() __SCREAMING_SNAKE_CASE : str = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=_A )
303
1
"""simple docstring""" def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase = 1_000 ) -> int: '''simple docstring''' lowerCAmelCase : str = -1 lowerCAmelCase : int = 0 for a in range(1, n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowerCAmelCase : str = (n * n - 2 * a * n) // (2 * n - 2 * a) lowerCAmelCase : str = n - a - b if c * c == (a * a + b * b): lowerCAmelCase : List[str] = a * b * c if candidate >= product: lowerCAmelCase : List[Any] = candidate return product if __name__ == "__main__": print(F'{solution() = }')
359
def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> int: '''simple docstring''' if len(_UpperCAmelCase ) != len(_UpperCAmelCase ): raise ValueError('String lengths must match!' ) lowerCAmelCase : Tuple = 0 for chara, chara in zip(_UpperCAmelCase, _UpperCAmelCase ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
323
0
"""simple docstring""" from PIL import Image def a_ ( lowerCamelCase , lowerCamelCase ): UpperCAmelCase__ = (2_5_9 * (level + 2_5_5)) / (2_5_5 * (2_5_9 - level)) def contrast(lowerCamelCase ) -> int: return int(1_2_8 + factor * (c - 1_2_8) ) return img.point(lowerCamelCase ) if __name__ == "__main__": # Load image with Image.open('image_data/lena.jpg') as img: # Change contrast to 170 lowerCAmelCase__ : Any = change_contrast(img, 170) cont_img.save('image_data/lena_high_contrast.png', format='png')
98
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : int = logging.get_logger(__name__) lowerCAmelCase__ : str = { 'facebook/xglm-564M': 'https://huggingface.co/facebook/xglm-564M/resolve/main/config.json', # See all XGLM models at https://huggingface.co/models?filter=xglm } class snake_case ( __UpperCAmelCase ): """simple docstring""" snake_case__ = "xglm" snake_case__ = ["past_key_values"] snake_case__ = { "num_attention_heads": "attention_heads", "hidden_size": "d_model", "num_hidden_layers": "num_layers", } def __init__( self : Any ,lowerCamelCase__ : Any=256_008 ,lowerCamelCase__ : Optional[Any]=2_048 ,lowerCamelCase__ : List[str]=1_024 ,lowerCamelCase__ : List[str]=4_096 ,lowerCamelCase__ : Tuple=24 ,lowerCamelCase__ : Optional[int]=16 ,lowerCamelCase__ : int="gelu" ,lowerCamelCase__ : Dict=0.1 ,lowerCamelCase__ : int=0.1 ,lowerCamelCase__ : List[Any]=0.0 ,lowerCamelCase__ : List[str]=0.0 ,lowerCamelCase__ : Optional[Any]=0.0_2 ,lowerCamelCase__ : List[str]=True ,lowerCamelCase__ : Optional[Any]=True ,lowerCamelCase__ : str=2 ,lowerCamelCase__ : Dict=1 ,lowerCamelCase__ : Optional[int]=0 ,lowerCamelCase__ : Tuple=2 ,**lowerCamelCase__ : List[Any] ,): UpperCAmelCase__ = vocab_size UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = d_model UpperCAmelCase__ = ffn_dim UpperCAmelCase__ = num_layers UpperCAmelCase__ = attention_heads UpperCAmelCase__ = activation_function UpperCAmelCase__ = dropout UpperCAmelCase__ = attention_dropout UpperCAmelCase__ = activation_dropout UpperCAmelCase__ = layerdrop UpperCAmelCase__ = init_std UpperCAmelCase__ = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase__ = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,)
98
1
import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() _A : Dict = logging.get_logger(__name__) def _a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: """simple docstring""" lowerCamelCase__ : str = WavaVecaForSequenceClassification.from_pretrained(UpperCAmelCase , config=UpperCAmelCase ) lowerCamelCase__ : List[Any] = downstream_dict['''projector.weight'''] lowerCamelCase__ : str = downstream_dict['''projector.bias'''] lowerCamelCase__ : Optional[Any] = downstream_dict['''model.post_net.linear.weight'''] lowerCamelCase__ : Any = downstream_dict['''model.post_net.linear.bias'''] return model def _a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[Any]: """simple docstring""" lowerCamelCase__ : int = WavaVecaForAudioFrameClassification.from_pretrained(UpperCAmelCase , config=UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = downstream_dict['''model.linear.weight'''] lowerCamelCase__ : Optional[Any] = downstream_dict['''model.linear.bias'''] return model def _a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> str: """simple docstring""" lowerCamelCase__ : List[Any] = WavaVecaForXVector.from_pretrained(UpperCAmelCase , config=UpperCAmelCase ) lowerCamelCase__ : Dict = downstream_dict['''connector.weight'''] lowerCamelCase__ : Dict = downstream_dict['''connector.bias'''] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): lowerCamelCase__ : Union[str, Any] = downstream_dict[ f"model.framelevel_feature_extractor.module.{i}.kernel.weight" ] lowerCamelCase__ : Union[str, Any] = downstream_dict[f"model.framelevel_feature_extractor.module.{i}.kernel.bias"] lowerCamelCase__ : int = downstream_dict['''model.utterancelevel_feature_extractor.linear1.weight'''] lowerCamelCase__ : Tuple = downstream_dict['''model.utterancelevel_feature_extractor.linear1.bias'''] lowerCamelCase__ : List[str] = downstream_dict['''model.utterancelevel_feature_extractor.linear2.weight'''] lowerCamelCase__ : List[str] = downstream_dict['''model.utterancelevel_feature_extractor.linear2.bias'''] lowerCamelCase__ : Optional[int] = downstream_dict['''objective.W'''] return model @torch.no_grad() def _a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Dict: """simple docstring""" lowerCamelCase__ : str = torch.load(UpperCAmelCase , map_location='''cpu''' ) lowerCamelCase__ : Optional[int] = checkpoint['''Downstream'''] lowerCamelCase__ : str = WavaVecaConfig.from_pretrained(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = WavaVecaFeatureExtractor.from_pretrained( UpperCAmelCase , return_attention_mask=UpperCAmelCase , do_normalize=UpperCAmelCase ) lowerCamelCase__ : int = hf_config.architectures[0] if arch.endswith('''ForSequenceClassification''' ): lowerCamelCase__ : Tuple = convert_classification(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) elif arch.endswith('''ForAudioFrameClassification''' ): lowerCamelCase__ : Any = convert_diarization(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) elif arch.endswith('''ForXVector''' ): lowerCamelCase__ : List[str] = convert_xvector(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) else: raise NotImplementedError(f"S3PRL weights conversion is not supported for {arch}" ) if hf_config.use_weighted_layer_sum: lowerCamelCase__ : str = checkpoint['''Featurizer''']['''weights'''] hf_feature_extractor.save_pretrained(UpperCAmelCase ) hf_model.save_pretrained(UpperCAmelCase ) if __name__ == "__main__": _A : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') _A : List[str] = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
265
import inspect from typing import Callable, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import DiffusionPipeline from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import logging _A : Optional[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): def __init__( self : Optional[Any] , A : AutoencoderKL , A : CLIPTextModel , A : CLIPTokenizer , A : UNetaDConditionModel , A : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , A : StableDiffusionSafetyChecker , A : CLIPImageProcessor , ) ->List[Any]: super().__init__() self.register_modules( vae=A , text_encoder=A , tokenizer=A , unet=A , scheduler=A , safety_checker=A , feature_extractor=A , ) def __lowerCamelCase ( self : Any , A : Optional[Union[str, int]] = "auto" ) ->Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowerCamelCase__ : List[Any] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(A ) def __lowerCamelCase ( self : List[Any] ) ->List[Any]: self.enable_attention_slicing(A ) @torch.no_grad() def __call__( self : Union[str, Any] , A : Union[str, List[str]] , A : int = 5_1_2 , A : int = 5_1_2 , A : int = 5_0 , A : float = 7.5 , A : Optional[Union[str, List[str]]] = None , A : Optional[int] = 1 , A : float = 0.0 , A : Optional[torch.Generator] = None , A : Optional[torch.FloatTensor] = None , A : Optional[str] = "pil" , A : bool = True , A : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , A : int = 1 , A : Optional[torch.FloatTensor] = None , **A : int , ) ->Tuple: if isinstance(A , A ): lowerCamelCase__ : str = 1 elif isinstance(A , A ): lowerCamelCase__ : Dict = len(A ) else: raise ValueError(F"`prompt` has to be of type `str` or `list` but is {type(A )}" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F"`height` and `width` have to be divisible by 8 but are {height} and {width}." ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(A , A ) or callback_steps <= 0) ): raise ValueError( F"`callback_steps` has to be a positive integer but is {callback_steps} of type" F" {type(A )}." ) # get prompt text embeddings lowerCamelCase__ : Optional[Any] = self.tokenizer( A , padding='''max_length''' , max_length=self.tokenizer.model_max_length , return_tensors='''pt''' , ) lowerCamelCase__ : Dict = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: lowerCamelCase__ : Optional[Any] = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' F" {self.tokenizer.model_max_length} tokens: {removed_text}" ) lowerCamelCase__ : Any = text_input_ids[:, : self.tokenizer.model_max_length] if text_embeddings is None: lowerCamelCase__ : str = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = text_embeddings.shape lowerCamelCase__ : int = text_embeddings.repeat(1 , A , 1 ) lowerCamelCase__ : str = text_embeddings.view(bs_embed * num_images_per_prompt , A , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. lowerCamelCase__ : Optional[Any] = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: lowerCamelCase__ : List[str] if negative_prompt is None: lowerCamelCase__ : Optional[int] = [''''''] elif type(A ) is not type(A ): raise TypeError( F"`negative_prompt` should be the same type to `prompt`, but got {type(A )} !=" F" {type(A )}." ) elif isinstance(A , A ): lowerCamelCase__ : Optional[int] = [negative_prompt] elif batch_size != len(A ): raise ValueError( F"`negative_prompt`: {negative_prompt} has batch size {len(A )}, but `prompt`:" F" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches" ''' the batch size of `prompt`.''' ) else: lowerCamelCase__ : List[Any] = negative_prompt lowerCamelCase__ : int = text_input_ids.shape[-1] lowerCamelCase__ : Optional[int] = self.tokenizer( A , padding='''max_length''' , max_length=A , truncation=A , return_tensors='''pt''' , ) lowerCamelCase__ : Optional[Any] = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method lowerCamelCase__ : Any = uncond_embeddings.shape[1] lowerCamelCase__ : Union[str, Any] = uncond_embeddings.repeat(A , A , 1 ) lowerCamelCase__ : str = uncond_embeddings.view(batch_size * num_images_per_prompt , A , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowerCamelCase__ : str = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. lowerCamelCase__ : Dict = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) lowerCamelCase__ : List[Any] = (batch_size * num_images_per_prompt, self.unet.config.in_channels, 6_4, 6_4) lowerCamelCase__ : List[Any] = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps lowerCamelCase__ : Dict = torch.randn( A , generator=A , device='''cpu''' , dtype=A ).to(self.device ) lowerCamelCase__ : Dict = torch.randn(A , generator=A , device='''cpu''' , dtype=A ).to( self.device ) else: lowerCamelCase__ : Optional[Any] = torch.randn( A , generator=A , device=self.device , dtype=A ) lowerCamelCase__ : Union[str, Any] = torch.randn(A , generator=A , device=self.device , dtype=A ) else: if latents_reference.shape != latents_shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {latents_shape}" ) lowerCamelCase__ : Optional[Any] = latents_reference.to(self.device ) lowerCamelCase__ : Optional[int] = latents.to(self.device ) # This is the key part of the pipeline where we # try to ensure that the generated images w/ the same seed # but different sizes actually result in similar images lowerCamelCase__ : List[str] = (latents_shape[3] - latents_shape_reference[3]) // 2 lowerCamelCase__ : List[Any] = (latents_shape[2] - latents_shape_reference[2]) // 2 lowerCamelCase__ : Dict = latents_shape_reference[3] if dx >= 0 else latents_shape_reference[3] + 2 * dx lowerCamelCase__ : str = latents_shape_reference[2] if dy >= 0 else latents_shape_reference[2] + 2 * dy lowerCamelCase__ : int = 0 if dx < 0 else dx lowerCamelCase__ : Optional[int] = 0 if dy < 0 else dy lowerCamelCase__ : Dict = max(-dx , 0 ) lowerCamelCase__ : int = max(-dy , 0 ) # import pdb # pdb.set_trace() lowerCamelCase__ : str = latents_reference[:, :, dy : dy + h, dx : dx + w] # set timesteps self.scheduler.set_timesteps(A ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand lowerCamelCase__ : Union[str, Any] = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler lowerCamelCase__ : List[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCamelCase__ : Union[str, Any] = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCamelCase__ : List[Any] = {} if accepts_eta: lowerCamelCase__ : Any = eta for i, t in enumerate(self.progress_bar(A ) ): # expand the latents if we are doing classifier free guidance lowerCamelCase__ : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase__ : Any = self.scheduler.scale_model_input(A , A ) # predict the noise residual lowerCamelCase__ : Union[str, Any] = self.unet(A , A , encoder_hidden_states=A ).sample # perform guidance if do_classifier_free_guidance: lowerCamelCase__ , lowerCamelCase__ : List[str] = noise_pred.chunk(2 ) lowerCamelCase__ : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 lowerCamelCase__ : Dict = self.scheduler.step(A , A , A , **A ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(A , A , A ) lowerCamelCase__ : Optional[Any] = 1 / 0.1_82_15 * latents lowerCamelCase__ : int = self.vae.decode(A ).sample lowerCamelCase__ : Dict = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowerCamelCase__ : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if self.safety_checker is not None: lowerCamelCase__ : Tuple = self.feature_extractor(self.numpy_to_pil(A ) , return_tensors='''pt''' ).to( self.device ) lowerCamelCase__ , lowerCamelCase__ : int = self.safety_checker( images=A , clip_input=safety_checker_input.pixel_values.to(text_embeddings.dtype ) ) else: lowerCamelCase__ : List[Any] = None if output_type == "pil": lowerCamelCase__ : Optional[Any] = self.numpy_to_pil(A ) if not return_dict: return (image, has_nsfw_concept) return StableDiffusionPipelineOutput(images=A , nsfw_content_detected=A )
265
1
"""simple docstring""" import argparse import logging import os import sys import numpy as np import onnxruntime import torch from bart_onnx.generation_onnx import BARTBeamSearchGenerator from bart_onnx.reduce_onnx_size import remove_dup_initializers import transformers from transformers import BartForConditionalGeneration, BartTokenizer logging.basicConfig( format="%(asctime)s | %(levelname)s | %(name)s | [%(filename)s:%(lineno)d] %(message)s", datefmt="%Y-%m-%d %H:%M:%S", level=os.environ.get("LOGLEVEL", "INFO").upper(), stream=sys.stdout, ) SCREAMING_SNAKE_CASE__ = logging.getLogger(__name__) SCREAMING_SNAKE_CASE__ = {"facebook/bart-base": BartForConditionalGeneration} SCREAMING_SNAKE_CASE__ = {"facebook/bart-base": BartTokenizer} def UpperCAmelCase__ ( ): '''simple docstring''' lowerCAmelCase = argparse.ArgumentParser(description="""Export Bart model + Beam Search to ONNX graph.""" ) parser.add_argument( """--validation_file""" , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help="""A csv or a json file containing the validation data.""" ) parser.add_argument( """--max_length""" , type=SCREAMING_SNAKE_CASE , default=5 , help="""The maximum total input sequence length after tokenization.""" , ) parser.add_argument( """--num_beams""" , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help=( """Number of beams to use for evaluation. This argument will be """ """passed to ``model.generate``, which is used during ``evaluate`` and ``predict``.""" ) , ) parser.add_argument( """--model_name_or_path""" , type=SCREAMING_SNAKE_CASE , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=SCREAMING_SNAKE_CASE , ) parser.add_argument( """--config_name""" , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help="""Pretrained config name or path if not the same as model_name""" , ) parser.add_argument( """--device""" , type=SCREAMING_SNAKE_CASE , default="""cpu""" , help="""Device where the model will be run""" , ) parser.add_argument("""--output_file_path""" , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help="""Where to store the final ONNX file.""" ) lowerCAmelCase = parser.parse_args() return args def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : Union[str, Any]="cpu" ): '''simple docstring''' lowerCAmelCase = model_dict[model_name].from_pretrained(SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ) lowerCAmelCase = tokenizer_dict[model_name].from_pretrained(SCREAMING_SNAKE_CASE ) if model_name in ["facebook/bart-base"]: lowerCAmelCase = 0 lowerCAmelCase = None lowerCAmelCase = 0 return huggingface_model, tokenizer def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : str ): '''simple docstring''' model.eval() lowerCAmelCase = None lowerCAmelCase = torch.jit.script(BARTBeamSearchGenerator(SCREAMING_SNAKE_CASE ) ) with torch.no_grad(): lowerCAmelCase = """My friends are cool but they eat too many carbs.""" lowerCAmelCase = tokenizer([ARTICLE_TO_SUMMARIZE] , max_length=10_24 , return_tensors="""pt""" ).to(model.device ) lowerCAmelCase = model.generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , num_beams=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , early_stopping=SCREAMING_SNAKE_CASE , decoder_start_token_id=model.config.decoder_start_token_id , ) torch.onnx.export( SCREAMING_SNAKE_CASE , ( inputs["""input_ids"""], inputs["""attention_mask"""], num_beams, max_length, model.config.decoder_start_token_id, ) , SCREAMING_SNAKE_CASE , opset_version=14 , input_names=["""input_ids""", """attention_mask""", """num_beams""", """max_length""", """decoder_start_token_id"""] , output_names=["""output_ids"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """seq"""}, """output_ids""": {0: """batch""", 1: """seq_out"""}, } , example_outputs=SCREAMING_SNAKE_CASE , ) logger.info("""Model exported to {}""".format(SCREAMING_SNAKE_CASE ) ) lowerCAmelCase = remove_dup_initializers(os.path.abspath(SCREAMING_SNAKE_CASE ) ) logger.info("""Deduplicated and optimized model written to {}""".format(SCREAMING_SNAKE_CASE ) ) lowerCAmelCase = onnxruntime.InferenceSession(SCREAMING_SNAKE_CASE ) lowerCAmelCase = ort_sess.run( SCREAMING_SNAKE_CASE , { """input_ids""": inputs["""input_ids"""].cpu().numpy(), """attention_mask""": inputs["""attention_mask"""].cpu().numpy(), """num_beams""": np.array(SCREAMING_SNAKE_CASE ), """max_length""": np.array(SCREAMING_SNAKE_CASE ), """decoder_start_token_id""": np.array(model.config.decoder_start_token_id ), } , ) np.testing.assert_allclose(summary_ids.cpu().numpy() , ort_out[0] , rtol=1e-3 , atol=1e-3 ) logger.info("""Model outputs from torch and ONNX Runtime are similar.""" ) logger.info("""Success.""" ) def UpperCAmelCase__ ( ): '''simple docstring''' lowerCAmelCase = parse_args() lowerCAmelCase = 5 lowerCAmelCase = 4 # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.setLevel(logging.INFO ) transformers.utils.logging.set_verbosity_error() lowerCAmelCase = torch.device(args.device ) lowerCAmelCase , lowerCAmelCase = load_model_tokenizer(args.model_name_or_path , SCREAMING_SNAKE_CASE ) if model.config.decoder_start_token_id is None: raise ValueError("""Make sure that `config.decoder_start_token_id` is correctly defined""" ) model.to(SCREAMING_SNAKE_CASE ) if args.max_length: lowerCAmelCase = args.max_length if args.num_beams: lowerCAmelCase = args.num_beams if args.output_file_path: lowerCAmelCase = args.output_file_path else: lowerCAmelCase = """BART.onnx""" logger.info("""Exporting model to ONNX""" ) export_and_validate_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
46
"""simple docstring""" import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FlaubertForMultipleChoice, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertModel, FlaubertWithLMHeadModel, ) from transformers.models.flaubert.modeling_flaubert import FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST class lowercase ( _UpperCAmelCase ): def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=False , lowercase=False , lowercase=False , lowercase=2 , lowercase=99 , lowercase=0 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=12 , lowercase=2 , lowercase=0.02 , lowercase=3 , lowercase=4 , lowercase="last" , lowercase=None , lowercase=None , ) -> int: lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = is_training lowerCAmelCase = use_input_lengths lowerCAmelCase = use_token_type_ids lowerCAmelCase = use_labels lowerCAmelCase = gelu_activation lowerCAmelCase = sinusoidal_embeddings lowerCAmelCase = causal lowerCAmelCase = asm lowerCAmelCase = n_langs lowerCAmelCase = vocab_size lowerCAmelCase = n_special lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = type_vocab_size lowerCAmelCase = type_sequence_label_size lowerCAmelCase = initializer_range lowerCAmelCase = num_labels lowerCAmelCase = num_choices lowerCAmelCase = summary_type lowerCAmelCase = use_proj lowerCAmelCase = scope def _snake_case ( self ) -> int: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase = None if self.use_input_lengths: lowerCAmelCase = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length lowerCAmelCase = None if self.use_token_type_ids: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase = ids_tensor([self.batch_size] , 2 ).float() lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _snake_case ( self ) -> List[Any]: return FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> Any: lowerCAmelCase = FlaubertModel(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase , lengths=lowercase , langs=lowercase ) lowerCAmelCase = model(lowercase , langs=lowercase ) lowerCAmelCase = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> Tuple: lowerCAmelCase = FlaubertWithLMHeadModel(lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> str: lowerCAmelCase = FlaubertForQuestionAnsweringSimple(lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase ) lowerCAmelCase = model(lowercase , start_positions=lowercase , end_positions=lowercase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> Dict: lowerCAmelCase = FlaubertForQuestionAnswering(lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase ) lowerCAmelCase = model( lowercase , start_positions=lowercase , end_positions=lowercase , cls_index=lowercase , is_impossible=lowercase , p_mask=lowercase , ) lowerCAmelCase = model( lowercase , start_positions=lowercase , end_positions=lowercase , cls_index=lowercase , is_impossible=lowercase , ) ((lowerCAmelCase) , ) = result_with_labels.to_tuple() lowerCAmelCase = model(lowercase , start_positions=lowercase , end_positions=lowercase ) ((lowerCAmelCase) , ) = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> int: lowerCAmelCase = FlaubertForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase ) lowerCAmelCase = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> int: lowerCAmelCase = self.num_labels lowerCAmelCase = FlaubertForTokenClassification(lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> Tuple: lowerCAmelCase = self.num_choices lowerCAmelCase = FlaubertForMultipleChoice(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _snake_case ( self ) -> List[Any]: lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ) = config_and_inputs lowerCAmelCase = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """lengths""": input_lengths, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class lowercase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE = ( ( FlaubertModel, FlaubertWithLMHeadModel, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertForMultipleChoice, ) if is_torch_available() else () ) _SCREAMING_SNAKE_CASE = ( { 'feature-extraction': FlaubertModel, 'fill-mask': FlaubertWithLMHeadModel, 'question-answering': FlaubertForQuestionAnsweringSimple, 'text-classification': FlaubertForSequenceClassification, 'token-classification': FlaubertForTokenClassification, 'zero-shot': FlaubertForSequenceClassification, } if is_torch_available() else {} ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _snake_case ( self , lowercase , lowercase , lowercase=False ) -> Optional[Any]: lowerCAmelCase = super()._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) if return_labels: if model_class.__name__ == "FlaubertForQuestionAnswering": lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase ) lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase ) return inputs_dict def _snake_case ( self ) -> List[str]: lowerCAmelCase = FlaubertModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=lowercase , emb_dim=37 ) def _snake_case ( self ) -> Tuple: self.config_tester.run_common_tests() def _snake_case ( self ) -> Tuple: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*lowercase ) def _snake_case ( self ) -> Tuple: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*lowercase ) def _snake_case ( self ) -> Optional[int]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_simple_qa(*lowercase ) def _snake_case ( self ) -> Union[str, Any]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*lowercase ) def _snake_case ( self ) -> Any: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*lowercase ) def _snake_case ( self ) -> Any: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_token_classif(*lowercase ) def _snake_case ( self ) -> Any: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_multiple_choice(*lowercase ) @slow def _snake_case ( self ) -> Tuple: for model_name in FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = FlaubertModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) @slow @require_torch_gpu def _snake_case ( self ) -> List[Any]: lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # FlauBertForMultipleChoice behaves incorrectly in JIT environments. if model_class == FlaubertForMultipleChoice: return lowerCAmelCase = True lowerCAmelCase = model_class(config=lowercase ) lowerCAmelCase = self._prepare_for_class(lowercase , lowercase ) lowerCAmelCase = torch.jit.trace( lowercase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowercase , os.path.join(lowercase , """traced_model.pt""" ) ) lowerCAmelCase = torch.jit.load(os.path.join(lowercase , """traced_model.pt""" ) , map_location=lowercase ) loaded(inputs_dict["""input_ids"""].to(lowercase ) , inputs_dict["""attention_mask"""].to(lowercase ) ) @require_torch class lowercase ( unittest.TestCase ): @slow def _snake_case ( self ) -> Optional[int]: lowerCAmelCase = FlaubertModel.from_pretrained("""flaubert/flaubert_base_cased""" ) lowerCAmelCase = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) with torch.no_grad(): lowerCAmelCase = model(lowercase )[0] lowerCAmelCase = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , lowercase ) lowerCAmelCase = torch.tensor( [[[-2.6_251, -1.4_298, -0.0_227], [-2.8_510, -1.6_387, 0.2_258], [-2.8_114, -1.1_832, -0.3_066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase , atol=1e-4 ) )
46
1
def a__ ( UpperCAmelCase : int = 600_851_475_143 ) -> int: try: UpperCAmelCase : int = int(_snake_case ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''' ) UpperCAmelCase : Tuple = 1 UpperCAmelCase : str = 2 while i * i <= n: while n % i == 0: UpperCAmelCase : List[Any] = i n //= i i += 1 if n > 1: UpperCAmelCase : Optional[int] = n return int(_snake_case ) if __name__ == "__main__": print(f"""{solution() = }""")
352
from ..utils import DummyObject, requires_backends class __UpperCAmelCase ( metaclass=lowerCamelCase__ ): UpperCamelCase = ["""onnx"""] def __init__( self : int, *__A : Optional[Any], **__A : Dict ): requires_backends(self, ['''onnx'''] ) @classmethod def __magic_name__ ( cls : Any, *__A : Any, **__A : Dict ): requires_backends(cls, ['''onnx'''] ) @classmethod def __magic_name__ ( cls : Tuple, *__A : List[str], **__A : List[str] ): requires_backends(cls, ['''onnx'''] )
99
0