code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" # Algorithm for the pigeonhole sorting def _snake_case ( _snake_case : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' _A = min(_snake_case ) # min() finds the minimum value _A = max(_snake_case ) # max() finds the maximum value _A = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size _A = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(_snake_case , _snake_case ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. _A = 0 for count in range(_snake_case ): while holes[count] > 0: holes[count] -= 1 _A = count + min_val i += 1 def _snake_case ( ) -> Dict: '''simple docstring''' _A = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(_snake_case ) print('Sorted order is:' , ' '.join(_snake_case ) ) if __name__ == "__main__": main()
315
"""simple docstring""" from __future__ import annotations def _snake_case ( _snake_case : int , _snake_case : int ) -> list[list[int]]: '''simple docstring''' _A = [] create_all_state(1 , _snake_case , _snake_case , [] , _snake_case ) return result def _snake_case ( _snake_case : int , _snake_case : int , _snake_case : int , _snake_case : list[int] , _snake_case : list[list[int]] , ) -> None: '''simple docstring''' if level == 0: total_list.append(current_list[:] ) return for i in range(_snake_case , total_number - level + 2 ): current_list.append(_snake_case ) create_all_state(i + 1 , _snake_case , level - 1 , _snake_case , _snake_case ) current_list.pop() def _snake_case ( _snake_case : list[list[int]] ) -> None: '''simple docstring''' for i in total_list: print(*_snake_case ) if __name__ == "__main__": a = 4 a = 2 a = generate_all_combinations(n, k) print_all_state(total_list)
315
1
def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(snake_case__ ): for j in range(snake_case__ ): if dist[i][j] != float("""inf""" ): print(int(dist[i][j] ) ,end="""\t""" ) else: print("""INF""" ,end="""\t""" ) print() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = [[float("""inf""" ) for _ in range(snake_case__ )] for _ in range(snake_case__ )] for i in range(snake_case__ ): for j in range(snake_case__ ): _SCREAMING_SNAKE_CASE = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(snake_case__ ): # looping through rows of graph array for i in range(snake_case__ ): # looping through columns of graph array for j in range(snake_case__ ): if ( dist[i][k] != float("""inf""" ) and dist[k][j] != float("""inf""" ) and dist[i][k] + dist[k][j] < dist[i][j] ): _SCREAMING_SNAKE_CASE = dist[i][k] + dist[k][j] _print_dist(snake_case__ ,snake_case__ ) return dist, v if __name__ == "__main__": UpperCamelCase = int(input('''Enter number of vertices: ''')) UpperCamelCase = int(input('''Enter number of edges: ''')) UpperCamelCase = [[float('''inf''') for i in range(v)] for j in range(v)] for i in range(v): UpperCamelCase = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print('''\nEdge ''', i + 1) UpperCamelCase = int(input('''Enter source:''')) UpperCamelCase = int(input('''Enter destination:''')) UpperCamelCase = float(input('''Enter weight:''')) UpperCamelCase = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
125
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { '''google/mobilenet_v1_1.0_224''': '''https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json''', '''google/mobilenet_v1_0.75_192''': '''https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Any = "mobilenet_v1" def __init__( self: List[str] , UpperCAmelCase_: List[Any]=3 , UpperCAmelCase_: List[str]=224 , UpperCAmelCase_: Optional[Any]=1.0 , UpperCAmelCase_: Dict=8 , UpperCAmelCase_: List[str]="relu6" , UpperCAmelCase_: List[Any]=True , UpperCAmelCase_: Dict=0.9_99 , UpperCAmelCase_: Union[str, Any]=0.02 , UpperCAmelCase_: Tuple=0.0_01 , **UpperCAmelCase_: List[str] , ): '''simple docstring''' super().__init__(**UpperCAmelCase_ ) if depth_multiplier <= 0: raise ValueError("""depth_multiplier must be greater than zero.""" ) _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = image_size _SCREAMING_SNAKE_CASE = depth_multiplier _SCREAMING_SNAKE_CASE = min_depth _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = tf_padding _SCREAMING_SNAKE_CASE = classifier_dropout_prob _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = layer_norm_eps class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : List[Any] = version.parse("1.11" ) @property def UpperCamelCase ( self: Tuple ): '''simple docstring''' return OrderedDict([("""pixel_values""", {0: """batch"""})] ) @property def UpperCamelCase ( self: Dict ): '''simple docstring''' if self.task == "image-classification": return OrderedDict([("""logits""", {0: """batch"""})] ) else: return OrderedDict([("""last_hidden_state""", {0: """batch"""}), ("""pooler_output""", {0: """batch"""})] ) @property def UpperCamelCase ( self: int ): '''simple docstring''' return 1E-4
125
1
import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class __magic_name__ ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : str = logging.get_logger() # the current default level is logging.WARNING A_ : int = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(snake_case ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' A_ : Optional[int] = logging.get_verbosity() A_ : List[Any] = logging.get_logger("transformers.models.bart.tokenization_bart" ) A_ : str = "Testing 1, 2, 3" # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(snake_case ) as cl: logger.warning(snake_case ) self.assertEqual(cl.out , msg + "\n" ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(snake_case ) as cl: logger.warning(snake_case ) self.assertEqual(cl.out , "" ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(snake_case ) as cl: logger.warning(snake_case ) self.assertEqual(cl.out , msg + "\n" ) # restore to the original level logging.set_verbosity(snake_case ) @mockenv(TRANSFORMERS_VERBOSITY="error" ) def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' transformers.utils.logging._reset_library_root_logger() # this action activates the env var A_ : List[str] = logging.get_logger("transformers.models.bart.tokenization_bart" ) A_ : Optional[int] = os.getenv("TRANSFORMERS_VERBOSITY" , snake_case ) A_ : Optional[Any] = logging.log_levels[env_level_str] A_ : Any = logging.get_verbosity() self.assertEqual( snake_case , snake_case , f"TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}" , ) # restore to the original level A_ : Tuple = "" transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY="super-error" ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' transformers.utils.logging._reset_library_root_logger() A_ : Dict = logging.logging.getLogger() with CaptureLogger(snake_case ) as cl: # this action activates the env var logging.get_logger("transformers.models.bart.tokenization_bart" ) self.assertIn("Unknown option TRANSFORMERS_VERBOSITY=super-error" , cl.out ) # no need to restore as nothing was changed def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' transformers.utils.logging._reset_library_root_logger() A_ : List[Any] = logging.get_logger("transformers.models.bart.tokenization_bart" ) A_ : Tuple = "Testing 1, 2, 3" with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="1" ): # nothing should be logged as env var disables this method with CaptureLogger(snake_case ) as cl: logger.warning_advice(snake_case ) self.assertEqual(cl.out , "" ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="" ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(snake_case ) as cl: logger.warning_advice(snake_case ) self.assertEqual(cl.out , msg + "\n" ) def __snake_case ( ) -> Tuple: disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
300
from __future__ import annotations def __snake_case ( _lowerCAmelCase : list[float] ) -> bool: if len(_lowerCAmelCase ) < 2: raise ValueError("Monogons and Digons are not polygons in the Euclidean space" ) if any(i <= 0 for i in nums ): raise ValueError("All values must be greater than 0" ) A_ : List[str] = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
300
1
'''simple docstring''' import os from distutils.util import strtobool def _UpperCAmelCase ( _UpperCamelCase : Any, _UpperCamelCase : Optional[Any] ) -> str: for e in env_keys: A_ = int(os.environ.get(_UpperCamelCase, -1 ) ) if val >= 0: return val return default def _UpperCAmelCase ( _UpperCamelCase : Union[str, Any], _UpperCamelCase : Optional[Any]=False ) -> int: A_ = os.environ.get(_UpperCamelCase, str(_UpperCamelCase ) ) return strtobool(_UpperCamelCase ) == 1 # As its name indicates `strtobool` actually returns an int... def _UpperCAmelCase ( _UpperCamelCase : List[str], _UpperCamelCase : List[str]="no" ) -> Any: A_ = os.environ.get(_UpperCamelCase, str(_UpperCamelCase ) ) return value
18
'''simple docstring''' # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def _UpperCAmelCase ( _UpperCamelCase : Tuple, _UpperCamelCase : Tuple, _UpperCamelCase : List[str] ) -> int: A_ = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, oder?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] A_ = { '''ru-en''': ['''[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)''', '''39.20'''], '''en-ru''': ['''[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)''', '''33.47'''], '''en-de''': ['''[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)''', '''42.83'''], '''de-en''': ['''[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)''', '''41.35'''], } A_ = F'''{src_lang}-{tgt_lang}''' A_ = F''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(_UpperCamelCase, exist_ok=_UpperCamelCase ) A_ = os.path.join(_UpperCamelCase, '''README.md''' ) print(F'''Generating {path}''' ) with open(_UpperCamelCase, '''w''', encoding='''utf-8''' ) as f: f.write(_UpperCamelCase ) # make sure we are under the root of the project __snake_case : Any = Path(__file__).resolve().parent.parent.parent __snake_case : Tuple = repo_dir / 'model_cards' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __snake_case , __snake_case , __snake_case : Any = model_name.split('-') __snake_case : int = model_cards_dir / 'facebook' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
18
1
"""simple docstring""" # Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union __A : List[str] = re.compile(R'''^(?P<major>\d+)''' R'''\.(?P<minor>\d+)''' R'''\.(?P<patch>\d+)$''') @total_ordering @dataclass class _UpperCAmelCase : SCREAMING_SNAKE_CASE_ : str SCREAMING_SNAKE_CASE_ : Optional[str] = None SCREAMING_SNAKE_CASE_ : Optional[Union[str, int]] = None SCREAMING_SNAKE_CASE_ : Optional[Union[str, int]] = None SCREAMING_SNAKE_CASE_ : Optional[Union[str, int]] = None def A ( self : Optional[int] ) -> Union[str, Any]: lowercase_ , lowercase_ , lowercase_ : Union[str, Any] = _str_to_version_tuple(self.version_str ) def __repr__( self : int ) -> List[Any]: return F'''{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}''' @property def A ( self : Any ) -> Union[str, Any]: return self.major, self.minor, self.patch def A ( self : Dict , A : List[Any] ) -> Tuple: if isinstance(A , A ): return Version(A ) elif isinstance(A , A ): return other raise TypeError(F'''{other} (type {type(A )}) cannot be compared to version.''' ) def __eq__( self : Union[str, Any] , A : str ) -> List[str]: try: lowercase_ : Optional[int] = self._validate_operand(A ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self : int , A : str ) -> Union[str, Any]: lowercase_ : Optional[Any] = self._validate_operand(A ) return self.tuple < other.tuple def __hash__( self : Dict ) -> Optional[Any]: return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def A ( cls : int , A : List[Any] ) -> List[str]: lowercase_ : Dict = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def A ( self : Optional[int] ) -> str: return self.version_str def lowercase ( __snake_case : Any ): lowercase_ : int = _VERSION_REG.match(__snake_case ) if not res: raise ValueError(F'''Invalid version \'{version_str}\'. Format should be x.y.z with {{x,y,z}} being digits.''' ) return tuple(int(__snake_case ) for v in [res.group('''major''' ), res.group('''minor''' ), res.group('''patch''' )] ) def lowercase ( __snake_case : int ): return ".".join(str(__snake_case ) for v in version_tuple )
33
"""simple docstring""" import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __A : str = argparse.ArgumentParser() parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--txt2img_unclip''', default='''kakaobrain/karlo-v1-alpha''', type=str, required=False, help='''The pretrained txt2img unclip.''', ) __A : str = parser.parse_args() __A : List[Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __A : Dict = CLIPImageProcessor() __A : Union[str, Any] = CLIPVisionModelWithProjection.from_pretrained('''openai/clip-vit-large-patch14''') __A : List[str] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
33
1
import argparse from collections import defaultdict def lowerCamelCase__ ( a , a , a , a , a ) -> Union[str, Any]: _A: Union[str, Any] = f"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(a , '''r''' ) as f: _A: List[Any] = f.readlines() _A: Tuple = f"""class {class_name}(""" _A: Tuple = f"""{4 * ' '}def {test_name}(""" _A: Union[str, Any] = f"""{8 * ' '}{correct_line.split()[0]}""" _A: Any = f"""{16 * ' '}{correct_line.split()[0]}""" _A: List[Any] = False _A: int = False _A: List[str] = False _A: Dict = False _A: Union[str, Any] = 0 _A: Optional[Any] = 0 _A: Dict = [] for line in lines: if line.startswith(a ): _A: Union[str, Any] = True elif in_class and line.startswith(a ): _A: Any = True elif in_class and in_func and (line.startswith(a ) or line.startswith(a )): _A: int = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: _A: List[str] = True if in_class and in_func and in_line: if ")" not in line: continue else: _A: List[str] = True if in_class and in_func and in_line and insert_line: new_lines.append(f"""{spaces * ' '}{correct_line}""" ) _A: List[Any] = False else: new_lines.append(a ) with open(a , '''w''' ) as f: for line in new_lines: f.write(a ) def lowerCamelCase__ ( a , a=None ) -> int: if fail is not None: with open(a , '''r''' ) as f: _A: Dict = {l.strip() for l in f.readlines()} else: _A: Any = None with open(a , '''r''' ) as f: _A: Union[str, Any] = f.readlines() _A: str = defaultdict(a ) for line in correct_lines: _A , _A , _A , _A: Optional[int] = line.split(''';''' ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(a , a , a , a , a ) if __name__ == "__main__": UpperCAmelCase__ : Optional[int] = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) UpperCAmelCase__ : Optional[int] = parser.parse_args() main(args.correct_filename, args.fail_filename)
301
from __future__ import annotations UpperCAmelCase__ : List[str] = list[list[int]] # assigning initial values to the grid UpperCAmelCase__ : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution UpperCAmelCase__ : Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def lowerCamelCase__ ( a , a , a , a ) -> bool: for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def lowerCamelCase__ ( a ) -> tuple[int, int] | None: for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def lowerCamelCase__ ( a ) -> Matrix | None: if location := find_empty_location(a ): _A , _A: Optional[Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(a , a , a , a ): _A: str = digit if sudoku(a ) is not None: return grid _A: Tuple = 0 return None def lowerCamelCase__ ( a ) -> None: for row in grid: for cell in row: print(a , end=''' ''' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('\nExample grid:\n' + '=' * 20) print_solution(example_grid) print('\nExample grid solution:') UpperCAmelCase__ : int = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('Cannot find a solution.')
301
1
from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar __lowerCamelCase = TypeVar("""T""") __lowerCamelCase = TypeVar("""U""") class UpperCAmelCase ( Generic[T, U] ): def __init__(self : Dict , snake_case__ : T | None , snake_case__ : U | None ) -> List[str]: '''simple docstring''' snake_case : Any = key snake_case : int = val snake_case : DoubleLinkedListNode[T, U] | None = None snake_case : DoubleLinkedListNode[T, U] | None = None def __repr__(self : List[Any] ) -> str: '''simple docstring''' return ( f"""Node: key: {self.key}, val: {self.val}, """ f"""has next: {bool(self.next )}, has prev: {bool(self.prev )}""" ) class UpperCAmelCase ( Generic[T, U] ): def __init__(self : List[Any] ) -> None: '''simple docstring''' snake_case : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(snake_case__ , snake_case__ ) snake_case : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(snake_case__ , snake_case__ ) snake_case , snake_case : Optional[int] = self.rear, self.head def __repr__(self : int ) -> str: '''simple docstring''' snake_case : Dict = ["DoubleLinkedList"] snake_case : Optional[int] = self.head while node.next is not None: rep.append(str(snake_case__ ) ) snake_case : List[str] = node.next rep.append(str(self.rear ) ) return ",\n ".join(snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Optional[Any] , snake_case__ : DoubleLinkedListNode[T, U] ) -> None: '''simple docstring''' snake_case : Any = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None snake_case : str = node snake_case : Any = previous snake_case : str = node snake_case : Dict = self.rear def _SCREAMING_SNAKE_CASE (self : Optional[Any] , snake_case__ : DoubleLinkedListNode[T, U] ) -> DoubleLinkedListNode[T, U] | None: '''simple docstring''' if node.prev is None or node.next is None: return None snake_case : Any = node.next snake_case : int = node.prev snake_case : str = None snake_case : str = None return node class UpperCAmelCase ( Generic[T, U] ): A__ : dict[Callable[[T], U], LRUCache[T, U]] = {} def __init__(self : Union[str, Any] , snake_case__ : int ) -> List[Any]: '''simple docstring''' snake_case : DoubleLinkedList[T, U] = DoubleLinkedList() snake_case : Tuple = capacity snake_case : Dict = 0 snake_case : Tuple = 0 snake_case : int = 0 snake_case : dict[T, DoubleLinkedListNode[T, U]] = {} def __repr__(self : Optional[Any] ) -> str: '''simple docstring''' return ( f"""CacheInfo(hits={self.hits}, misses={self.miss}, """ f"""capacity={self.capacity}, current size={self.num_keys})""" ) def __contains__(self : Dict , snake_case__ : T ) -> bool: '''simple docstring''' return key in self.cache def _SCREAMING_SNAKE_CASE (self : Tuple , snake_case__ : T ) -> U | None: '''simple docstring''' if key in self.cache: self.hits += 1 snake_case : DoubleLinkedListNode[T, U] = self.cache[key] snake_case : Any = self.list.remove(self.cache[key] ) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(snake_case__ ) return node.val self.miss += 1 return None def _SCREAMING_SNAKE_CASE (self : int , snake_case__ : T , snake_case__ : U ) -> None: '''simple docstring''' if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity snake_case : Optional[int] = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(snake_case__ ) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 snake_case : Tuple = DoubleLinkedListNode(snake_case__ , snake_case__ ) self.list.add(self.cache[key] ) self.num_keys += 1 else: # bump node to the end of the list, update value snake_case : str = self.list.remove(self.cache[key] ) assert node is not None # node guaranteed to be in list snake_case : str = value self.list.add(snake_case__ ) @classmethod def _SCREAMING_SNAKE_CASE (cls : List[str] , snake_case__ : int = 1_28 ) -> Callable[[Callable[[T], U]], Callable[..., U]]: '''simple docstring''' def cache_decorator_inner(snake_case__ : Callable[[T], U] ) -> Callable[..., U]: def cache_decorator_wrapper(*snake_case__ : T ) -> U: if func not in cls.decorator_function_to_instance_map: snake_case : Optional[int] = LRUCache(snake_case__ ) snake_case : Any = cls.decorator_function_to_instance_map[func].get(args[0] ) if result is None: snake_case : Tuple = func(*snake_case__ ) cls.decorator_function_to_instance_map[func].put(args[0] , snake_case__ ) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(snake_case__ , "cache_info" , snake_case__ ) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
59
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __lowerCamelCase = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __lowerCamelCase = TaTokenizerFast __lowerCamelCase = {"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = ["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = ["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __lowerCamelCase = _LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
59
1
"""simple docstring""" import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase__ : Any = get_tests_dir('fixtures/spiece.model') @require_sentencepiece @require_tokenizers class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase : int = DebertaVaTokenizer __UpperCamelCase : List[str] = DebertaVaTokenizerFast __UpperCamelCase : Tuple = True __UpperCamelCase : Any = True def __magic_name__ ( self : Optional[int] ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing _A: int = DebertaVaTokenizer(lowerCAmelCase_ , unk_token='''<unk>''' ) tokenizer.save_pretrained(self.tmpdirname ) def __magic_name__ ( self : Optional[int] , lowerCAmelCase_ : List[Any] ): """simple docstring""" _A: Optional[int] = '''this is a test''' _A: Any = '''this is a test''' return input_text, output_text def __magic_name__ ( self : Union[str, Any] ): """simple docstring""" _A: Dict = '''<pad>''' _A: Optional[Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase_ ) , lowerCAmelCase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase_ ) , lowerCAmelCase_ ) def __magic_name__ ( self : Optional[Any] ): """simple docstring""" _A: int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''<unk>''' ) self.assertEqual(vocab_keys[-1] , '''[PAD]''' ) self.assertEqual(len(lowerCAmelCase_ ) , 3_0_0_0_1 ) def __magic_name__ ( self : List[str] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 3_0_0_0_0 ) def __magic_name__ ( self : List[Any] ): """simple docstring""" _A: List[str] = ''' \tHeLLo!how \n Are yoU? ''' _A: Optional[Any] = ['''▁hello''', '''!''', '''how''', '''▁are''', '''▁you''', '''?'''] # fmt: on _A: Optional[int] = DebertaVaTokenizer(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ ) _A: List[str] = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: Tuple = DebertaVaTokenizerFast(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ ) _A: Any = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @unittest.skip('''There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.''' ) def __magic_name__ ( self : List[Any] ): """simple docstring""" pass @unittest.skip('''There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.''' ) def __magic_name__ ( self : Dict ): """simple docstring""" pass def __magic_name__ ( self : Tuple ): """simple docstring""" _A: Optional[Any] = '''I was born in 92000, and this is falsé.''' _A: int = ['''▁''', '''<unk>''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', '''▁''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''▁''', '''.''', ] # fmt: on _A: Optional[Any] = DebertaVaTokenizer(lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: Union[str, Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: int = DebertaVaTokenizerFast(lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: Dict = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : Optional[int] ): """simple docstring""" _A: Optional[int] = '''I was born in 92000, and this is falsé.''' _A: Optional[Any] = ['''▁i''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', '''▁''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''▁''', '''.''', ] # fmt: on _A: Any = DebertaVaTokenizer(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: Any = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: str = DebertaVaTokenizerFast(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: int = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : str ): """simple docstring""" _A: List[str] = '''I was born in 92000, and this is falsé.''' _A: Union[str, Any] = ['''▁i''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''.''', ] # fmt: on _A: Union[str, Any] = DebertaVaTokenizer(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: int = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: List[Any] = DebertaVaTokenizerFast(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: str = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : Tuple ): """simple docstring""" _A: str = '''I was born in 92000, and this is falsé.''' _A: List[str] = ['''▁''', '''<unk>''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', '''▁''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''▁''', '''.''', ] # fmt: on _A: Any = DebertaVaTokenizer(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: Dict = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: int = DebertaVaTokenizerFast(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: Optional[Any] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : Dict ): """simple docstring""" _A: str = ''' \tHeLLo!how \n Are yoU? ''' _A: Union[str, Any] = ['''▁''', '''<unk>''', '''e''', '''<unk>''', '''o''', '''!''', '''how''', '''▁''', '''<unk>''', '''re''', '''▁yo''', '''<unk>''', '''?'''] # fmt: on _A: int = DebertaVaTokenizer(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: Union[str, Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: List[Any] = DebertaVaTokenizerFast(lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , split_by_punct=lowerCAmelCase_ ) _A: Optional[int] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : Dict ): """simple docstring""" _A: Any = self.get_tokenizer() _A: List[str] = self.get_rust_tokenizer() _A: str = '''I was born in 92000, and this is falsé.''' _A: Optional[Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) _A: List[str] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: Optional[int] = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) _A: Optional[Any] = rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: str = self.get_rust_tokenizer() _A: int = tokenizer.encode(lowerCAmelCase_ ) _A: Any = rust_tokenizer.encode(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : str ): """simple docstring""" _A: Optional[Any] = '''This is a test''' _A: List[Any] = [1_3, 1, 4_3_9_8, 2_5, 2_1, 1_2_8_9] _A: Dict = ['''▁''', '''T''', '''his''', '''▁is''', '''▁a''', '''▁test'''] _A: Optional[int] = ['''▁''', '''<unk>''', '''his''', '''▁is''', '''▁a''', '''▁test'''] _A: Union[str, Any] = DebertaVaTokenizer(lowerCAmelCase_ , keep_accents=lowerCAmelCase_ ) _A: Tuple = DebertaVaTokenizerFast(lowerCAmelCase_ , keep_accents=lowerCAmelCase_ ) _A: Optional[int] = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: List[str] = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: int = tokenizer.convert_ids_to_tokens(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: Dict = rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: int = rust_tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: int = rust_tokenizer.convert_ids_to_tokens(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) # fmt: off _A: Tuple = '''I was born in 92000, and this is falsé.''' _A: str = [1_3, 1, 2_3, 3_8_6, 1_9, 5_6_1, 3_0_5_0, 1_5, 1_7, 4_8, 2_5, 8_2_5_6, 1_8, 1, 9] _A: str = ['''▁''', '''I''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''é''', '''.''', ] _A: str = ['''▁''', '''<unk>''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''.''', ] # fmt: on _A: Optional[int] = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: Tuple = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: str = tokenizer.convert_ids_to_tokens(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: Union[str, Any] = rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: List[str] = rust_tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) _A: Tuple = rust_tokenizer.convert_ids_to_tokens(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : str ): """simple docstring""" _A: Tuple = DebertaVaTokenizer(lowerCAmelCase_ ) _A: List[Any] = tokenizer.encode('''sequence builders''' ) _A: Tuple = tokenizer.encode('''multi-sequence build''' ) _A: Union[str, Any] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) _A: Optional[int] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] , lowerCAmelCase_ ) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id] , lowerCAmelCase_ , ) @slow def __magic_name__ ( self : Tuple ): """simple docstring""" _A: Any = {'''input_ids''': [[1, 3_9_8_6_7, 3_6, 1_9_3_9_0, 4_8_6, 2_7, 3_5_0_5_2, 8_1_4_3_6, 1_8, 6_0_6_8_5, 1_2_2_5, 7, 3_5_0_5_2, 8_1_4_3_6, 1_8, 9_3_6_7, 1_6_8_9_9, 1_8, 1_5_9_3_7, 5_3, 5_9_4, 7_7_3, 1_8, 1_6_2_8_7, 3_0_4_6_5, 3_6, 1_5_9_3_7, 6, 4_1_1_3_9, 3_8, 3_6_9_7_9, 6_0_7_6_3, 1_9_1, 6, 3_4_1_3_2, 9_9, 6, 5_0_5_3_8, 3_9_0, 4_3_2_3_0, 6, 3_4_1_3_2, 2_7_7_9, 2_0_8_5_0, 1_4, 6_9_9, 1_0_7_2, 1_1_9_4, 3_6, 3_8_2, 1_0_9_0_1, 5_3, 7, 6_9_9, 1_0_7_2, 2_0_8_4, 3_6, 2_0_4_2_2, 6_3_0, 5_3, 1_9, 1_0_5, 3_0_4_9, 1_8_9_6, 1_0_5_3, 1_6_8_9_9, 1_5_0_6, 1_1, 3_7_9_7_8, 4_2_4_3, 7, 1_2_3_7, 3_1_8_6_9, 2_0_0, 1_6_5_6_6, 6_5_4, 6, 3_5_0_5_2, 8_1_4_3_6, 7, 5_5_6_3_0, 1_3_5_9_3, 4, 2], [1, 2_6, 1_5_0_1_1, 1_3, 6_6_7, 8, 1_0_5_3, 1_8, 2_3_6_1_1, 1_2_3_7, 7_2_3_5_6, 1_2_8_2_0, 3_4, 1_0_4_1_3_4, 1_2_0_9, 3_5, 1_3_3_1_3, 6_6_2_7, 2_1, 2_0_2, 3_4_7, 7, 1_6_4, 2_3_9_9, 1_1, 4_6, 4_4_8_5, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 1_2_3_2, 2_8_6_4, 1_5_7_8_5, 1_4_9_5_1, 1_0_5, 5, 8_5_8_1, 1_2_5_0, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''token_type_ids''': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase_ , model_name='''microsoft/deberta-v2-xlarge''' , revision='''ad6e42c1532ddf3a15c39246b63f5559d558b670''' , )
352
from __future__ import annotations UpperCAmelCase__ : List[str] = list[list[int]] # assigning initial values to the grid UpperCAmelCase__ : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution UpperCAmelCase__ : Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def lowerCamelCase__ ( a , a , a , a ) -> bool: for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def lowerCamelCase__ ( a ) -> tuple[int, int] | None: for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def lowerCamelCase__ ( a ) -> Matrix | None: if location := find_empty_location(a ): _A , _A: Optional[Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(a , a , a , a ): _A: str = digit if sudoku(a ) is not None: return grid _A: Tuple = 0 return None def lowerCamelCase__ ( a ) -> None: for row in grid: for cell in row: print(a , end=''' ''' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('\nExample grid:\n' + '=' * 20) print_solution(example_grid) print('\nExample grid solution:') UpperCAmelCase__ : int = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('Cannot find a solution.')
301
0
'''simple docstring''' import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, PLBartTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin __A =get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.plbart.modeling_plbart import shift_tokens_right __A =5_00_03 __A =5_00_02 @require_sentencepiece @require_tokenizers class _snake_case ( a__ , unittest.TestCase ): lowerCAmelCase :Union[str, Any] = PLBartTokenizer lowerCAmelCase :Union[str, Any] = None lowerCAmelCase :Optional[int] = False def snake_case__ ( self): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ : Union[str, Any] = PLBartTokenizer(__lowerCamelCase , language_codes="""base""" , keep_accents=__lowerCamelCase) tokenizer.save_pretrained(self.tmpdirname) def snake_case__ ( self): UpperCAmelCase__ : List[Any] = PLBartTokenizer(__lowerCamelCase , language_codes="""base""" , keep_accents=__lowerCamelCase) UpperCAmelCase__ : List[Any] = tokenizer.tokenize("""This is a test""") self.assertListEqual(__lowerCamelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""]) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCamelCase) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) UpperCAmelCase__ : Dict = tokenizer.tokenize("""I was born in 92000, and this is falsé.""") self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase__ : Optional[int] = tokenizer.convert_tokens_to_ids(__lowerCamelCase) self.assertListEqual( __lowerCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) UpperCAmelCase__ : Union[str, Any] = tokenizer.convert_ids_to_tokens(__lowerCamelCase) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) UpperCAmelCase__ : Tuple = tokenizer.vocab_size UpperCAmelCase__ : int = [tokenizer.convert_ids_to_tokens(__lowerCamelCase) for x in range(end - 4 , __lowerCamelCase)] self.assertListEqual(__lowerCamelCase , ["""__java__""", """__python__""", """__en_XX__""", """<mask>"""]) UpperCAmelCase__ : Tuple = "java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" UpperCAmelCase__ : Dict = tokenizer(__lowerCamelCase).input_ids self.assertEqual( tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase) , __lowerCamelCase , ) def snake_case__ ( self): UpperCAmelCase__ : str = PLBartTokenizer(__lowerCamelCase , language_codes="""multi""" , keep_accents=__lowerCamelCase) UpperCAmelCase__ : Optional[Any] = tokenizer.tokenize("""This is a test""") self.assertListEqual(__lowerCamelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""]) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCamelCase) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) UpperCAmelCase__ : List[str] = tokenizer.tokenize("""I was born in 92000, and this is falsé.""") self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase__ : List[Any] = tokenizer.convert_tokens_to_ids(__lowerCamelCase) self.assertListEqual( __lowerCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) UpperCAmelCase__ : str = tokenizer.convert_ids_to_tokens(__lowerCamelCase) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) UpperCAmelCase__ : Dict = tokenizer.vocab_size UpperCAmelCase__ : List[Any] = [tokenizer.convert_ids_to_tokens(__lowerCamelCase) for x in range(end - 7 , __lowerCamelCase)] self.assertListEqual( __lowerCamelCase , ["""__java__""", """__python__""", """__en_XX__""", """__javascript__""", """__php__""", """__ruby__""", """__go__"""]) UpperCAmelCase__ : Tuple = "java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" UpperCAmelCase__ : Any = tokenizer(__lowerCamelCase).input_ids self.assertEqual( tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase) , __lowerCamelCase , ) @require_torch @require_sentencepiece @require_tokenizers class _snake_case ( unittest.TestCase ): lowerCAmelCase :List[Any] = '''uclanlp/plbart-python-en_XX''' lowerCAmelCase :Any = [ '''def maximum(a,b,c):NEW_LINE_INDENTreturn max([a,b,c])''', '''def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])''', ] lowerCAmelCase :Union[str, Any] = [ '''Returns the maximum value of a b c.''', '''Sums the values of a b c.''', ] lowerCAmelCase :Any = [ 134, 5_452, 33_460, 33_441, 33_463, 33_465, 33_463, 33_449, 988, 20, 33_456, 19, 33_456, 771, 39, 4_258, 889, 3_318, 33_441, 33_463, 33_465, 33_463, 33_449, 2_471, 2, PYTHON_CODE, ] @classmethod def snake_case__ ( cls): UpperCAmelCase__ : PLBartTokenizer = PLBartTokenizer.from_pretrained( cls.checkpoint_name , language_codes="""base""" , src_lang="""python""" , tgt_lang="""en_XX""") UpperCAmelCase__ : str = 1 return cls def snake_case__ ( self): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""__java__"""] , 5_0001) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""__python__"""] , 5_0002) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""__en_XX__"""] , 5_0003) def snake_case__ ( self): UpperCAmelCase__ : Optional[Any] = self.tokenizer.batch_encode_plus(self.src_text).input_ids[0] self.assertListEqual(self.expected_src_tokens , __lowerCamelCase) def snake_case__ ( self): self.assertIn(__lowerCamelCase , self.tokenizer.all_special_ids) UpperCAmelCase__ : int = [EN_CODE, 9037, 3_3442, 57, 752, 153, 14, 56, 18, 9, 2] UpperCAmelCase__ : Dict = self.tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase) UpperCAmelCase__ : Optional[Any] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__lowerCamelCase) self.assertEqual(__lowerCamelCase , __lowerCamelCase) self.assertNotIn(self.tokenizer.eos_token , __lowerCamelCase) def snake_case__ ( self): UpperCAmelCase__ : Optional[int] = ["def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])" * 20] self.assertIsInstance(src_text[0] , __lowerCamelCase) UpperCAmelCase__ : str = 10 UpperCAmelCase__ : str = self.tokenizer(__lowerCamelCase , max_length=__lowerCamelCase , truncation=__lowerCamelCase).input_ids[0] self.assertEqual(ids[-2] , 2) self.assertEqual(ids[-1] , __lowerCamelCase) self.assertEqual(len(__lowerCamelCase) , __lowerCamelCase) def snake_case__ ( self): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["""<mask>""", """__java__"""]) , [5_0004, 5_0001]) def snake_case__ ( self): UpperCAmelCase__ : Any = tempfile.mkdtemp() UpperCAmelCase__ : Dict = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(__lowerCamelCase) UpperCAmelCase__ : int = PLBartTokenizer.from_pretrained(__lowerCamelCase) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , __lowerCamelCase) @require_torch def snake_case__ ( self): UpperCAmelCase__ : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__lowerCamelCase , return_tensors="""pt""") UpperCAmelCase__ : Union[str, Any] = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 self.assertEqual(batch.input_ids[1][-2:].tolist() , [2, PYTHON_CODE]) self.assertEqual(batch.decoder_input_ids[1][0] , __lowerCamelCase) self.assertEqual(batch.decoder_input_ids[1][-1] , 2) self.assertEqual(batch.labels[1][-2:].tolist() , [2, EN_CODE]) @require_torch def snake_case__ ( self): UpperCAmelCase__ : List[Any] = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=len(self.expected_src_tokens) , return_tensors="""pt""" , ) UpperCAmelCase__ : Tuple = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase) self.assertEqual((2, 26) , batch.input_ids.shape) self.assertEqual((2, 26) , batch.attention_mask.shape) UpperCAmelCase__ : Tuple = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , __lowerCamelCase) self.assertEqual(2 , batch.decoder_input_ids[0, -1]) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , []) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, PYTHON_CODE]) def snake_case__ ( self): UpperCAmelCase__ : List[Any] = self.tokenizer(self.src_text , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=3 , return_tensors="""pt""") UpperCAmelCase__ : Dict = self.tokenizer( text_target=self.tgt_text , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=10 , return_tensors="""pt""") UpperCAmelCase__ : Any = targets["input_ids"] UpperCAmelCase__ : Union[str, Any] = shift_tokens_right(__lowerCamelCase , self.tokenizer.pad_token_id) self.assertEqual(batch.input_ids.shape[1] , 3) self.assertEqual(batch.decoder_input_ids.shape[1] , 10) @require_torch def snake_case__ ( self): UpperCAmelCase__ : str = self.tokenizer._build_translation_inputs( """A test""" , return_tensors="""pt""" , src_lang="""en_XX""" , tgt_lang="""java""") self.assertEqual( nested_simplify(__lowerCamelCase) , { # A, test, EOS, en_XX """input_ids""": [[150, 242, 2, 5_0003]], """attention_mask""": [[1, 1, 1, 1]], # java """forced_bos_token_id""": 5_0001, } , )
163
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False")) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env") @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 650, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "pytorch", "script": "run_ddp.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 600, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "tensorflow", "script": "run_tf_dist.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 600, "eval_accuracy": 0.6, "eval_loss": 0.7}, }, ]) class lowerCAmelCase__ ( unittest.TestCase): '''simple docstring''' def _lowerCamelCase ( self) -> str: if self.framework == "pytorch": subprocess.run( F"cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py".split() , encoding="utf-8" , check=__lowerCamelCase , ) assert hasattr(self , "env") def _lowerCamelCase ( self , __lowerCamelCase) -> Tuple: _A : Dict = F"{self.env.base_job_name}-{instance_count}-{'ddp' if 'ddp' in self.script else 'smd'}" # distributed data settings _A : Optional[Any] = {"smdistributed": {"dataparallel": {"enabled": True}}} if self.script != "run_ddp.py" else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=__lowerCamelCase , instance_count=__lowerCamelCase , instance_type=self.instance_type , debugger_hook_config=__lowerCamelCase , hyperparameters={**self.env.distributed_hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=__lowerCamelCase , py_version="py36" , ) def _lowerCamelCase ( self , __lowerCamelCase) -> Optional[Any]: TrainingJobAnalytics(__lowerCamelCase).export_csv(F"{self.env.test_path}/{job_name}_metrics.csv") @parameterized.expand([(2,)]) def _lowerCamelCase ( self , __lowerCamelCase) -> Any: # create estimator _A : Union[str, Any] = self.create_estimator(__lowerCamelCase) # run training estimator.fit() # result dataframe _A : Optional[Any] = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis _A : List[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) _A : Dict = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping _A : Optional[Any] = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 9_9_9_9_9_9) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"{estimator.latest_training_job.name}.json" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , __lowerCamelCase)
11
0
"""simple docstring""" import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup _SCREAMING_SNAKE_CASE = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36' ' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582' } def snake_case ( snake_case__ :str = "dhaka" , snake_case__ :int = 5) -> int: _A = min(snake_case__ , 50) # Prevent abuse! _A = { """q""": query, """tbm""": """isch""", """hl""": """en""", """ijn""": """0""", } _A = requests.get("""https://www.google.com/search""" , params=snake_case__ , headers=snake_case__) _A = BeautifulSoup(html.text , """html.parser""") _A = """""".join( re.findall(R"""AF_initDataCallback\(([^<]+)\);""" , str(soup.select("""script""")))) _A = json.dumps(snake_case__) _A = json.loads(snake_case__) _A = re.findall( R"""\[\"GRID_STATE0\",null,\[\[1,\[0,\".*?\",(.*),\"All\",""" , snake_case__ , ) if not matched_google_image_data: return 0 _A = re.sub( R"""\[\"(https\:\/\/encrypted-tbn0\.gstatic\.com\/images\?.*?)\",\d+,\d+\]""" , """""" , str(snake_case__) , ) _A = re.findall( R"""(?:'|,),\[\"(https:|http.*?)\",\d+,\d+\]""" , snake_case__ , ) for index, fixed_full_res_image in enumerate(snake_case__): if index >= max_images: return index _A = bytes(snake_case__ , """ascii""").decode( """unicode-escape""") _A = bytes(snake_case__ , """ascii""").decode( """unicode-escape""") _A = urllib.request.build_opener() _A = [ ( """User-Agent""", """Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36""" """ (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582""", ) ] urllib.request.install_opener(snake_case__) _A = F'''query_{query.replace(' ' , '_')}''' if not os.path.exists(snake_case__): os.makedirs(snake_case__) urllib.request.urlretrieve( # noqa: S310 snake_case__ , F'''{path_name}/original_size_img_{index}.jpg''') return index if __name__ == "__main__": try: _SCREAMING_SNAKE_CASE = download_images_from_google_query(sys.argv[1]) print(F'''{image_count} images were downloaded to disk.''') except IndexError: print('Please provide a search term.') raise
351
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :int = (UnCLIPScheduler,) def UpperCAmelCase ( self , **lowerCAmelCase_ ) -> List[Any]: _A = { """num_train_timesteps""": 10_00, """variance_type""": """fixed_small_log""", """clip_sample""": True, """clip_sample_range""": 1.0, """prediction_type""": """epsilon""", } config.update(**lowerCAmelCase_ ) return config def UpperCAmelCase ( self ) -> Union[str, Any]: for timesteps in [1, 5, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Union[str, Any]: for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> int: for clip_sample in [True, False]: self.check_over_configs(clip_sample=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> List[Any]: for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Any: for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Optional[int]: for time_step in [0, 5_00, 9_99]: for prev_timestep in [None, 5, 1_00, 2_50, 5_00, 7_50]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=lowerCAmelCase_ , prev_timestep=lowerCAmelCase_ ) def UpperCAmelCase ( self ) -> Union[str, Any]: _A = self.scheduler_classes[0] _A = self.get_scheduler_config(variance_type="""fixed_small_log""" ) _A = scheduler_class(**lowerCAmelCase_ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0000E-10 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(4_87 ) - 0.054_9625 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(9_99 ) - 0.999_4987 ) ) < 1E-5 def UpperCAmelCase ( self ) -> Optional[int]: _A = self.scheduler_classes[0] _A = self.get_scheduler_config(variance_type="""learned_range""" ) _A = scheduler_class(**lowerCAmelCase_ ) _A = 0.5 assert scheduler._get_variance(1 , predicted_variance=lowerCAmelCase_ ) - -10.171_2790 < 1E-5 assert scheduler._get_variance(4_87 , predicted_variance=lowerCAmelCase_ ) - -5.799_8052 < 1E-5 assert scheduler._get_variance(9_99 , predicted_variance=lowerCAmelCase_ ) - -0.001_0011 < 1E-5 def UpperCAmelCase ( self ) -> List[Any]: _A = self.scheduler_classes[0] _A = self.get_scheduler_config() _A = scheduler_class(**lowerCAmelCase_ ) _A = scheduler.timesteps _A = self.dummy_model() _A = self.dummy_sample_deter _A = torch.manual_seed(0 ) for i, t in enumerate(lowerCAmelCase_ ): # 1. predict noise residual _A = model(lowerCAmelCase_ , lowerCAmelCase_ ) # 2. predict previous mean of sample x_t-1 _A = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_ ).prev_sample _A = pred_prev_sample _A = torch.sum(torch.abs(lowerCAmelCase_ ) ) _A = torch.mean(torch.abs(lowerCAmelCase_ ) ) assert abs(result_sum.item() - 252.268_2495 ) < 1E-2 assert abs(result_mean.item() - 0.328_4743 ) < 1E-3 def UpperCAmelCase ( self ) -> Optional[int]: _A = self.scheduler_classes[0] _A = self.get_scheduler_config() _A = scheduler_class(**lowerCAmelCase_ ) scheduler.set_timesteps(25 ) _A = scheduler.timesteps _A = self.dummy_model() _A = self.dummy_sample_deter _A = torch.manual_seed(0 ) for i, t in enumerate(lowerCAmelCase_ ): # 1. predict noise residual _A = model(lowerCAmelCase_ , lowerCAmelCase_ ) if i + 1 == timesteps.shape[0]: _A = None else: _A = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 _A = scheduler.step( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , prev_timestep=lowerCAmelCase_ , generator=lowerCAmelCase_ ).prev_sample _A = pred_prev_sample _A = torch.sum(torch.abs(lowerCAmelCase_ ) ) _A = torch.mean(torch.abs(lowerCAmelCase_ ) ) assert abs(result_sum.item() - 258.204_4983 ) < 1E-2 assert abs(result_mean.item() - 0.336_2038 ) < 1E-3 def UpperCAmelCase ( self ) -> Dict: pass def UpperCAmelCase ( self ) -> List[Any]: pass
81
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __magic_name__ = { "configuration_graphormer": ["GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "GraphormerConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ "GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "GraphormerForGraphClassification", "GraphormerModel", "GraphormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys __magic_name__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
100
"""simple docstring""" from __future__ import annotations from fractions import Fraction def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = 11 __SCREAMING_SNAKE_CASE = int("""1""" + """0""" * digit_len ) for num in range(UpperCamelCase_ , UpperCamelCase_ ): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(UpperCamelCase_ , UpperCamelCase_ ): solutions.append(f"{num}/{den}" ) den += 1 num += 1 __SCREAMING_SNAKE_CASE = 10 return solutions def _lowerCAmelCase ( UpperCamelCase_ = 2 ): __SCREAMING_SNAKE_CASE = 1.0 for fraction in fraction_list(UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = Fraction(UpperCamelCase_ ) result *= frac.denominator / frac.numerator return int(UpperCamelCase_ ) if __name__ == "__main__": print(solution())
100
1
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class a__ : def __init__( self : Union[str, Any], lowerCAmelCase : Any, lowerCAmelCase : Tuple=13, lowerCAmelCase : List[Any]=2, lowerCAmelCase : Tuple=24, lowerCAmelCase : Any=16, lowerCAmelCase : Optional[Any]=True, lowerCAmelCase : Tuple=True, lowerCAmelCase : Optional[int]=32, lowerCAmelCase : Optional[int]=5, lowerCAmelCase : Optional[int]=4, lowerCAmelCase : Optional[int]=37, lowerCAmelCase : Tuple="gelu", lowerCAmelCase : str=0.1, lowerCAmelCase : Tuple=0.1, lowerCAmelCase : List[Any]=10, lowerCAmelCase : List[Any]=0.02, lowerCAmelCase : List[str]=None, lowerCAmelCase : Any=2, lowerCAmelCase : str=2, ) -> Union[str, Any]: lowercase : str = parent lowercase : Optional[int] = batch_size lowercase : str = patch_size lowercase : List[Any] = max_length lowercase : Optional[Any] = num_mel_bins lowercase : int = is_training lowercase : Dict = use_labels lowercase : List[str] = hidden_size lowercase : str = num_hidden_layers lowercase : Any = num_attention_heads lowercase : List[str] = intermediate_size lowercase : int = hidden_act lowercase : Optional[Any] = hidden_dropout_prob lowercase : Optional[Any] = attention_probs_dropout_prob lowercase : int = type_sequence_label_size lowercase : Optional[int] = initializer_range lowercase : int = scope lowercase : int = frequency_stride lowercase : Dict = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) lowercase : Tuple = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 lowercase : Dict = (self.max_length - self.patch_size) // self.time_stride + 1 lowercase : Any = frequency_out_dimension * time_out_dimension lowercase : List[str] = num_patches + 2 def lowercase ( self : int ) -> Optional[int]: lowercase : List[Any] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) lowercase : List[Any] = None if self.use_labels: lowercase : Optional[Any] = ids_tensor([self.batch_size], self.type_sequence_label_size ) lowercase : str = self.get_config() return config, input_values, labels def lowercase ( self : List[str] ) -> Any: return ASTConfig( patch_size=self.patch_size, max_length=self.max_length, num_mel_bins=self.num_mel_bins, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, is_decoder=lowerCAmelCase, initializer_range=self.initializer_range, frequency_stride=self.frequency_stride, time_stride=self.time_stride, ) def lowercase ( self : str, lowerCAmelCase : List[Any], lowerCAmelCase : Optional[Any], lowerCAmelCase : Union[str, Any] ) -> Optional[int]: lowercase : Any = ASTModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() lowercase : Any = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase ( self : Any ) -> Tuple: lowercase : List[Any] = self.prepare_config_and_inputs() ( lowercase ) : Dict = config_and_inputs lowercase : Union[str, Any] = {'input_values': input_values} return config, inputs_dict @require_torch class a__ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, unittest.TestCase ): _lowerCamelCase = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) _lowerCamelCase = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def lowercase ( self : Any, lowerCAmelCase : Any, lowerCAmelCase : Tuple, lowerCAmelCase : Dict, lowerCAmelCase : List[str], lowerCAmelCase : int ) -> Tuple: if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowercase ( self : Optional[Any] ) -> Dict: lowercase : List[Any] = ASTModelTester(self ) lowercase : Any = ConfigTester(self, config_class=lowerCAmelCase, has_text_modality=lowerCAmelCase, hidden_size=37 ) def lowercase ( self : Tuple ) -> Optional[int]: self.config_tester.run_common_tests() @unittest.skip(reason='AST does not use inputs_embeds' ) def lowercase ( self : Tuple ) -> List[Any]: pass def lowercase ( self : Union[str, Any] ) -> List[str]: lowercase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Optional[Any] = model_class(lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings(), (nn.Module) ) lowercase : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase, nn.Linear ) ) def lowercase ( self : Union[str, Any] ) -> Optional[Any]: lowercase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Optional[int] = model_class(lowerCAmelCase ) lowercase : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase : List[Any] = [*signature.parameters.keys()] lowercase : str = ['input_values'] self.assertListEqual(arg_names[:1], lowerCAmelCase ) def lowercase ( self : Optional[int] ) -> Tuple: lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) @slow def lowercase ( self : List[str] ) -> Optional[Any]: for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase : Dict = ASTModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) def lowercase__ ( ) -> Any: '''simple docstring''' lowercase : Tuple = hf_hub_download( repo_id='nielsr/audio-spectogram-transformer-checkpoint' , filename='sample_audio.flac' , repo_type='dataset' ) lowercase : List[str] = torchaudio.load(_UpperCAmelCase ) return audio, sampling_rate @require_torch @require_torchaudio class a__ ( unittest.TestCase ): @cached_property def lowercase ( self : Union[str, Any] ) -> Optional[int]: return ( ASTFeatureExtractor.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ) if is_torchaudio_available() else None ) @slow def lowercase ( self : Any ) -> Optional[Any]: lowercase : List[str] = self.default_feature_extractor lowercase : Tuple = ASTForAudioClassification.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ).to(lowerCAmelCase ) lowercase : List[str] = self.default_feature_extractor lowercase : Optional[int] = prepare_audio() lowercase : List[str] = audio.squeeze().numpy() lowercase : List[Any] = feature_extractor(lowerCAmelCase, sampling_rate=lowerCAmelCase, return_tensors='pt' ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): lowercase : List[Any] = model(**lowerCAmelCase ) # verify the logits lowercase : Union[str, Any] = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape, lowerCAmelCase ) lowercase : Any = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3], lowerCAmelCase, atol=1e-4 ) )
365
"""simple docstring""" import unittest from transformers import DonutProcessor _UpperCamelCase: Any = 'naver-clova-ix/donut-base' class a__ ( unittest.TestCase ): def lowercase ( self : Optional[Any] ) -> Tuple: lowercase : Any = DonutProcessor.from_pretrained(lowerCAmelCase ) def lowercase ( self : Dict ) -> Union[str, Any]: lowercase : Tuple = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowercase : Tuple = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowercase : Any = self.processor.tokenajson(lowerCAmelCase ) self.assertDictEqual(lowerCAmelCase, lowerCAmelCase )
53
0
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging _snake_case = { "cola": 2, "mnli": 3, "mrpc": 2, "sst-2": 2, "sts-b": 1, "qqp": 2, "qnli": 2, "rte": 2, "wnli": 2, } logging.set_verbosity_info() def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_,snake_case_=None ): _A : Optional[int] = XLNetConfig.from_json_file(lowerCamelCase__ ) _A : int = finetuning_task.lower() if finetuning_task is not None else """""" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'''Building PyTorch XLNetForSequenceClassification model from configuration: {config}''' ) _A : str = finetuning_task _A : str = GLUE_TASKS_NUM_LABELS[finetuning_task] _A : List[str] = XLNetForSequenceClassification(lowerCamelCase__ ) elif "squad" in finetuning_task: _A : Union[str, Any] = finetuning_task _A : List[str] = XLNetForQuestionAnswering(lowerCamelCase__ ) else: _A : int = XLNetLMHeadModel(lowerCamelCase__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(lowerCamelCase__,lowerCamelCase__,lowerCamelCase__ ) # Save pytorch-model _A : Tuple = os.path.join(lowerCamelCase__,lowerCamelCase__ ) _A : Optional[Any] = os.path.join(lowerCamelCase__,lowerCamelCase__ ) print(f'''Save PyTorch model to {os.path.abspath(lowerCamelCase__ )}''' ) torch.save(model.state_dict(),lowerCamelCase__ ) print(f'''Save configuration file to {os.path.abspath(lowerCamelCase__ )}''' ) with open(lowerCamelCase__,"""w""",encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--xlnet_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained XLNet model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the folder to store the PyTorch model or dataset/vocab.", ) parser.add_argument( "--finetuning_task", default=None, type=str, help="Name of a task on which the XLNet TensorFlow model was fine-tuned", ) _snake_case = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
26
import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values UpperCAmelCase : str = argparse.ArgumentParser() parser.add_argument("--user", type=str, default="ubuntu") parser.add_argument("--host", type=str, default="localhost") parser.add_argument("--key_path", type=str, default=None) parser.add_argument("--instance", type=str, default="V100:1") parser.add_argument("--provider", type=str, default="cheapest") parser.add_argument("--use_spot", type=bool, default=False) parser.add_argument("--example", type=str, default="pytorch/text-generation/run_generation.py") UpperCAmelCase, UpperCAmelCase : Optional[Any] = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError("Cannot specify both BYO and on-demand cluster args") UpperCAmelCase : Dict = rh.cluster( name="rh-cluster", ips=[args.host], ssh_creds={"ssh_user": args.user, "ssh_private_key": args.key_path} ) else: UpperCAmelCase : str = rh.cluster( name="rh-cluster", instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) UpperCAmelCase : str = args.example.rsplit("/", 1)[0] # Set up remote environment cluster.install_packages(["pip:./"]) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([f"""pip install -r transformers/examples/{example_dir}/requirements.txt"""]) cluster.run(["pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117"]) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([f"""python transformers/examples/{args.example} {" ".join(shlex.quote(arg) for arg in unknown)}"""]) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
252
0
'''simple docstring''' from datetime import datetime as dt import os from github import Github __A : int = [ "good first issue", "good second issue", "good difficult issue", "feature request", "new model", "wip", ] def UpperCamelCase_ ( ): '''simple docstring''' lowerCAmelCase_ : Dict = Github(os.environ["""GITHUB_TOKEN"""] ) lowerCAmelCase_ : Any = g.get_repo("""huggingface/transformers""" ) lowerCAmelCase_ : Any = repo.get_issues(state="""open""" ) for issue in open_issues: lowerCAmelCase_ : Union[str, Any] = sorted([comment for comment in issue.get_comments()] , key=lambda A__ : i.created_at , reverse=A__ ) lowerCAmelCase_ : Union[str, Any] = comments[0] if len(A__ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="""closed""" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""" ) if __name__ == "__main__": main()
89
'''simple docstring''' from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Dict = logging.get_logger(__name__) __A : int = { "snap-research/efficientformer-l1-300": ( "https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json" ), } class __snake_case ( _SCREAMING_SNAKE_CASE): """simple docstring""" lowercase = 'efficientformer' def __init__( self : Any , lowerCamelCase : List[int] = [3, 2, 6, 4] , lowerCamelCase : List[int] = [48, 96, 2_24, 4_48] , lowerCamelCase : List[bool] = [True, True, True, True] , lowerCamelCase : int = 4_48 , lowerCamelCase : int = 32 , lowerCamelCase : int = 4 , lowerCamelCase : int = 7 , lowerCamelCase : int = 5 , lowerCamelCase : int = 8 , lowerCamelCase : int = 4 , lowerCamelCase : float = 0.0 , lowerCamelCase : int = 16 , lowerCamelCase : int = 3 , lowerCamelCase : int = 3 , lowerCamelCase : int = 3 , lowerCamelCase : int = 2 , lowerCamelCase : int = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : int = 1 , lowerCamelCase : bool = True , lowerCamelCase : bool = True , lowerCamelCase : float = 1E-5 , lowerCamelCase : str = "gelu" , lowerCamelCase : float = 0.02 , lowerCamelCase : float = 1E-12 , lowerCamelCase : int = 2_24 , lowerCamelCase : float = 1E-05 , **lowerCamelCase : int , ) -> None: super().__init__(**lowerCamelCase ) lowerCAmelCase_ : List[str] = hidden_act lowerCAmelCase_ : List[Any] = hidden_dropout_prob lowerCAmelCase_ : Any = hidden_sizes lowerCAmelCase_ : List[Any] = num_hidden_layers lowerCAmelCase_ : Union[str, Any] = num_attention_heads lowerCAmelCase_ : Tuple = initializer_range lowerCAmelCase_ : Union[str, Any] = layer_norm_eps lowerCAmelCase_ : int = patch_size lowerCAmelCase_ : List[str] = num_channels lowerCAmelCase_ : Dict = depths lowerCAmelCase_ : int = mlp_expansion_ratio lowerCAmelCase_ : Optional[Any] = downsamples lowerCAmelCase_ : Union[str, Any] = dim lowerCAmelCase_ : Union[str, Any] = key_dim lowerCAmelCase_ : str = attention_ratio lowerCAmelCase_ : Tuple = resolution lowerCAmelCase_ : Optional[Any] = pool_size lowerCAmelCase_ : str = downsample_patch_size lowerCAmelCase_ : Dict = downsample_stride lowerCAmelCase_ : str = downsample_pad lowerCAmelCase_ : str = drop_path_rate lowerCAmelCase_ : List[Any] = num_metaad_blocks lowerCAmelCase_ : Tuple = distillation lowerCAmelCase_ : Optional[Any] = use_layer_scale lowerCAmelCase_ : Dict = layer_scale_init_value lowerCAmelCase_ : Optional[Any] = image_size lowerCAmelCase_ : Optional[Any] = batch_norm_eps
89
1
'''simple docstring''' import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def _UpperCamelCase ( __A ) -> Union[str, Any]: '''simple docstring''' return EnvironmentCommand() class lowercase_ ( a__ ): @staticmethod def __a ( a ): UpperCamelCase__ = parser.add_parser("env" ) download_parser.set_defaults(func=a ) def __a ( self ): UpperCamelCase__ = huggingface_hub.__version__ UpperCamelCase__ = "not installed" UpperCamelCase__ = "NA" if is_torch_available(): import torch UpperCamelCase__ = torch.__version__ UpperCamelCase__ = torch.cuda.is_available() UpperCamelCase__ = "not installed" if is_transformers_available(): import transformers UpperCamelCase__ = transformers.__version__ UpperCamelCase__ = "not installed" if is_accelerate_available(): import accelerate UpperCamelCase__ = accelerate.__version__ UpperCamelCase__ = "not installed" if is_xformers_available(): import xformers UpperCamelCase__ = xformers.__version__ UpperCamelCase__ = { "`diffusers` version": version, "Platform": platform.platform(), "Python version": platform.python_version(), "PyTorch version (GPU?)": f'''{pt_version} ({pt_cuda_available})''', "Huggingface_hub version": hub_version, "Transformers version": transformers_version, "Accelerate version": accelerate_version, "xFormers version": xformers_version, "Using GPU in script?": "<fill in>", "Using distributed or parallel set-up in script?": "<fill in>", } print("\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n" ) print(self.format_dict(a ) ) return info @staticmethod def __a ( a ): return "\n".join([f'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
80
'''simple docstring''' import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available A__: int = logging.getLogger(__name__) @dataclass class A__ : __UpperCamelCase : str __UpperCamelCase : List[str] __UpperCamelCase : Optional[List[str]] @dataclass class A__ : __UpperCamelCase : List[int] __UpperCamelCase : List[int] __UpperCamelCase : Optional[List[int]] = None __UpperCamelCase : Optional[List[int]] = None class A__ ( UpperCAmelCase__ ): __UpperCamelCase : str = "train" __UpperCamelCase : Tuple = "dev" __UpperCamelCase : str = "test" class A__ : @staticmethod def __UpperCAmelCase ( SCREAMING_SNAKE_CASE :Dict , SCREAMING_SNAKE_CASE :Union[Split, str] ) -> List[InputExample]: '''simple docstring''' raise NotImplementedError @staticmethod def __UpperCAmelCase ( SCREAMING_SNAKE_CASE :str ) -> List[str]: '''simple docstring''' raise NotImplementedError @staticmethod def __UpperCAmelCase ( SCREAMING_SNAKE_CASE :List[InputExample] , SCREAMING_SNAKE_CASE :List[str] , SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :PreTrainedTokenizer , SCREAMING_SNAKE_CASE :str=False , SCREAMING_SNAKE_CASE :Optional[Any]="[CLS]" , SCREAMING_SNAKE_CASE :Optional[int]=1 , SCREAMING_SNAKE_CASE :Any="[SEP]" , SCREAMING_SNAKE_CASE :List[Any]=False , SCREAMING_SNAKE_CASE :Union[str, Any]=False , SCREAMING_SNAKE_CASE :List[str]=0 , SCREAMING_SNAKE_CASE :str=0 , SCREAMING_SNAKE_CASE :Dict=-1_0_0 , SCREAMING_SNAKE_CASE :Optional[int]=0 , SCREAMING_SNAKE_CASE :Tuple=True , ) -> List[InputFeatures]: '''simple docstring''' _a : str ={label: i for i, label in enumerate(SCREAMING_SNAKE_CASE )} _a : Tuple =[] for ex_index, example in enumerate(SCREAMING_SNAKE_CASE ): if ex_index % 1_0_0_0_0 == 0: logger.info("""Writing example %d of %d""" , SCREAMING_SNAKE_CASE , len(SCREAMING_SNAKE_CASE ) ) _a : Optional[Any] =[] _a : List[Any] =[] for word, label in zip(example.words , example.labels ): _a : Optional[int] =tokenizer.tokenize(SCREAMING_SNAKE_CASE ) # bert-base-multilingual-cased sometimes output "nothing ([]) when calling tokenize with just a space. if len(SCREAMING_SNAKE_CASE ) > 0: tokens.extend(SCREAMING_SNAKE_CASE ) # Use the real label id for the first token of the word, and padding ids for the remaining tokens label_ids.extend([label_map[label]] + [pad_token_label_id] * (len(SCREAMING_SNAKE_CASE ) - 1) ) # Account for [CLS] and [SEP] with "- 2" and with "- 3" for RoBERTa. _a : Optional[int] =tokenizer.num_special_tokens_to_add() if len(SCREAMING_SNAKE_CASE ) > max_seq_length - special_tokens_count: _a : List[Any] =tokens[: (max_seq_length - special_tokens_count)] _a : Tuple =label_ids[: (max_seq_length - special_tokens_count)] # The convention in BERT is: # (a) For sequence pairs: # tokens: [CLS] is this jack ##son ##ville ? [SEP] no it is not . [SEP] # type_ids: 0 0 0 0 0 0 0 0 1 1 1 1 1 1 # (b) For single sequences: # tokens: [CLS] the dog is hairy . [SEP] # type_ids: 0 0 0 0 0 0 0 # # Where "type_ids" are used to indicate whether this is the first # sequence or the second sequence. The embedding vectors for `type=0` and # `type=1` were learned during pre-training and are added to the wordpiece # embedding vector (and position vector). This is not *strictly* necessary # since the [SEP] token unambiguously separates the sequences, but it makes # it easier for the model to learn the concept of sequences. # # For classification tasks, the first vector (corresponding to [CLS]) is # used as the "sentence vector". Note that this only makes sense because # the entire model is fine-tuned. tokens += [sep_token] label_ids += [pad_token_label_id] if sep_token_extra: # roberta uses an extra separator b/w pairs of sentences tokens += [sep_token] label_ids += [pad_token_label_id] _a : Dict =[sequence_a_segment_id] * len(SCREAMING_SNAKE_CASE ) if cls_token_at_end: tokens += [cls_token] label_ids += [pad_token_label_id] segment_ids += [cls_token_segment_id] else: _a : Any =[cls_token] + tokens _a : Dict =[pad_token_label_id] + label_ids _a : Union[str, Any] =[cls_token_segment_id] + segment_ids _a : List[str] =tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE ) # The mask has 1 for real tokens and 0 for padding tokens. Only real # tokens are attended to. _a : Optional[int] =[1 if mask_padding_with_zero else 0] * len(SCREAMING_SNAKE_CASE ) # Zero-pad up to the sequence length. _a : Union[str, Any] =max_seq_length - len(SCREAMING_SNAKE_CASE ) if pad_on_left: _a : Optional[Any] =([pad_token] * padding_length) + input_ids _a : Optional[int] =([0 if mask_padding_with_zero else 1] * padding_length) + input_mask _a : Union[str, Any] =([pad_token_segment_id] * padding_length) + segment_ids _a : Dict =([pad_token_label_id] * padding_length) + label_ids else: input_ids += [pad_token] * padding_length input_mask += [0 if mask_padding_with_zero else 1] * padding_length segment_ids += [pad_token_segment_id] * padding_length label_ids += [pad_token_label_id] * padding_length assert len(SCREAMING_SNAKE_CASE ) == max_seq_length assert len(SCREAMING_SNAKE_CASE ) == max_seq_length assert len(SCREAMING_SNAKE_CASE ) == max_seq_length assert len(SCREAMING_SNAKE_CASE ) == max_seq_length if ex_index < 5: logger.info("""*** Example ***""" ) logger.info("""guid: %s""" , example.guid ) logger.info("""tokens: %s""" , """ """.join([str(SCREAMING_SNAKE_CASE ) for x in tokens] ) ) logger.info("""input_ids: %s""" , """ """.join([str(SCREAMING_SNAKE_CASE ) for x in input_ids] ) ) logger.info("""input_mask: %s""" , """ """.join([str(SCREAMING_SNAKE_CASE ) for x in input_mask] ) ) logger.info("""segment_ids: %s""" , """ """.join([str(SCREAMING_SNAKE_CASE ) for x in segment_ids] ) ) logger.info("""label_ids: %s""" , """ """.join([str(SCREAMING_SNAKE_CASE ) for x in label_ids] ) ) if "token_type_ids" not in tokenizer.model_input_names: _a : Tuple =None features.append( InputFeatures( input_ids=SCREAMING_SNAKE_CASE , attention_mask=SCREAMING_SNAKE_CASE , token_type_ids=SCREAMING_SNAKE_CASE , label_ids=SCREAMING_SNAKE_CASE ) ) return features if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset class A__ ( UpperCAmelCase__ ): __UpperCamelCase : List[InputFeatures] __UpperCamelCase : int = nn.CrossEntropyLoss().ignore_index def __init__( self :Dict , SCREAMING_SNAKE_CASE :TokenClassificationTask , SCREAMING_SNAKE_CASE :str , SCREAMING_SNAKE_CASE :PreTrainedTokenizer , SCREAMING_SNAKE_CASE :List[str] , SCREAMING_SNAKE_CASE :str , SCREAMING_SNAKE_CASE :Optional[int] = None , SCREAMING_SNAKE_CASE :int=False , SCREAMING_SNAKE_CASE :Split = Split.train , ) -> List[str]: '''simple docstring''' # Load data features from cache or dataset file _a : Optional[Any] =os.path.join( SCREAMING_SNAKE_CASE , """cached_{}_{}_{}""".format(mode.value , tokenizer.__class__.__name__ , str(SCREAMING_SNAKE_CASE ) ) , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _a : List[str] =cached_features_file + """.lock""" with FileLock(SCREAMING_SNAKE_CASE ): if os.path.exists(SCREAMING_SNAKE_CASE ) and not overwrite_cache: logger.info(f"Loading features from cached file {cached_features_file}" ) _a : Any =torch.load(SCREAMING_SNAKE_CASE ) else: logger.info(f"Creating features from dataset file at {data_dir}" ) _a : Any =token_classification_task.read_examples_from_file(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # TODO clean up all this to leverage built-in features of tokenizers _a : List[str] =token_classification_task.convert_examples_to_features( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , cls_token_at_end=bool(model_type in ["""xlnet"""] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ["""xlnet"""] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=SCREAMING_SNAKE_CASE , pad_on_left=bool(tokenizer.padding_side == """left""" ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info(f"Saving features into cached file {cached_features_file}" ) torch.save(self.features , SCREAMING_SNAKE_CASE ) def __len__( self :Optional[int] ) -> Union[str, Any]: '''simple docstring''' return len(self.features ) def __getitem__( self :Dict , SCREAMING_SNAKE_CASE :int ) -> InputFeatures: '''simple docstring''' return self.features[i] if is_tf_available(): import tensorflow as tf class A__ : __UpperCamelCase : List[InputFeatures] __UpperCamelCase : int = -100 def __init__( self :str , SCREAMING_SNAKE_CASE :TokenClassificationTask , SCREAMING_SNAKE_CASE :str , SCREAMING_SNAKE_CASE :PreTrainedTokenizer , SCREAMING_SNAKE_CASE :List[str] , SCREAMING_SNAKE_CASE :str , SCREAMING_SNAKE_CASE :Optional[int] = None , SCREAMING_SNAKE_CASE :str=False , SCREAMING_SNAKE_CASE :Split = Split.train , ) -> Any: '''simple docstring''' _a : Tuple =token_classification_task.read_examples_from_file(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # TODO clean up all this to leverage built-in features of tokenizers _a : List[Any] =token_classification_task.convert_examples_to_features( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , cls_token_at_end=bool(model_type in ["""xlnet"""] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ["""xlnet"""] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=SCREAMING_SNAKE_CASE , pad_on_left=bool(tokenizer.padding_side == """left""" ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) def gen(): for ex in self.features: if ex.token_type_ids is None: yield ( {"input_ids": ex.input_ids, "attention_mask": ex.attention_mask}, ex.label_ids, ) else: yield ( { "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label_ids, ) if "token_type_ids" not in tokenizer.model_input_names: _a : Union[str, Any] =tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({"""input_ids""": tf.intaa, """attention_mask""": tf.intaa}, tf.intaa) , ( {"""input_ids""": tf.TensorShape([None] ), """attention_mask""": tf.TensorShape([None] )}, tf.TensorShape([None] ), ) , ) else: _a : Union[str, Any] =tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({"""input_ids""": tf.intaa, """attention_mask""": tf.intaa, """token_type_ids""": tf.intaa}, tf.intaa) , ( { """input_ids""": tf.TensorShape([None] ), """attention_mask""": tf.TensorShape([None] ), """token_type_ids""": tf.TensorShape([None] ), }, tf.TensorShape([None] ), ) , ) def __UpperCAmelCase ( self :Tuple ) -> Any: '''simple docstring''' _a : List[Any] =self.dataset.apply(tf.data.experimental.assert_cardinality(len(self.features ) ) ) return self.dataset def __len__( self :str ) -> Optional[int]: '''simple docstring''' return len(self.features ) def __getitem__( self :int , SCREAMING_SNAKE_CASE :str ) -> InputFeatures: '''simple docstring''' return self.features[i]
276
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class lowercase_ : '''simple docstring''' def __init__( self : str , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Any=2 , _UpperCAmelCase : Any=True , _UpperCAmelCase : str=False , _UpperCAmelCase : Union[str, Any]=10 , _UpperCAmelCase : Optional[Any]=3 , _UpperCAmelCase : Dict=32 * 8 , _UpperCAmelCase : Optional[int]=32 * 8 , _UpperCAmelCase : List[str]=4 , _UpperCAmelCase : Any=64 , ): _A = parent _A = batch_size _A = is_training _A = use_auxiliary_loss _A = num_queries _A = num_channels _A = min_size _A = max_size _A = num_labels _A = hidden_dim _A = hidden_dim def lowerCAmelCase_ ( self : List[Any] ): _A = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _UpperCAmelCase ) _A = torch.ones([self.batch_size, self.min_size, self.max_size] , device=_UpperCAmelCase ) _A = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=_UpperCAmelCase ) > 0.5 ).float() _A = (torch.rand((self.batch_size, self.num_labels) , device=_UpperCAmelCase ) > 0.5).long() _A = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowerCAmelCase_ ( self : Any ): _A = MaskaFormerConfig( hidden_size=self.hidden_dim , ) _A = self.num_queries _A = self.num_labels _A = [1, 1, 1, 1] _A = self.num_channels _A = 64 _A = 128 _A = self.hidden_dim _A = self.hidden_dim _A = self.hidden_dim return config def lowerCAmelCase_ ( self : Tuple ): _A , _A , _A , _A , _A = self.prepare_config_and_inputs() _A = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def lowerCAmelCase_ ( self : Any , _UpperCAmelCase : Dict , _UpperCAmelCase : List[Any] ): _A = output.encoder_hidden_states _A = output.pixel_decoder_hidden_states _A = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_UpperCAmelCase ) , config.decoder_layers ) def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple=False ): with torch.no_grad(): _A = MaskaFormerModel(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _A = model(pixel_values=_UpperCAmelCase , pixel_mask=_UpperCAmelCase ) _A = model(_UpperCAmelCase , output_hidden_states=_UpperCAmelCase ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_UpperCAmelCase , _UpperCAmelCase ) def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Dict , _UpperCAmelCase : List[Any] , _UpperCAmelCase : int , _UpperCAmelCase : str ): _A = MaskaFormerForUniversalSegmentation(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() def comm_check_on_output(_UpperCAmelCase : List[Any] ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _A = model(pixel_values=_UpperCAmelCase , pixel_mask=_UpperCAmelCase ) _A = model(_UpperCAmelCase ) comm_check_on_output(_UpperCAmelCase ) _A = model( pixel_values=_UpperCAmelCase , pixel_mask=_UpperCAmelCase , mask_labels=_UpperCAmelCase , class_labels=_UpperCAmelCase ) comm_check_on_output(_UpperCAmelCase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class lowercase_ ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : List[str] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () UpperCAmelCase : List[Any] = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} UpperCAmelCase : str = False UpperCAmelCase : List[str] = False UpperCAmelCase : List[str] = False UpperCAmelCase : Optional[Any] = False def lowerCAmelCase_ ( self : str ): _A = MaskaFormerModelTester(self ) _A = ConfigTester(self , config_class=_UpperCAmelCase , has_text_modality=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Tuple ): self.config_tester.run_common_tests() def lowerCAmelCase_ ( self : Optional[Any] ): _A , _A = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_UpperCAmelCase , **_UpperCAmelCase , output_hidden_states=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Dict ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_UpperCAmelCase ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def lowerCAmelCase_ ( self : List[str] ): pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def lowerCAmelCase_ ( self : Tuple ): pass @unittest.skip(reason='Mask2Former is not a generative model' ) def lowerCAmelCase_ ( self : int ): pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def lowerCAmelCase_ ( self : int ): pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def lowerCAmelCase_ ( self : Any ): pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def lowerCAmelCase_ ( self : Dict ): pass def lowerCAmelCase_ ( self : Tuple ): _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(_UpperCAmelCase ) _A = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _A = [*signature.parameters.keys()] _A = ['pixel_values'] self.assertListEqual(arg_names[:1] , _UpperCAmelCase ) @slow def lowerCAmelCase_ ( self : str ): for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _A = MaskaFormerModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] ): _A = (self.model_tester.min_size,) * 2 _A = { 'pixel_values': torch.randn((2, 3, *size) , device=_UpperCAmelCase ), 'mask_labels': torch.randn((2, 10, *size) , device=_UpperCAmelCase ), 'class_labels': torch.zeros(2 , 10 , device=_UpperCAmelCase ).long(), } _A = self.model_tester.get_config() _A = MaskaFormerForUniversalSegmentation(_UpperCAmelCase ).to(_UpperCAmelCase ) _A = model(**_UpperCAmelCase ) self.assertTrue(outputs.loss is not None ) def lowerCAmelCase_ ( self : Optional[Any] ): _A , _A = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_UpperCAmelCase , **_UpperCAmelCase , output_hidden_states=_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(_UpperCAmelCase ).to(_UpperCAmelCase ) _A = model(**_UpperCAmelCase , output_attentions=_UpperCAmelCase ) self.assertTrue(outputs.attentions is not None ) def lowerCAmelCase_ ( self : Optional[int] ): if not self.model_tester.is_training: return _A = self.all_model_classes[1] _A , _A , _A , _A , _A = self.model_tester.prepare_config_and_inputs() _A = model_class(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.train() _A = model(_UpperCAmelCase , mask_labels=_UpperCAmelCase , class_labels=_UpperCAmelCase ).loss loss.backward() def lowerCAmelCase_ ( self : str ): _A = self.all_model_classes[1] _A , _A , _A , _A , _A = self.model_tester.prepare_config_and_inputs() _A = True _A = True _A = model_class(_UpperCAmelCase ).to(_UpperCAmelCase ) model.train() _A = model(_UpperCAmelCase , mask_labels=_UpperCAmelCase , class_labels=_UpperCAmelCase ) _A = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _A = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _A = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _A = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_UpperCAmelCase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) a = 1e-4 def _snake_case ( ) -> Optional[Any]: '''simple docstring''' _A = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class lowercase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCAmelCase_ ( self : List[Any] ): return "facebook/mask2former-swin-small-coco-instance" @cached_property def lowerCAmelCase_ ( self : List[Any] ): return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def lowerCAmelCase_ ( self : Any ): _A = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_UpperCAmelCase ) _A = self.default_image_processor _A = prepare_img() _A = image_processor(_UpperCAmelCase , return_tensors='pt' ).to(_UpperCAmelCase ) _A = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_UpperCAmelCase , (1, 3, 384, 384) ) with torch.no_grad(): _A = model(**_UpperCAmelCase ) _A = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) _A = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) _A = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) def lowerCAmelCase_ ( self : List[Any] ): _A = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_UpperCAmelCase ).eval() _A = self.default_image_processor _A = prepare_img() _A = image_processor(_UpperCAmelCase , return_tensors='pt' ).to(_UpperCAmelCase ) _A = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_UpperCAmelCase , (1, 3, 384, 384) ) with torch.no_grad(): _A = model(**_UpperCAmelCase ) # masks_queries_logits _A = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _A = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _A = torch.tensor(_UpperCAmelCase ).to(_UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) # class_queries_logits _A = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) _A = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) def lowerCAmelCase_ ( self : Union[str, Any] ): _A = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_UpperCAmelCase ).eval() _A = self.default_image_processor _A = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors='pt' , ) _A = inputs['pixel_values'].to(_UpperCAmelCase ) _A = [el.to(_UpperCAmelCase ) for el in inputs['mask_labels']] _A = [el.to(_UpperCAmelCase ) for el in inputs['class_labels']] with torch.no_grad(): _A = model(**_UpperCAmelCase ) self.assertTrue(outputs.loss is not None )
358
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : "DiagonalGaussianDistribution" class lowercase_ ( __lowerCAmelCase , __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : List[Any] = True @register_to_config def __init__( self : List[str] , _UpperCAmelCase : int = 3 , _UpperCAmelCase : int = 3 , _UpperCAmelCase : Tuple[str] = ("DownEncoderBlock2D",) , _UpperCAmelCase : Tuple[str] = ("UpDecoderBlock2D",) , _UpperCAmelCase : Tuple[int] = (64,) , _UpperCAmelCase : int = 1 , _UpperCAmelCase : str = "silu" , _UpperCAmelCase : int = 4 , _UpperCAmelCase : int = 32 , _UpperCAmelCase : int = 32 , _UpperCAmelCase : float = 0.1_8215 , ): super().__init__() # pass init params to Encoder _A = Encoder( in_channels=_UpperCAmelCase , out_channels=_UpperCAmelCase , down_block_types=_UpperCAmelCase , block_out_channels=_UpperCAmelCase , layers_per_block=_UpperCAmelCase , act_fn=_UpperCAmelCase , norm_num_groups=_UpperCAmelCase , double_z=_UpperCAmelCase , ) # pass init params to Decoder _A = Decoder( in_channels=_UpperCAmelCase , out_channels=_UpperCAmelCase , up_block_types=_UpperCAmelCase , block_out_channels=_UpperCAmelCase , layers_per_block=_UpperCAmelCase , norm_num_groups=_UpperCAmelCase , act_fn=_UpperCAmelCase , ) _A = nn.Convad(2 * latent_channels , 2 * latent_channels , 1 ) _A = nn.Convad(_UpperCAmelCase , _UpperCAmelCase , 1 ) _A = False _A = False # only relevant if vae tiling is enabled _A = self.config.sample_size _A = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple) ) else self.config.sample_size ) _A = int(sample_size / (2 ** (len(self.config.block_out_channels ) - 1)) ) _A = 0.25 def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : int , _UpperCAmelCase : Tuple=False ): if isinstance(_UpperCAmelCase , (Encoder, Decoder) ): _A = value def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : bool = True ): _A = use_tiling def lowerCAmelCase_ ( self : Union[str, Any] ): self.enable_tiling(_UpperCAmelCase ) def lowerCAmelCase_ ( self : Dict ): _A = True def lowerCAmelCase_ ( self : str ): _A = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def lowerCAmelCase_ ( self : str ): _A = {} def fn_recursive_add_processors(_UpperCAmelCase : str , _UpperCAmelCase : torch.nn.Module , _UpperCAmelCase : Dict[str, AttentionProcessor] ): if hasattr(_UpperCAmelCase , 'set_processor' ): _A = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(F'''{name}.{sub_name}''' , _UpperCAmelCase , _UpperCAmelCase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return processors def lowerCAmelCase_ ( self : str , _UpperCAmelCase : Union[AttentionProcessor, Dict[str, AttentionProcessor]] ): _A = len(self.attn_processors.keys() ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) and len(_UpperCAmelCase ) != count: raise ValueError( F'''A dict of processors was passed, but the number of processors {len(_UpperCAmelCase )} does not match the''' F''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(_UpperCAmelCase : str , _UpperCAmelCase : torch.nn.Module , _UpperCAmelCase : int ): if hasattr(_UpperCAmelCase , 'set_processor' ): if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): module.set_processor(_UpperCAmelCase ) else: module.set_processor(processor.pop(F'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(F'''{name}.{sub_name}''' , _UpperCAmelCase , _UpperCAmelCase ) for name, module in self.named_children(): fn_recursive_attn_processor(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def lowerCAmelCase_ ( self : List[str] ): self.set_attn_processor(AttnProcessor() ) @apply_forward_hook def lowerCAmelCase_ ( self : int , _UpperCAmelCase : torch.FloatTensor , _UpperCAmelCase : bool = True ): if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(_UpperCAmelCase , return_dict=_UpperCAmelCase ) if self.use_slicing and x.shape[0] > 1: _A = [self.encoder(_UpperCAmelCase ) for x_slice in x.split(1 )] _A = torch.cat(_UpperCAmelCase ) else: _A = self.encoder(_UpperCAmelCase ) _A = self.quant_conv(_UpperCAmelCase ) _A = DiagonalGaussianDistribution(_UpperCAmelCase ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Any , _UpperCAmelCase : torch.FloatTensor , _UpperCAmelCase : bool = True ): if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(_UpperCAmelCase , return_dict=_UpperCAmelCase ) _A = self.post_quant_conv(_UpperCAmelCase ) _A = self.decoder(_UpperCAmelCase ) if not return_dict: return (dec,) return DecoderOutput(sample=_UpperCAmelCase ) @apply_forward_hook def lowerCAmelCase_ ( self : str , _UpperCAmelCase : torch.FloatTensor , _UpperCAmelCase : bool = True ): if self.use_slicing and z.shape[0] > 1: _A = [self._decode(_UpperCAmelCase ).sample for z_slice in z.split(1 )] _A = torch.cat(_UpperCAmelCase ) else: _A = self._decode(_UpperCAmelCase ).sample if not return_dict: return (decoded,) return DecoderOutput(sample=_UpperCAmelCase ) def lowerCAmelCase_ ( self : str , _UpperCAmelCase : int , _UpperCAmelCase : Dict , _UpperCAmelCase : Any ): _A = min(a.shape[2] , b.shape[2] , _UpperCAmelCase ) for y in range(_UpperCAmelCase ): _A = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : List[Any] ): _A = min(a.shape[3] , b.shape[3] , _UpperCAmelCase ) for x in range(_UpperCAmelCase ): _A = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def lowerCAmelCase_ ( self : str , _UpperCAmelCase : torch.FloatTensor , _UpperCAmelCase : bool = True ): _A = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor) ) _A = int(self.tile_latent_min_size * self.tile_overlap_factor ) _A = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. _A = [] for i in range(0 , x.shape[2] , _UpperCAmelCase ): _A = [] for j in range(0 , x.shape[3] , _UpperCAmelCase ): _A = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] _A = self.encoder(_UpperCAmelCase ) _A = self.quant_conv(_UpperCAmelCase ) row.append(_UpperCAmelCase ) rows.append(_UpperCAmelCase ) _A = [] for i, row in enumerate(_UpperCAmelCase ): _A = [] for j, tile in enumerate(_UpperCAmelCase ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: _A = self.blend_v(rows[i - 1][j] , _UpperCAmelCase , _UpperCAmelCase ) if j > 0: _A = self.blend_h(row[j - 1] , _UpperCAmelCase , _UpperCAmelCase ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(_UpperCAmelCase , dim=3 ) ) _A = torch.cat(_UpperCAmelCase , dim=2 ) _A = DiagonalGaussianDistribution(_UpperCAmelCase ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Tuple , _UpperCAmelCase : torch.FloatTensor , _UpperCAmelCase : bool = True ): _A = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor) ) _A = int(self.tile_sample_min_size * self.tile_overlap_factor ) _A = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. _A = [] for i in range(0 , z.shape[2] , _UpperCAmelCase ): _A = [] for j in range(0 , z.shape[3] , _UpperCAmelCase ): _A = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] _A = self.post_quant_conv(_UpperCAmelCase ) _A = self.decoder(_UpperCAmelCase ) row.append(_UpperCAmelCase ) rows.append(_UpperCAmelCase ) _A = [] for i, row in enumerate(_UpperCAmelCase ): _A = [] for j, tile in enumerate(_UpperCAmelCase ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: _A = self.blend_v(rows[i - 1][j] , _UpperCAmelCase , _UpperCAmelCase ) if j > 0: _A = self.blend_h(row[j - 1] , _UpperCAmelCase , _UpperCAmelCase ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(_UpperCAmelCase , dim=3 ) ) _A = torch.cat(_UpperCAmelCase , dim=2 ) if not return_dict: return (dec,) return DecoderOutput(sample=_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : torch.FloatTensor , _UpperCAmelCase : bool = False , _UpperCAmelCase : bool = True , _UpperCAmelCase : Optional[torch.Generator] = None , ): _A = sample _A = self.encode(_UpperCAmelCase ).latent_dist if sample_posterior: _A = posterior.sample(generator=_UpperCAmelCase ) else: _A = posterior.mode() _A = self.decode(_UpperCAmelCase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=_UpperCAmelCase )
271
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _snake_case = { 'configuration_xlm_roberta': [ 'XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaConfig', 'XLMRobertaOnnxConfig', ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['XLMRobertaTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['XLMRobertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMRobertaForCausalLM', 'XLMRobertaForMaskedLM', 'XLMRobertaForMultipleChoice', 'XLMRobertaForQuestionAnswering', 'XLMRobertaForSequenceClassification', 'XLMRobertaForTokenClassification', 'XLMRobertaModel', 'XLMRobertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMRobertaForCausalLM', 'TFXLMRobertaForMaskedLM', 'TFXLMRobertaForMultipleChoice', 'TFXLMRobertaForQuestionAnswering', 'TFXLMRobertaForSequenceClassification', 'TFXLMRobertaForTokenClassification', 'TFXLMRobertaModel', 'TFXLMRobertaPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'FlaxXLMRobertaForMaskedLM', 'FlaxXLMRobertaForCausalLM', 'FlaxXLMRobertaForMultipleChoice', 'FlaxXLMRobertaForQuestionAnswering', 'FlaxXLMRobertaForSequenceClassification', 'FlaxXLMRobertaForTokenClassification', 'FlaxXLMRobertaModel', 'FlaxXLMRobertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
250
'''simple docstring''' def _A ( snake_case , snake_case ) -> float: return price * (1 + tax_rate) if __name__ == "__main__": print(F'''{price_plus_tax(100, 0.2_5) = }''') print(F'''{price_plus_tax(1_2_5.5_0, 0.0_5) = }''')
250
1
# flake8: noqa # Lint as: python3 __UpperCAmelCase = [ 'VerificationMode', 'Version', 'disable_progress_bar', 'enable_progress_bar', 'is_progress_bar_enabled', 'experimental', ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
145
from __future__ import annotations def lowercase__ ( __snake_case : list[int] , __snake_case : int ): '''simple docstring''' if len(__snake_case ) == 0: return False UpperCAmelCase_ : Optional[int] = len(__snake_case ) // 2 if a_list[midpoint] == item: return True if item < a_list[midpoint]: return binary_search(a_list[:midpoint] , __snake_case ) else: return binary_search(a_list[midpoint + 1 :] , __snake_case ) if __name__ == "__main__": __UpperCAmelCase = input('Enter numbers separated by comma:\n').strip() __UpperCAmelCase = [int(item.strip()) for item in user_input.split(',')] __UpperCAmelCase = int(input('Enter the number to be found in the list:\n').strip()) __UpperCAmelCase = '' if binary_search(sequence, target) else 'not ' print(F'{target} was {not_str}found in {sequence}')
145
1
import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class lowerCamelCase (unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self ) -> Tuple: UpperCAmelCase_ : List[str] = logging.get_logger() # the current default level is logging.WARNING UpperCAmelCase_ : str = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Dict: UpperCAmelCase_ : Dict = logging.get_verbosity() UpperCAmelCase_ : List[str] = logging.get_logger('transformers.models.bart.tokenization_bart' ) UpperCAmelCase_ : Dict = 'Testing 1, 2, 3' # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(_UpperCamelCase ) as cl: logger.warning(_UpperCamelCase ) self.assertEqual(cl.out , msg + '\n' ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(_UpperCamelCase ) as cl: logger.warning(_UpperCamelCase ) self.assertEqual(cl.out , '' ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(_UpperCamelCase ) as cl: logger.warning(_UpperCamelCase ) self.assertEqual(cl.out , msg + '\n' ) # restore to the original level logging.set_verbosity(_UpperCamelCase ) @mockenv(TRANSFORMERS_VERBOSITY='error' ) def __UpperCAmelCase ( self ) -> Union[str, Any]: # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() # this action activates the env var UpperCAmelCase_ : List[str] = logging.get_logger('transformers.models.bart.tokenization_bart' ) UpperCAmelCase_ : Tuple = os.getenv('TRANSFORMERS_VERBOSITY' , _UpperCamelCase ) UpperCAmelCase_ : List[str] = logging.log_levels[env_level_str] UpperCAmelCase_ : Dict = logging.get_verbosity() self.assertEqual( _UpperCamelCase , _UpperCamelCase , f"TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}" , ) # restore to the original level UpperCAmelCase_ : Tuple = '' transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY='super-error' ) def __UpperCAmelCase ( self ) -> List[Any]: # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() UpperCAmelCase_ : Any = logging.logging.getLogger() with CaptureLogger(_UpperCamelCase ) as cl: # this action activates the env var logging.get_logger('transformers.models.bart.tokenization_bart' ) self.assertIn('Unknown option TRANSFORMERS_VERBOSITY=super-error' , cl.out ) # no need to restore as nothing was changed def __UpperCAmelCase ( self ) -> Union[str, Any]: # testing `logger.warning_advice()` transformers.utils.logging._reset_library_root_logger() UpperCAmelCase_ : Optional[Any] = logging.get_logger('transformers.models.bart.tokenization_bart' ) UpperCAmelCase_ : Dict = 'Testing 1, 2, 3' with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='1' ): # nothing should be logged as env var disables this method with CaptureLogger(_UpperCamelCase ) as cl: logger.warning_advice(_UpperCamelCase ) self.assertEqual(cl.out , '' ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='' ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(_UpperCamelCase ) as cl: logger.warning_advice(_UpperCamelCase ) self.assertEqual(cl.out , msg + '\n' ) def lowercase__ ( ): '''simple docstring''' disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
29
from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'snap-research/efficientformer-l1-300': ( 'https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json' ), } class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Optional[int] = '''efficientformer''' def __init__( self , _UpperCamelCase = [3, 2, 6, 4] , _UpperCamelCase = [4_8, 9_6, 2_2_4, 4_4_8] , _UpperCamelCase = [True, True, True, True] , _UpperCamelCase = 4_4_8 , _UpperCamelCase = 3_2 , _UpperCamelCase = 4 , _UpperCamelCase = 7 , _UpperCamelCase = 5 , _UpperCamelCase = 8 , _UpperCamelCase = 4 , _UpperCamelCase = 0.0 , _UpperCamelCase = 1_6 , _UpperCamelCase = 3 , _UpperCamelCase = 3 , _UpperCamelCase = 3 , _UpperCamelCase = 2 , _UpperCamelCase = 1 , _UpperCamelCase = 0.0 , _UpperCamelCase = 1 , _UpperCamelCase = True , _UpperCamelCase = True , _UpperCamelCase = 1E-5 , _UpperCamelCase = "gelu" , _UpperCamelCase = 0.02 , _UpperCamelCase = 1E-12 , _UpperCamelCase = 2_2_4 , _UpperCamelCase = 1E-05 , **_UpperCamelCase , ) -> None: super().__init__(**_UpperCamelCase ) UpperCAmelCase_ : int = hidden_act UpperCAmelCase_ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase_ : Tuple = hidden_sizes UpperCAmelCase_ : Union[str, Any] = num_hidden_layers UpperCAmelCase_ : List[str] = num_attention_heads UpperCAmelCase_ : List[Any] = initializer_range UpperCAmelCase_ : int = layer_norm_eps UpperCAmelCase_ : List[str] = patch_size UpperCAmelCase_ : Union[str, Any] = num_channels UpperCAmelCase_ : Optional[Any] = depths UpperCAmelCase_ : List[Any] = mlp_expansion_ratio UpperCAmelCase_ : List[str] = downsamples UpperCAmelCase_ : List[Any] = dim UpperCAmelCase_ : Tuple = key_dim UpperCAmelCase_ : Optional[int] = attention_ratio UpperCAmelCase_ : str = resolution UpperCAmelCase_ : Dict = pool_size UpperCAmelCase_ : Union[str, Any] = downsample_patch_size UpperCAmelCase_ : List[str] = downsample_stride UpperCAmelCase_ : List[str] = downsample_pad UpperCAmelCase_ : Any = drop_path_rate UpperCAmelCase_ : Dict = num_metaad_blocks UpperCAmelCase_ : Dict = distillation UpperCAmelCase_ : int = use_layer_scale UpperCAmelCase_ : Any = layer_scale_init_value UpperCAmelCase_ : Any = image_size UpperCAmelCase_ : Dict = batch_norm_eps
29
1
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): lowerCamelCase_ = True from torch.cuda.amp import autocast lowerCamelCase_ = logging.getLogger(__name__) def UpperCamelCase( lowercase_=None , lowercase_=None ) -> int: '''simple docstring''' return field(default_factory=lambda: default , metadata=lowercase_ ) @dataclass class __lowerCamelCase : lowerCamelCase_ : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) lowerCamelCase_ : Optional[str] = field( default=__snake_case , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) lowerCamelCase_ : Optional[bool] = field( default=__snake_case , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) lowerCamelCase_ : Optional[float] = field( default=0.1 , metadata={'help': 'The dropout ratio for the attention probabilities.'} ) lowerCamelCase_ : Optional[float] = field( default=0.1 , metadata={'help': 'The dropout ratio for activations inside the fully connected layer.'} ) lowerCamelCase_ : Optional[float] = field( default=0.1 , metadata={ 'help': 'The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.' } , ) lowerCamelCase_ : Optional[float] = field( default=0.1 , metadata={'help': 'The dropout probabilitiy for all 1D convolutional layers in feature extractor.'} , ) lowerCamelCase_ : Optional[float] = field( default=0.0_5 , metadata={ 'help': ( 'Propability of each feature vector along the time axis to be chosen as the start of the vector' 'span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature' 'vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.' ) } , ) lowerCamelCase_ : Optional[float] = field(default=0.0 , metadata={'help': 'The LayerDrop probability.'} ) @dataclass class __lowerCamelCase : lowerCamelCase_ : Optional[str] = field( default=__snake_case , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) lowerCamelCase_ : Optional[str] = field( default='train+validation' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) lowerCamelCase_ : bool = field( default=__snake_case , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) lowerCamelCase_ : Optional[int] = field( default=__snake_case , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) lowerCamelCase_ : Optional[int] = field( default=__snake_case , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) lowerCamelCase_ : Optional[int] = field( default=__snake_case , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of validation examples to this ' 'value if set.' ) } , ) lowerCamelCase_ : List[str] = list_field( default=[',', '?', '.', '!', '-', ';', ':', '""', '%', '\'', '"', '�'] , metadata={'help': 'A list of characters to remove from the transcripts.'} , ) @dataclass class __lowerCamelCase : lowerCamelCase_ : WavaVecaProcessor lowerCamelCase_ : Union[bool, str] = True lowerCamelCase_ : Optional[int] = None lowerCamelCase_ : Optional[int] = None lowerCamelCase_ : Optional[int] = None lowerCamelCase_ : Optional[int] = None def __call__( self , lowerCamelCase ) -> Dict[str, torch.Tensor]: # split inputs and labels since they have to be of different lenghts and need # different padding methods snake_case_ = [{"""input_values""": feature["""input_values"""]} for feature in features] snake_case_ = [{"""input_ids""": feature["""labels"""]} for feature in features] snake_case_ = self.processor.pad( lowerCamelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" , ) snake_case_ = self.processor.pad( labels=lowerCamelCase , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors="""pt""" , ) # replace padding with -100 to ignore loss correctly snake_case_ = labels_batch["""input_ids"""].masked_fill(labels_batch.attention_mask.ne(1 ) , -100 ) snake_case_ = labels return batch class __lowerCamelCase ( __snake_case ): def lowerCAmelCase_ ( self , lowerCamelCase , lowerCamelCase ) -> torch.Tensor: model.train() snake_case_ = self._prepare_inputs(lowerCamelCase ) if self.use_amp: with autocast(): snake_case_ = self.compute_loss(lowerCamelCase , lowerCamelCase ) else: snake_case_ = self.compute_loss(lowerCamelCase , lowerCamelCase ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": snake_case_ = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": snake_case_ = loss.sum() / (inputs["""labels"""] >= 0).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: snake_case_ = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowerCamelCase ).backward() elif self.use_apex: with amp.scale_loss(lowerCamelCase , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowerCamelCase ) else: loss.backward() return loss.detach() def UpperCamelCase( ) -> Tuple: '''simple docstring''' snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case_ , snake_case_ , snake_case_ = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case_ , snake_case_ , snake_case_ = parser.parse_args_into_dataclasses() # Detecting last checkpoint. snake_case_ = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case_ = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None: logger.info( f'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( f'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + f'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info("""Training/evaluation parameters %s""" , lowercase_ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: snake_case_ = datasets.load_dataset( """common_voice""" , data_args.dataset_config_name , split=data_args.train_split_name ) snake_case_ = datasets.load_dataset("""common_voice""" , data_args.dataset_config_name , split="""test""" ) # Create and save tokenizer snake_case_ = f'''[{"".join(data_args.chars_to_ignore )}]''' def remove_special_characters(lowercase_ ): snake_case_ = re.sub(lowercase_ , """""" , batch["""sentence"""] ).lower() + """ """ return batch snake_case_ = train_dataset.map(lowercase_ , remove_columns=["""sentence"""] ) snake_case_ = eval_dataset.map(lowercase_ , remove_columns=["""sentence"""] ) def extract_all_chars(lowercase_ ): snake_case_ = """ """.join(batch["""text"""] ) snake_case_ = list(set(lowercase_ ) ) return {"vocab": [vocab], "all_text": [all_text]} snake_case_ = train_dataset.map( lowercase_ , batched=lowercase_ , batch_size=-1 , keep_in_memory=lowercase_ , remove_columns=train_dataset.column_names , ) snake_case_ = train_dataset.map( lowercase_ , batched=lowercase_ , batch_size=-1 , keep_in_memory=lowercase_ , remove_columns=eval_dataset.column_names , ) snake_case_ = list(set(vocab_train["""vocab"""][0] ) | set(vocab_test["""vocab"""][0] ) ) snake_case_ = {v: k for k, v in enumerate(lowercase_ )} snake_case_ = vocab_dict[""" """] del vocab_dict[" "] snake_case_ = len(lowercase_ ) snake_case_ = len(lowercase_ ) with open("""vocab.json""" , """w""" ) as vocab_file: json.dump(lowercase_ , lowercase_ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = WavaVecaCTCTokenizer( """vocab.json""" , unk_token="""[UNK]""" , pad_token="""[PAD]""" , word_delimiter_token="""|""" , ) snake_case_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0.0 , do_normalize=lowercase_ , return_attention_mask=lowercase_ ) snake_case_ = WavaVecaProcessor(feature_extractor=lowercase_ , tokenizer=lowercase_ ) snake_case_ = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction="""mean""" , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: snake_case_ = min(len(lowercase_ ) , data_args.max_train_samples ) snake_case_ = train_dataset.select(range(lowercase_ ) ) if data_args.max_val_samples is not None: snake_case_ = eval_dataset.select(range(data_args.max_val_samples ) ) snake_case_ = torchaudio.transforms.Resample(48000 , 16000 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(lowercase_ ): snake_case_ , snake_case_ = torchaudio.load(batch["""path"""] ) snake_case_ = resampler(lowercase_ ).squeeze().numpy() snake_case_ = 16000 snake_case_ = batch["""text"""] return batch snake_case_ = train_dataset.map( lowercase_ , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) snake_case_ = eval_dataset.map( lowercase_ , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(lowercase_ ): # check that all files have the correct sampling rate assert ( len(set(batch["""sampling_rate"""] ) ) == 1 ), f'''Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.''' snake_case_ = processor( audio=batch["""speech"""] , text=batch["""target_text"""] , sampling_rate=batch["""sampling_rate"""][0] ) batch.update(lowercase_ ) return batch snake_case_ = train_dataset.map( lowercase_ , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=lowercase_ , num_proc=data_args.preprocessing_num_workers , ) snake_case_ = eval_dataset.map( lowercase_ , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=lowercase_ , num_proc=data_args.preprocessing_num_workers , ) # Metric snake_case_ = datasets.load_metric("""wer""" ) def compute_metrics(lowercase_ ): snake_case_ = pred.predictions snake_case_ = np.argmax(lowercase_ , axis=-1 ) snake_case_ = processor.tokenizer.pad_token_id snake_case_ = processor.batch_decode(lowercase_ ) # we do not want to group tokens when computing the metrics snake_case_ = processor.batch_decode(pred.label_ids , group_tokens=lowercase_ ) snake_case_ = wer_metric.compute(predictions=lowercase_ , references=lowercase_ ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator snake_case_ = DataCollatorCTCWithPadding(processor=lowercase_ , padding=lowercase_ ) # Initialize our Trainer snake_case_ = CTCTrainer( model=lowercase_ , data_collator=lowercase_ , args=lowercase_ , compute_metrics=lowercase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: snake_case_ = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): snake_case_ = model_args.model_name_or_path else: snake_case_ = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) snake_case_ = trainer.train(resume_from_checkpoint=lowercase_ ) trainer.save_model() snake_case_ = train_result.metrics snake_case_ = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowercase_ ) ) snake_case_ = min(lowercase_ , len(lowercase_ ) ) trainer.log_metrics("""train""" , lowercase_ ) trainer.save_metrics("""train""" , lowercase_ ) trainer.save_state() # Evaluation snake_case_ = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) snake_case_ = trainer.evaluate() snake_case_ = data_args.max_val_samples if data_args.max_val_samples is not None else len(lowercase_ ) snake_case_ = min(lowercase_ , len(lowercase_ ) ) trainer.log_metrics("""eval""" , lowercase_ ) trainer.save_metrics("""eval""" , lowercase_ ) return results if __name__ == "__main__": main()
365
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''google/mobilenet_v2_1.4_224''': '''https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json''', '''google/mobilenet_v2_1.0_224''': '''https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json''', '''google/mobilenet_v2_0.75_160''': '''https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json''', '''google/mobilenet_v2_0.35_96''': '''https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json''', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class __lowerCamelCase ( __snake_case ): lowerCamelCase_ : Optional[Any] = 'mobilenet_v2' def __init__( self , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=1.0 , lowerCamelCase=8 , lowerCamelCase=8 , lowerCamelCase=6 , lowerCamelCase=32 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="relu6" , lowerCamelCase=True , lowerCamelCase=0.8 , lowerCamelCase=0.02 , lowerCamelCase=0.001 , lowerCamelCase=255 , **lowerCamelCase , ) -> Union[str, Any]: super().__init__(**lowerCamelCase ) if depth_multiplier <= 0: raise ValueError("""depth_multiplier must be greater than zero.""" ) snake_case_ = num_channels snake_case_ = image_size snake_case_ = depth_multiplier snake_case_ = depth_divisible_by snake_case_ = min_depth snake_case_ = expand_ratio snake_case_ = output_stride snake_case_ = first_layer_is_expansion snake_case_ = finegrained_output snake_case_ = hidden_act snake_case_ = tf_padding snake_case_ = classifier_dropout_prob snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = semantic_loss_ignore_index class __lowerCamelCase ( __snake_case ): lowerCamelCase_ : Dict = version.parse('1.11' ) @property def lowerCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict([("""pixel_values""", {0: """batch"""})] ) @property def lowerCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "image-classification": return OrderedDict([("""logits""", {0: """batch"""})] ) else: return OrderedDict([("""last_hidden_state""", {0: """batch"""}), ("""pooler_output""", {0: """batch"""})] ) @property def lowerCAmelCase_ ( self ) -> float: return 1e-4
34
0
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser snake_case_ : Dict = re.compile(r"\s+") def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : Tuple ) -> Union[str, Any]: return {"hash": hashlib.mda(re.sub(SCREAMING_SNAKE_CASE__, '''''', example['''content'''] ).encode('''utf-8''' ) ).hexdigest()} def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase_ : List[str] = [len(SCREAMING_SNAKE_CASE__ ) for line in example['''content'''].splitlines()] return {"line_mean": np.mean(SCREAMING_SNAKE_CASE__ ), "line_max": max(SCREAMING_SNAKE_CASE__ )} def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : str ) -> str: UpperCAmelCase_ : Optional[Any] = np.mean([c.isalnum() for c in example['''content''']] ) return {"alpha_frac": alpha_frac} def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : List[Any], SCREAMING_SNAKE_CASE__ : Tuple ) -> List[str]: if example["hash"] in uniques: uniques.remove(example['''hash'''] ) return True else: return False def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : List[Any], SCREAMING_SNAKE_CASE__ : Tuple=5 ) -> str: UpperCAmelCase_ : Union[str, Any] = ['''auto-generated''', '''autogenerated''', '''automatically generated'''] UpperCAmelCase_ : Dict = example['''content'''].splitlines() for _, line in zip(range(SCREAMING_SNAKE_CASE__ ), SCREAMING_SNAKE_CASE__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : Optional[Any], SCREAMING_SNAKE_CASE__ : Union[str, Any]=5, SCREAMING_SNAKE_CASE__ : Optional[Any]=0.05 ) -> int: UpperCAmelCase_ : List[str] = ['''unit tests''', '''test file''', '''configuration file'''] UpperCAmelCase_ : str = example['''content'''].splitlines() UpperCAmelCase_ : List[str] = 0 UpperCAmelCase_ : List[str] = 0 # first test for _, line in zip(range(SCREAMING_SNAKE_CASE__ ), SCREAMING_SNAKE_CASE__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test UpperCAmelCase_ : Dict = example['''content'''].count('''\n''' ) UpperCAmelCase_ : str = int(coeff * nlines ) for line in lines: count_config += line.lower().count('''config''' ) count_test += line.lower().count('''test''' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> List[Any]: UpperCAmelCase_ : Tuple = ['''def ''', '''class ''', '''for ''', '''while '''] UpperCAmelCase_ : int = example['''content'''].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : Optional[Any], SCREAMING_SNAKE_CASE__ : List[str]=4 ) -> Dict: UpperCAmelCase_ : List[str] = example['''content'''].splitlines() UpperCAmelCase_ : str = 0 for line in lines: counter += line.lower().count('''=''' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : Optional[int] ) -> Union[str, Any]: UpperCAmelCase_ : int = tokenizer(example['''content'''], truncation=SCREAMING_SNAKE_CASE__ )['''input_ids'''] UpperCAmelCase_ : Optional[int] = len(example['''content'''] ) / len(SCREAMING_SNAKE_CASE__ ) return {"ratio": ratio} def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : List[Any] ) -> Dict: UpperCAmelCase_ : int = {} results.update(get_hash(SCREAMING_SNAKE_CASE__ ) ) results.update(line_stats(SCREAMING_SNAKE_CASE__ ) ) results.update(alpha_stats(SCREAMING_SNAKE_CASE__ ) ) results.update(char_token_ratio(SCREAMING_SNAKE_CASE__ ) ) results.update(is_autogenerated(SCREAMING_SNAKE_CASE__ ) ) results.update(is_config_or_test(SCREAMING_SNAKE_CASE__ ) ) results.update(has_no_keywords(SCREAMING_SNAKE_CASE__ ) ) results.update(has_few_assignments(SCREAMING_SNAKE_CASE__ ) ) return results def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : List[str], SCREAMING_SNAKE_CASE__ : Optional[int], SCREAMING_SNAKE_CASE__ : List[Any] ) -> Tuple: if not check_uniques(SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : List[str] ) -> Tuple: with open(SCREAMING_SNAKE_CASE__, '''rb''' ) as f_in: with gzip.open(str(SCREAMING_SNAKE_CASE__ ) + '''.gz''', '''wb''', compresslevel=6 ) as f_out: shutil.copyfileobj(SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) os.unlink(SCREAMING_SNAKE_CASE__ ) # Settings snake_case_ : Dict = HfArgumentParser(PreprocessingArguments) snake_case_ : Dict = parser.parse_args() if args.num_workers is None: snake_case_ : List[Any] = multiprocessing.cpu_count() snake_case_ : Any = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset snake_case_ : Tuple = time.time() snake_case_ : str = load_dataset(args.dataset_name, split="train") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing snake_case_ : Any = time.time() snake_case_ : Optional[int] = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes snake_case_ : List[Any] = set(ds.unique("hash")) snake_case_ : int = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics snake_case_ : Any = time.time() snake_case_ : Optional[Any] = ds.filter(filter, fn_kwargs={"uniques": uniques, "args": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: snake_case_ : List[str] = time.time() snake_case_ ,snake_case_ : str = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file snake_case_ : List[str] = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / "duplicate_clusters.json", "w") as f: json.dump(duplicate_clusters, f) snake_case_ : str = output_dir / "data" data_dir.mkdir(exist_ok=True) snake_case_ : int = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): snake_case_ : List[str] = str(data_dir / f'''file-{file_number+1:012}.json''') snake_case_ : Union[str, Any] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
125
'''simple docstring''' snake_case_ : List[str] = { "Pillow": "Pillow<10.0.0", "accelerate": "accelerate>=0.20.3", "av": "av==9.2.0", "beautifulsoup4": "beautifulsoup4", "black": "black~=23.1", "codecarbon": "codecarbon==1.2.0", "cookiecutter": "cookiecutter==1.7.3", "dataclasses": "dataclasses", "datasets": "datasets!=2.5.0", "decord": "decord==0.6.0", "deepspeed": "deepspeed>=0.9.3", "diffusers": "diffusers", "dill": "dill<0.3.5", "evaluate": "evaluate>=0.2.0", "fairscale": "fairscale>0.3", "faiss-cpu": "faiss-cpu", "fastapi": "fastapi", "filelock": "filelock", "flax": "flax>=0.4.1,<=0.7.0", "ftfy": "ftfy", "fugashi": "fugashi>=1.0", "GitPython": "GitPython<3.1.19", "hf-doc-builder": "hf-doc-builder>=0.3.0", "huggingface-hub": "huggingface-hub>=0.14.1,<1.0", "importlib_metadata": "importlib_metadata", "ipadic": "ipadic>=1.0.0,<2.0", "isort": "isort>=5.5.4", "jax": "jax>=0.2.8,!=0.3.2,<=0.4.13", "jaxlib": "jaxlib>=0.1.65,<=0.4.13", "jieba": "jieba", "kenlm": "kenlm", "keras-nlp": "keras-nlp>=0.3.1", "librosa": "librosa", "nltk": "nltk", "natten": "natten>=0.14.6", "numpy": "numpy>=1.17", "onnxconverter-common": "onnxconverter-common", "onnxruntime-tools": "onnxruntime-tools>=1.4.2", "onnxruntime": "onnxruntime>=1.4.0", "opencv-python": "opencv-python", "optuna": "optuna", "optax": "optax>=0.0.8,<=0.1.4", "packaging": "packaging>=20.0", "parameterized": "parameterized", "phonemizer": "phonemizer", "protobuf": "protobuf", "psutil": "psutil", "pyyaml": "pyyaml>=5.1", "pydantic": "pydantic<2", "pytest": "pytest>=7.2.0", "pytest-timeout": "pytest-timeout", "pytest-xdist": "pytest-xdist", "python": "python>=3.8.0", "ray[tune]": "ray[tune]", "regex": "regex!=2019.12.17", "requests": "requests", "rhoknp": "rhoknp>=1.1.0,<1.3.1", "rjieba": "rjieba", "rouge-score": "rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1", "ruff": "ruff>=0.0.241,<=0.0.259", "sacrebleu": "sacrebleu>=1.4.12,<2.0.0", "sacremoses": "sacremoses", "safetensors": "safetensors>=0.3.1", "sagemaker": "sagemaker>=2.31.0", "scikit-learn": "scikit-learn", "sentencepiece": "sentencepiece>=0.1.91,!=0.1.92", "sigopt": "sigopt", "starlette": "starlette", "sudachipy": "sudachipy>=0.6.6", "sudachidict_core": "sudachidict_core>=20220729", "tensorflow-cpu": "tensorflow-cpu>=2.6,<2.14", "tensorflow": "tensorflow>=2.6,<2.14", "tensorflow-text": "tensorflow-text<2.14", "tf2onnx": "tf2onnx", "timeout-decorator": "timeout-decorator", "timm": "timm", "tokenizers": "tokenizers>=0.11.1,!=0.11.3,<0.14", "torch": "torch>=1.9,!=1.12.0", "torchaudio": "torchaudio", "torchvision": "torchvision", "pyctcdecode": "pyctcdecode>=0.4.0", "tqdm": "tqdm>=4.27", "unidic": "unidic>=1.0.2", "unidic_lite": "unidic_lite>=1.0.7", "urllib3": "urllib3<2.0.0", "uvicorn": "uvicorn", }
125
1
import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def __lowercase ( a__ ) -> int: # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0x4_E00 and cp <= 0x9_FFF) or (cp >= 0x3_400 and cp <= 0x4_DBF) # or (cp >= 0x20_000 and cp <= 0x2A_6DF) # or (cp >= 0x2A_700 and cp <= 0x2B_73F) # or (cp >= 0x2B_740 and cp <= 0x2B_81F) # or (cp >= 0x2B_820 and cp <= 0x2C_EAF) # or (cp >= 0xF_900 and cp <= 0xF_AFF) or (cp >= 0x2F_800 and cp <= 0x2F_A1F) # ): # return True return False def __lowercase ( a__ ) -> List[str]: # word like '180' or '身高' or '神' for char in word: __SCREAMING_SNAKE_CASE = ord(a__ ) if not _is_chinese_char(a__ ): return 0 return 1 def __lowercase ( a__ ) -> Optional[int]: __SCREAMING_SNAKE_CASE = set() for token in tokens: __SCREAMING_SNAKE_CASE = len(a__ ) > 1 and is_chinese(a__ ) if chinese_word: word_set.add(a__ ) __SCREAMING_SNAKE_CASE = list(a__ ) return word_list def __lowercase ( a__ , a__ ) -> List[str]: if not chinese_word_set: return bert_tokens __SCREAMING_SNAKE_CASE = max([len(a__ ) for w in chinese_word_set] ) __SCREAMING_SNAKE_CASE = bert_tokens __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 0, len(a__ ) while start < end: __SCREAMING_SNAKE_CASE = True if is_chinese(bert_word[start] ): __SCREAMING_SNAKE_CASE = min(end - start , a__ ) for i in range(a__ , 1 , -1 ): __SCREAMING_SNAKE_CASE = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): __SCREAMING_SNAKE_CASE = '##' + bert_word[j] __SCREAMING_SNAKE_CASE = start + i __SCREAMING_SNAKE_CASE = False break if single_word: start += 1 return bert_word def __lowercase ( a__ , a__ , a__ ) -> List[str]: __SCREAMING_SNAKE_CASE = [] for i in range(0 , len(a__ ) , 1_00 ): __SCREAMING_SNAKE_CASE = ltp_tokenizer.pipeline(lines[i : i + 1_00] , tasks=['cws'] ).cws __SCREAMING_SNAKE_CASE = [get_chinese_word(a__ ) for r in res] ltp_res.extend(a__ ) assert len(a__ ) == len(a__ ) __SCREAMING_SNAKE_CASE = [] for i in range(0 , len(a__ ) , 1_00 ): __SCREAMING_SNAKE_CASE = bert_tokenizer(lines[i : i + 1_00] , add_special_tokens=a__ , truncation=a__ , max_length=5_12 ) bert_res.extend(res['input_ids'] ) assert len(a__ ) == len(a__ ) __SCREAMING_SNAKE_CASE = [] for input_ids, chinese_word in zip(a__ , a__ ): __SCREAMING_SNAKE_CASE = [] for id in input_ids: __SCREAMING_SNAKE_CASE = bert_tokenizer._convert_id_to_token(a__ ) input_tokens.append(a__ ) __SCREAMING_SNAKE_CASE = add_sub_symbol(a__ , a__ ) __SCREAMING_SNAKE_CASE = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(a__ ): if token[:2] == "##": __SCREAMING_SNAKE_CASE = token[2:] # save chinese tokens' pos if len(a__ ) == 1 and _is_chinese_char(ord(a__ ) ): ref_id.append(a__ ) ref_ids.append(a__ ) assert len(a__ ) == len(a__ ) return ref_ids def __lowercase ( a__ ) -> Dict: # For Chinese (Ro)Bert, the best result is from : RoBERTa-wwm-ext (https://github.com/ymcui/Chinese-BERT-wwm) # If we want to fine-tune these model, we have to use same tokenizer : LTP (https://github.com/HIT-SCIR/ltp) with open(args.file_name , 'r' , encoding='utf-8' ) as f: __SCREAMING_SNAKE_CASE = f.readlines() __SCREAMING_SNAKE_CASE = [line.strip() for line in data if len(a__ ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' __SCREAMING_SNAKE_CASE = LTP(args.ltp ) # faster in GPU device __SCREAMING_SNAKE_CASE = BertTokenizer.from_pretrained(args.bert ) __SCREAMING_SNAKE_CASE = prepare_ref(a__ , a__ , a__ ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: __SCREAMING_SNAKE_CASE = [json.dumps(a__ ) + '\n' for ref in ref_ids] f.writelines(a__ ) if __name__ == "__main__": lowerCAmelCase__ : str =argparse.ArgumentParser(description='''prepare_chinese_ref''') parser.add_argument( '''--file_name''', required=False, type=str, default='''./resources/chinese-demo.txt''', help='''file need process, same as training data in lm''', ) parser.add_argument( '''--ltp''', required=False, type=str, default='''./resources/ltp''', help='''resources for LTP tokenizer, usually a path''', ) parser.add_argument( '''--bert''', required=False, type=str, default='''./resources/robert''', help='''resources for Bert tokenizer''', ) parser.add_argument( '''--save_path''', required=False, type=str, default='''./resources/ref.txt''', help='''path to save res''', ) lowerCAmelCase__ : Any =parser.parse_args() main(args)
118
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase_ ( UpperCamelCase_ , unittest.TestCase ): '''simple docstring''' UpperCamelCase__ : Optional[int] = LayoutLMTokenizer UpperCamelCase__ : Any = LayoutLMTokenizerFast UpperCamelCase__ : Optional[int] = True UpperCamelCase__ : int = True def _A ( self ): '''simple docstring''' super().setUp() __SCREAMING_SNAKE_CASE = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def _A ( self , **_A ): '''simple docstring''' return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_A ) def _A ( self , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = 'UNwant\u00E9d,running' __SCREAMING_SNAKE_CASE = 'unwanted, running' return input_text, output_text def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.tokenizer_class(self.vocab_file ) __SCREAMING_SNAKE_CASE = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_A , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_A ) , [7, 4, 5, 10, 8, 9] ) def _A ( self ): '''simple docstring''' pass
118
1
from math import pow, sqrt def _snake_case ( *lowerCAmelCase : float ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = len(lowerCAmelCase ) > 0 and all(value > 0.0 for value in values ) return result def _snake_case ( lowerCAmelCase : float , lowerCAmelCase : float ): """simple docstring""" return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(lowerCAmelCase , lowerCAmelCase ) else ValueError("Input Error: Molar mass values must greater than 0." ) ) def _snake_case ( lowerCAmelCase : float , lowerCAmelCase : float , lowerCAmelCase : float ): """simple docstring""" return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def _snake_case ( lowerCAmelCase : float , lowerCAmelCase : float , lowerCAmelCase : float ): """simple docstring""" return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def _snake_case ( lowerCAmelCase : float , lowerCAmelCase : float , lowerCAmelCase : float ): """simple docstring""" return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def _snake_case ( lowerCAmelCase : float , lowerCAmelCase : float , lowerCAmelCase : float ): """simple docstring""" return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) )
18
from __future__ import annotations from math import pi, sqrt def _snake_case ( lowerCAmelCase : float , lowerCAmelCase : float ): """simple docstring""" if inductance <= 0: raise ValueError("Inductance cannot be 0 or negative" ) elif capacitance <= 0: raise ValueError("Capacitance cannot be 0 or negative" ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
18
1
def A(__a: Union[str, Any] ): def merge(__a: Tuple , __a: Optional[Any] ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(__a ) <= 1: return collection lowerCAmelCase_ = len(__a ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase__ = input('''Enter numbers separated by a comma:\n''').strip() lowerCamelCase__ = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
357
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def A(__a: Any , __a: Union[str, Any] , __a: List[str] ): lowerCAmelCase_ = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] lowerCAmelCase_ = { "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } lowerCAmelCase_ = F"{src_lang}-{tgt_lang}" lowerCAmelCase_ = F"\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR's WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = \"facebook/wmt19-{src_lang}-{tgt_lang}\"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = \"{texts[src_lang]}\"\ninput_ids = tokenizer.encode(input, return_tensors=\"pt\")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn't seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn't support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR's WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n" os.makedirs(__a , exist_ok=__a ) lowerCAmelCase_ = os.path.join(__a , "README.md" ) print(F"Generating {path}" ) with open(__a , "w" , encoding="utf-8" ) as f: f.write(__a ) # make sure we are under the root of the project lowerCamelCase__ = Path(__file__).resolve().parent.parent.parent lowerCamelCase__ = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = model_name.split('''-''') lowerCamelCase__ = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
22
0
"""simple docstring""" def lowercase (_lowerCAmelCase ): __lowerCAmelCase = [[0 for _ in range(_lowerCAmelCase )] for _ in range(m + 1 )] for i in range(m + 1 ): __lowerCAmelCase = 1 for n in range(m + 1 ): for k in range(1 , _lowerCAmelCase ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: SCREAMING_SNAKE_CASE_ = int(input('''Enter a number: ''').strip()) print(partition(n)) except ValueError: print('''Please enter a number.''') else: try: SCREAMING_SNAKE_CASE_ = int(sys.argv[1]) print(partition(n)) except ValueError: print('''Please pass a number.''')
301
"""simple docstring""" import math def lowercase (_lowerCAmelCase ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_lowerCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowercase (_lowerCAmelCase = 0.1 ): __lowerCAmelCase = 3 __lowerCAmelCase = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_lowerCAmelCase ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
301
1
from itertools import count def UpperCamelCase ( __lowercase : int = 50 ): '''simple docstring''' A_ : Optional[Any] = [1] * min_block_length for n in count(__lowercase ): fill_count_functions.append(1 ) for block_length in range(__lowercase ,n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 1_00_00_00: break return n if __name__ == "__main__": print(F"""{solution() = }""")
192
import math def UpperCamelCase ( __lowercase : int = 1_00 ): '''simple docstring''' A_ : List[Any] = sum(i * i for i in range(1 ,n + 1 ) ) A_ : int = int(math.pow(sum(range(1 ,n + 1 ) ) ,2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
192
1
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING lowercase__ : str = logging.get_logger(__name__) lowercase__ : Any = { 'SenseTime/deformable-detr': 'https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json', # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" _snake_case : Tuple = 'deformable_detr' _snake_case : Dict = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Optional[Any] , lowerCAmelCase__ : str=True , lowerCAmelCase__ : List[str]=None , lowerCAmelCase__ : Dict=3 , lowerCAmelCase__ : List[str]=300 , lowerCAmelCase__ : Union[str, Any]=1024 , lowerCAmelCase__ : Tuple=6 , lowerCAmelCase__ : Union[str, Any]=1024 , lowerCAmelCase__ : List[Any]=8 , lowerCAmelCase__ : List[Any]=6 , lowerCAmelCase__ : Tuple=1024 , lowerCAmelCase__ : List[Any]=8 , lowerCAmelCase__ : Union[str, Any]=0.0 , lowerCAmelCase__ : Tuple=True , lowerCAmelCase__ : Any="relu" , lowerCAmelCase__ : int=256 , lowerCAmelCase__ : Dict=0.1 , lowerCAmelCase__ : Tuple=0.0 , lowerCAmelCase__ : str=0.0 , lowerCAmelCase__ : int=0.02 , lowerCAmelCase__ : Any=1.0 , lowerCAmelCase__ : Optional[Any]=True , lowerCAmelCase__ : int=False , lowerCAmelCase__ : str="sine" , lowerCAmelCase__ : List[Any]="resnet50" , lowerCAmelCase__ : str=True , lowerCAmelCase__ : str=False , lowerCAmelCase__ : List[str]=4 , lowerCAmelCase__ : List[str]=4 , lowerCAmelCase__ : Optional[Any]=4 , lowerCAmelCase__ : Optional[Any]=False , lowerCAmelCase__ : Optional[int]=300 , lowerCAmelCase__ : int=False , lowerCAmelCase__ : Optional[Any]=1 , lowerCAmelCase__ : Dict=5 , lowerCAmelCase__ : int=2 , lowerCAmelCase__ : Tuple=1 , lowerCAmelCase__ : Optional[Any]=1 , lowerCAmelCase__ : Optional[int]=5 , lowerCAmelCase__ : Dict=2 , lowerCAmelCase__ : int=0.1 , lowerCAmelCase__ : int=0.25 , lowerCAmelCase__ : Any=False , **lowerCAmelCase__ : Optional[Any] , ) -> str: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) _UpperCamelCase = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): _UpperCamelCase = backbone_config.get('''model_type''' ) _UpperCamelCase = CONFIG_MAPPING[backbone_model_type] _UpperCamelCase = config_class.from_dict(lowerCAmelCase__ ) _UpperCamelCase = use_timm_backbone _UpperCamelCase = backbone_config _UpperCamelCase = num_channels _UpperCamelCase = num_queries _UpperCamelCase = max_position_embeddings _UpperCamelCase = d_model _UpperCamelCase = encoder_ffn_dim _UpperCamelCase = encoder_layers _UpperCamelCase = encoder_attention_heads _UpperCamelCase = decoder_ffn_dim _UpperCamelCase = decoder_layers _UpperCamelCase = decoder_attention_heads _UpperCamelCase = dropout _UpperCamelCase = attention_dropout _UpperCamelCase = activation_dropout _UpperCamelCase = activation_function _UpperCamelCase = init_std _UpperCamelCase = init_xavier_std _UpperCamelCase = encoder_layerdrop _UpperCamelCase = auxiliary_loss _UpperCamelCase = position_embedding_type _UpperCamelCase = backbone _UpperCamelCase = use_pretrained_backbone _UpperCamelCase = dilation # deformable attributes _UpperCamelCase = num_feature_levels _UpperCamelCase = encoder_n_points _UpperCamelCase = decoder_n_points _UpperCamelCase = two_stage _UpperCamelCase = two_stage_num_proposals _UpperCamelCase = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''' ) # Hungarian matcher _UpperCamelCase = class_cost _UpperCamelCase = bbox_cost _UpperCamelCase = giou_cost # Loss coefficients _UpperCamelCase = mask_loss_coefficient _UpperCamelCase = dice_loss_coefficient _UpperCamelCase = bbox_loss_coefficient _UpperCamelCase = giou_loss_coefficient _UpperCamelCase = eos_coefficient _UpperCamelCase = focal_alpha _UpperCamelCase = disable_custom_kernels super().__init__(is_encoder_decoder=lowerCAmelCase__ , **lowerCAmelCase__ ) @property def snake_case__ ( self : List[str] ) -> int: '''simple docstring''' return self.encoder_attention_heads @property def snake_case__ ( self : int ) -> int: '''simple docstring''' return self.d_model def snake_case__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: _UpperCamelCase = self.backbone_config.to_dict() _UpperCamelCase = self.__class__.model_type return output
324
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass class A_ : '''simple docstring''' __snake_case = None __snake_case = None __snake_case = None # sigma(t_i) @classmethod def _snake_case ( cls: Tuple ): return cls() @dataclass class A_ ( __UpperCamelCase ): '''simple docstring''' __snake_case = 42 __snake_case = 42 __snake_case = 42 class A_ ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' @property def _snake_case ( self: Tuple ): return True @register_to_config def __init__( self: Union[str, Any] , a: float = 0.0_2 , a: float = 100 , a: float = 1.0_0_7 , a: float = 80 , a: float = 0.0_5 , a: float = 50 , ): pass def _snake_case ( self: Tuple ): return KarrasVeSchedulerState.create() def _snake_case ( self: int , a: KarrasVeSchedulerState , a: int , a: Tuple = () ): __lowerCamelCase : str = jnp.arange(0 , a )[::-1].copy() __lowerCamelCase : Optional[Any] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in timesteps ] return state.replace( num_inference_steps=a , schedule=jnp.array(a , dtype=jnp.floataa ) , timesteps=a , ) def _snake_case ( self: List[str] , a: KarrasVeSchedulerState , a: jnp.ndarray , a: float , a: random.KeyArray , ): if self.config.s_min <= sigma <= self.config.s_max: __lowerCamelCase : Union[str, Any] = min(self.config.s_churn / state.num_inference_steps , 2**0.5 - 1 ) else: __lowerCamelCase : Optional[Any] = 0 # sample eps ~ N(0, S_noise^2 * I) __lowerCamelCase : List[Any] = random.split(a , num=1 ) __lowerCamelCase : List[str] = self.config.s_noise * random.normal(key=a , shape=sample.shape ) __lowerCamelCase : List[Any] = sigma + gamma * sigma __lowerCamelCase : Dict = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def _snake_case ( self: Optional[int] , a: KarrasVeSchedulerState , a: jnp.ndarray , a: float , a: float , a: jnp.ndarray , a: bool = True , ): __lowerCamelCase : str = sample_hat + sigma_hat * model_output __lowerCamelCase : Tuple = (sample_hat - pred_original_sample) / sigma_hat __lowerCamelCase : Dict = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=a , derivative=a , state=a ) def _snake_case ( self: List[Any] , a: KarrasVeSchedulerState , a: jnp.ndarray , a: float , a: float , a: jnp.ndarray , a: jnp.ndarray , a: jnp.ndarray , a: bool = True , ): __lowerCamelCase : Dict = sample_prev + sigma_prev * model_output __lowerCamelCase : int = (sample_prev - pred_original_sample) / sigma_prev __lowerCamelCase : Tuple = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=a , derivative=a , state=a ) def _snake_case ( self: Union[str, Any] , a: KarrasVeSchedulerState , a: int , a: Tuple , a: str ): raise NotImplementedError()
194
import warnings from .generation import TFGenerationMixin class A_ ( __UpperCamelCase ): '''simple docstring''' warnings.warn( """Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will """ """be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.""" , __UpperCamelCase , )
194
1
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Optional[int] = '| <pad> <unk> <s> </s> a b c d e f g h i j k'.split() a__: str = dict(zip(__A , range(len(__A)))) a__: int = { 'unk_token': '<unk>', 'bos_token': '<s>', 'eos_token': '</s>', } a__: Optional[Any] = { 'feature_size': 1, 'padding_value': 0.0, 'sampling_rate': 1_60_00, 'return_attention_mask': False, 'do_normalize': True, } a__: Dict = tempfile.mkdtemp() a__: Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file']) a__: Tuple = os.path.join(self.tmpdirname , __A) with open(self.vocab_file , 'w' , encoding='utf-8') as fp: fp.write(json.dumps(__A) + '\n') with open(self.feature_extraction_file , 'w' , encoding='utf-8') as fp: fp.write(json.dumps(__A) + '\n') # load decoder from hub a__: Any = 'hf-internal-testing/ngram-beam-search-decoder' def lowerCamelCase_ ( self , **lowercase) -> Dict: '''simple docstring''' a__: Tuple = self.add_kwargs_tokens_map.copy() kwargs.update(__A) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **__A) def lowerCamelCase_ ( self , **lowercase) -> Optional[Any]: '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **__A) def lowerCamelCase_ ( self , **lowercase) -> Dict: '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **__A) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Dict = self.get_tokenizer() a__: List[str] = self.get_feature_extractor() a__: List[Any] = self.get_decoder() a__: Optional[Any] = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A) processor.save_pretrained(self.tmpdirname) a__: str = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab()) self.assertIsInstance(processor.tokenizer , __A) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string()) self.assertIsInstance(processor.feature_extractor , __A) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , __A) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: str = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder()) processor.save_pretrained(self.tmpdirname) # make sure that error is thrown when decoder alphabet doesn't match a__: int = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3) # decoder self.assertEqual(processor.language_model.alpha , 5.0) self.assertEqual(processor.language_model.beta , 3.0) self.assertEqual(processor.language_model.score_boundary , -7.0) self.assertEqual(processor.language_model.unk_score_offset , 3) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: str = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['xx']) with self.assertRaisesRegex(__A , 'include'): WavaVecaProcessorWithLM( tokenizer=__A , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder()) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Any = self.get_feature_extractor() a__: int = self.get_tokenizer() a__: List[Any] = self.get_decoder() a__: Any = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A) a__: Any = floats_list((3, 10_00)) a__: Dict = feature_extractor(__A , return_tensors='np') a__: Tuple = processor(__A , return_tensors='np') for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Dict = self.get_feature_extractor() a__: List[str] = self.get_tokenizer() a__: int = self.get_decoder() a__: List[Any] = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A) a__: Optional[Any] = 'This is a test string' a__: Optional[Any] = processor(text=__A) a__: Tuple = tokenizer(__A) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def lowerCamelCase_ ( self , lowercase=(2, 10, 16) , lowercase=77) -> Optional[int]: '''simple docstring''' np.random.seed(__A) return np.random.rand(*__A) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Dict = self.get_feature_extractor() a__: Union[str, Any] = self.get_tokenizer() a__: List[str] = self.get_decoder() a__: str = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A) a__: Dict = self._get_dummy_logits(shape=(10, 16) , seed=13) a__: Optional[int] = processor.decode(__A) a__: List[Any] = decoder.decode_beams(__A)[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text) self.assertEqual('</s> <s> </s>' , decoded_processor.text) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score) @parameterized.expand([[None], ['fork'], ['spawn']]) def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' a__: Optional[Any] = self.get_feature_extractor() a__: Tuple = self.get_tokenizer() a__: Optional[Any] = self.get_decoder() a__: int = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A) a__: List[str] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: a__: List[Any] = processor.batch_decode(__A) else: with get_context(__A).Pool() as pool: a__: int = processor.batch_decode(__A , __A) a__: Optional[Any] = list(__A) with get_context('fork').Pool() as p: a__: Any = decoder.decode_beams_batch(__A , __A) a__ , a__ , a__: Optional[int] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0]) logit_scores_decoder.append(beams[0][-2]) lm_scores_decoder.append(beams[0][-1]) self.assertListEqual(__A , decoded_processor.text) self.assertListEqual(['<s> <s> </s>', '<s> <s> <s>'] , decoded_processor.text) self.assertListEqual(__A , decoded_processor.logit_score) self.assertListEqual(__A , decoded_processor.lm_score) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Any = self.get_feature_extractor() a__: str = self.get_tokenizer() a__: List[Any] = self.get_decoder() a__: Any = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A) a__: List[str] = self._get_dummy_logits() a__: Tuple = 15 a__: List[str] = -20.0 a__: Dict = -4.0 a__: Optional[Any] = processor.batch_decode( __A , beam_width=__A , beam_prune_logp=__A , token_min_logp=__A , ) a__: Tuple = decoded_processor_out.text a__: Tuple = list(__A) with get_context('fork').Pool() as pool: a__: Optional[int] = decoder.decode_beams_batch( __A , __A , beam_width=__A , beam_prune_logp=__A , token_min_logp=__A , ) a__: Tuple = [d[0][0] for d in decoded_decoder_out] a__: int = [d[0][2] for d in decoded_decoder_out] a__: Union[str, Any] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(__A , __A) self.assertListEqual(['</s> <s> <s>', '<s> <s> <s>'] , __A) self.assertTrue(np.array_equal(__A , decoded_processor_out.logit_score)) self.assertTrue(np.allclose([-20.054, -18.447] , __A , atol=1e-3)) self.assertTrue(np.array_equal(__A , decoded_processor_out.lm_score)) self.assertTrue(np.allclose([-15.554, -13.9474] , __A , atol=1e-3)) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Union[str, Any] = self.get_feature_extractor() a__: Union[str, Any] = self.get_tokenizer() a__: str = self.get_decoder() a__: Tuple = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A) a__: Tuple = self._get_dummy_logits() a__: List[Any] = 2.0 a__: str = 5.0 a__: List[Any] = -20.0 a__: Optional[int] = True a__: List[str] = processor.batch_decode( __A , alpha=__A , beta=__A , unk_score_offset=__A , lm_score_boundary=__A , ) a__: Optional[Any] = decoded_processor_out.text a__: Any = list(__A) decoder.reset_params( alpha=__A , beta=__A , unk_score_offset=__A , lm_score_boundary=__A , ) with get_context('fork').Pool() as pool: a__: List[Any] = decoder.decode_beams_batch( __A , __A , ) a__: Union[str, Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(__A , __A) self.assertListEqual(['<s> </s> <s> </s> </s>', '</s> </s> <s> </s> </s>'] , __A) a__: str = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0) self.assertEqual(lm_model.beta , 5.0) self.assertEqual(lm_model.unk_score_offset , -20.0) self.assertEqual(lm_model.score_boundary , __A) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained('hf-internal-testing/processor_with_lm') a__: List[Any] = processor.decoder.model_container[processor.decoder._model_key] a__: Optional[int] = Path(language_model._kenlm_model.path.decode('utf-8')).parent.parent.absolute() a__: List[str] = os.listdir(__A) a__: List[Any] = ['alphabet.json', 'language_model'] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(__A , __A) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = snapshot_download('hf-internal-testing/processor_with_lm') a__: int = WavaVecaProcessorWithLM.from_pretrained(__A) a__: Optional[Any] = processor.decoder.model_container[processor.decoder._model_key] a__: Optional[int] = Path(language_model._kenlm_model.path.decode('utf-8')).parent.parent.absolute() a__: str = os.listdir(__A) a__: Dict = os.listdir(__A) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(__A , __A) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: str = WavaVecaProcessorWithLM.from_pretrained('hf-internal-testing/processor_with_lm') a__: Optional[int] = AutoProcessor.from_pretrained('hf-internal-testing/processor_with_lm') a__: int = floats_list((3, 10_00)) a__: Union[str, Any] = processor_wavaveca(__A , return_tensors='np') a__: List[str] = processor_auto(__A , return_tensors='np') for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1e-2) a__: Any = self._get_dummy_logits() a__: str = processor_wavaveca.batch_decode(__A) a__: int = processor_auto.batch_decode(__A) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Dict = self.get_feature_extractor() a__: Optional[int] = self.get_tokenizer() a__: Dict = self.get_decoder() a__: Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='`processor` and `feature_extractor` model input names do not match' , ) @staticmethod def lowerCamelCase_ ( lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Union[str, Any] = [d[key] for d in offsets] return retrieved_list def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: List[Any] = WavaVecaProcessorWithLM.from_pretrained('hf-internal-testing/processor_with_lm') a__: Optional[int] = self._get_dummy_logits()[0] a__: int = processor.decode(__A , output_word_offsets=__A) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys()) , 4) self.assertTrue('text' in outputs) self.assertTrue('word_offsets' in outputs) self.assertTrue(isinstance(__A , __A)) self.assertEqual(' '.join(self.get_from_offsets(outputs['word_offsets'] , 'word')) , outputs.text) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'] , 'word') , ['<s>', '<s>', '</s>']) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'] , 'start_offset') , [0, 2, 4]) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'] , 'end_offset') , [1, 3, 5]) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained('hf-internal-testing/processor_with_lm') a__: Any = self._get_dummy_logits() a__: List[str] = processor.batch_decode(__A , output_word_offsets=__A) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys()) , 4) self.assertTrue('text' in outputs) self.assertTrue('word_offsets' in outputs) self.assertTrue(isinstance(__A , __A)) self.assertListEqual( [' '.join(self.get_from_offsets(__A , 'word')) for o in outputs['word_offsets']] , outputs.text) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'][0] , 'word') , ['<s>', '<s>', '</s>']) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'][0] , 'start_offset') , [0, 2, 4]) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'][0] , 'end_offset') , [1, 3, 5]) @slow @require_torch @require_torchaudio def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' import torch a__: str = load_dataset('common_voice' , 'en' , split='train' , streaming=__A) a__: Union[str, Any] = ds.cast_column('audio' , datasets.Audio(sampling_rate=1_60_00)) a__: Any = iter(__A) a__: Any = next(__A) a__: Any = AutoProcessor.from_pretrained('patrickvonplaten/wav2vec2-base-100h-with-lm') a__: Any = WavaVecaForCTC.from_pretrained('patrickvonplaten/wav2vec2-base-100h-with-lm') # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train a__: List[str] = processor(sample['audio']['array'] , return_tensors='pt').input_values with torch.no_grad(): a__: List[str] = model(__A).logits.cpu().numpy() a__: Tuple = processor.decode(logits[0] , output_word_offsets=__A) a__: int = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate a__: Tuple = [ { 'start_time': d['start_offset'] * time_offset, 'end_time': d['end_offset'] * time_offset, 'word': d['word'], } for d in output['word_offsets'] ] a__: int = 'WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL' # output words self.assertEqual(' '.join(self.get_from_offsets(__A , 'word')) , __A) self.assertEqual(' '.join(self.get_from_offsets(__A , 'word')) , output.text) # output times a__: List[Any] = torch.tensor(self.get_from_offsets(__A , 'start_time')) a__: List[str] = torch.tensor(self.get_from_offsets(__A , 'end_time')) # fmt: off a__: int = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599]) a__: Dict = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94]) # fmt: on self.assertTrue(torch.allclose(__A , __A , atol=0.01)) self.assertTrue(torch.allclose(__A , __A , atol=0.01))
290
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCamelCase_ : Any = logging.get_logger(__name__) lowerCamelCase_ : Optional[Any] = """▁""" lowerCamelCase_ : Union[str, Any] = {"""vocab_file""": """sentencepiece.bpe.model"""} lowerCamelCase_ : Any = { """vocab_file""": { """xlm-roberta-base""": """https://huggingface.co/xlm-roberta-base/resolve/main/sentencepiece.bpe.model""", """xlm-roberta-large""": """https://huggingface.co/xlm-roberta-large/resolve/main/sentencepiece.bpe.model""", """xlm-roberta-large-finetuned-conll02-dutch""": ( """https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/sentencepiece.bpe.model""" ), """xlm-roberta-large-finetuned-conll02-spanish""": ( """https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/sentencepiece.bpe.model""" ), """xlm-roberta-large-finetuned-conll03-english""": ( """https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/sentencepiece.bpe.model""" ), """xlm-roberta-large-finetuned-conll03-german""": ( """https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/sentencepiece.bpe.model""" ), } } lowerCamelCase_ : Tuple = { """xlm-roberta-base""": 5_1_2, """xlm-roberta-large""": 5_1_2, """xlm-roberta-large-finetuned-conll02-dutch""": 5_1_2, """xlm-roberta-large-finetuned-conll02-spanish""": 5_1_2, """xlm-roberta-large-finetuned-conll03-english""": 5_1_2, """xlm-roberta-large-finetuned-conll03-german""": 5_1_2, } class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = ["input_ids", "attention_mask"] def __init__( self , __A , __A="<s>" , __A="</s>" , __A="</s>" , __A="<s>" , __A="<unk>" , __A="<pad>" , __A="<mask>" , __A = None , **__A , ) -> None: # Mask token behave like a normal word, i.e. include the space before it a =AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else mask_token a ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__A , eos_token=__A , unk_token=__A , sep_token=__A , cls_token=__A , pad_token=__A , mask_token=__A , sp_model_kwargs=self.sp_model_kwargs , **__A , ) a =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__A ) ) a =vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token a ={'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab a =1 a =len(self.sp_model ) + self.fairseq_offset a ={v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ) -> Any: a =self.__dict__.copy() a =None a =self.sp_model.serialized_model_proto() return state def __setstate__( self , __A ) -> List[Any]: a =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): a ={} a =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def SCREAMING_SNAKE_CASE ( self , __A , __A = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] a =[self.cls_token_id] a =[self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , __A , __A = None , __A = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__A , token_ids_a=__A , already_has_special_tokens=__A ) if token_ids_a is None: return [1] + ([0] * len(__A )) + [1] return [1] + ([0] * len(__A )) + [1, 1] + ([0] * len(__A )) + [1] def SCREAMING_SNAKE_CASE ( self , __A , __A = None ) -> List[int]: a =[self.sep_token_id] a =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE ( self ) -> List[str]: return len(self.sp_model ) + self.fairseq_offset + 1 # Add the <mask> token def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a ={self.convert_ids_to_tokens(__A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def SCREAMING_SNAKE_CASE ( self , __A ) -> List[str]: return self.sp_model.encode(__A , out_type=__A ) def SCREAMING_SNAKE_CASE ( self , __A ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] a =self.sp_model.PieceToId(__A ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def SCREAMING_SNAKE_CASE ( self , __A ) -> List[str]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def SCREAMING_SNAKE_CASE ( self , __A ) -> Optional[Any]: a =''''''.join(__A ).replace(__A , ''' ''' ).strip() return out_string def SCREAMING_SNAKE_CASE ( self , __A , __A = None ) -> Tuple[str]: if not os.path.isdir(__A ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return a =os.path.join( __A , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __A ) elif not os.path.isfile(self.vocab_file ): with open(__A , '''wb''' ) as fi: a =self.sp_model.serialized_model_proto() fi.write(__A ) return (out_vocab_file,)
81
0
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _UpperCAmelCase : Union[str, Any] = "\\n\n" _UpperCAmelCase : List[str] = "\nPerplexity (PPL) is one of the most common metrics for evaluating language models.\nIt is defined as the exponentiated average negative log-likelihood of a sequence.\n\nFor more information, see https://huggingface.co/docs/transformers/perplexity\n" _UpperCAmelCase : List[str] = "\nArgs:\n model_id (str): model used for calculating Perplexity\n NOTE: Perplexity can only be calculated for causal language models.\n This includes models such as gpt2, causal variations of bert,\n causal versions of t5, and more (the full list can be found\n in the AutoModelForCausalLM documentation here:\n https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM )\n\n input_texts (list of str): input text, each separate text snippet\n is one list entry.\n batch_size (int): the batch size to run texts through the model. Defaults to 16.\n add_start_token (bool): whether to add the start token to the texts,\n so the perplexity can include the probability of the first word. Defaults to True.\n device (str): device to run on, defaults to 'cuda' when available\nReturns:\n perplexity: dictionary containing the perplexity scores for the texts\n in the input list, as well as the mean perplexity. If one of the input texts is\n longer than the max input length of the model, then it is truncated to the\n max length for the perplexity computation.\nExamples:\n Example 1:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"]\n >>> results = perplexity.compute(model_id='gpt2',\n ... add_start_token=False,\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 78.22\n >>> print(round(results[\"perplexities\"][0], 2))\n 11.11\n\n Example 2:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = datasets.load_dataset(\"wikitext\",\n ... \"wikitext-2-raw-v1\",\n ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS\n [...]\n >>> input_texts = [s for s in input_texts if s!='']\n >>> results = perplexity.compute(model_id='gpt2',\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 60.35\n >>> print(round(results[\"perplexities\"][0], 2))\n 81.12\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase ( datasets.Metric ): def __UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'input_texts': datasets.Value('string' ), } ) , reference_urls=['https://huggingface.co/docs/transformers/perplexity'] , ) def __UpperCamelCase ( self , A_ , A_ , A_ = 16 , A_ = True , A_=None ) -> str: """simple docstring""" if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": UpperCamelCase = 'cuda' else: UpperCamelCase = 'cuda' if torch.cuda.is_available() else 'cpu' UpperCamelCase = AutoModelForCausalLM.from_pretrained(A_ ) UpperCamelCase = model.to(A_ ) UpperCamelCase = AutoTokenizer.from_pretrained(A_ ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: UpperCamelCase = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(A_ ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'pad_token': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" UpperCamelCase = model.config.max_length - 1 else: UpperCamelCase = model.config.max_length UpperCamelCase = tokenizer( A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , return_tensors='pt' , return_attention_mask=A_ , ).to(A_ ) UpperCamelCase = encodings['input_ids'] UpperCamelCase = encodings['attention_mask'] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." UpperCamelCase = [] UpperCamelCase = CrossEntropyLoss(reduction='none' ) for start_index in logging.tqdm(range(0 , len(A_ ) , A_ ) ): UpperCamelCase = min(start_index + batch_size , len(A_ ) ) UpperCamelCase = encoded_texts[start_index:end_index] UpperCamelCase = attn_masks[start_index:end_index] if add_start_token: UpperCamelCase = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(A_ ) UpperCamelCase = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) UpperCamelCase = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(A_ ), attn_mask] , dim=1 ) UpperCamelCase = encoded_batch with torch.no_grad(): UpperCamelCase = model(A_ , attention_mask=A_ ).logits UpperCamelCase = out_logits[..., :-1, :].contiguous() UpperCamelCase = labels[..., 1:].contiguous() UpperCamelCase = attn_mask[..., 1:].contiguous() UpperCamelCase = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , A_ ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(A_ )}
365
from __future__ import annotations class lowercase : def __init__( self , A_ , A_ ) -> Any: """simple docstring""" UpperCamelCase , UpperCamelCase = text, pattern UpperCamelCase , UpperCamelCase = len(A_ ), len(A_ ) def __UpperCamelCase ( self , A_ ) -> int: """simple docstring""" for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def __UpperCamelCase ( self , A_ ) -> int: """simple docstring""" for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def __UpperCamelCase ( self ) -> list[int]: """simple docstring""" # searches pattern in text and returns index positions UpperCamelCase = [] for i in range(self.textLen - self.patLen + 1 ): UpperCamelCase = self.mismatch_in_text(A_ ) if mismatch_index == -1: positions.append(A_ ) else: UpperCamelCase = self.match_in_pattern(self.text[mismatch_index] ) UpperCamelCase = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _UpperCAmelCase : Union[str, Any] = "ABAABA" _UpperCAmelCase : Any = "AB" _UpperCAmelCase : Dict = BoyerMooreSearch(text, pattern) _UpperCAmelCase : Optional[int] = bms.bad_character_heuristic() if len(positions) == 0: print("No match found") else: print("Pattern found in following positions: ") print(positions)
110
0
import os import shutil import sys import tempfile import unittest from pathlib import Path import pytest import transformers from transformers import ( BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoTokenizer, BertConfig, BertTokenizer, BertTokenizerFast, CTRLTokenizer, GPTaTokenizer, GPTaTokenizerFast, PreTrainedTokenizerFast, RobertaTokenizer, RobertaTokenizerFast, is_tokenizers_available, ) from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.auto.tokenization_auto import ( TOKENIZER_MAPPING, get_tokenizer_config, tokenizer_class_from_name, ) from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import ( DUMMY_DIFF_TOKENIZER_IDENTIFIER, DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tokenizers, slow, ) sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def _a ( self ) -> List[Any]: __UpperCamelCase =0 @slow def _a ( self ) -> Optional[Any]: for model_name in (x for x in BERT_PRETRAINED_CONFIG_ARCHIVE_MAP.keys() if "japanese" not in x): __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) self.assertIsNotNone(__A ) self.assertIsInstance(__A , (BertTokenizer, BertTokenizerFast) ) self.assertGreater(len(__A ) , 0 ) for model_name in GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP.keys(): __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) self.assertIsNotNone(__A ) self.assertIsInstance(__A , (GPTaTokenizer, GPTaTokenizerFast) ) self.assertGreater(len(__A ) , 0 ) def _a ( self ) -> Any: __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) self.assertIsInstance(__A , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def _a ( self ) -> Any: __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) self.assertIsInstance(__A , (RobertaTokenizer, RobertaTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 20 ) def _a ( self ) -> str: __UpperCamelCase =AutoConfig.from_pretrained(__A ) self.assertIsInstance(__A , __A ) # Check that tokenizer_type ≠ model_type __UpperCamelCase =AutoTokenizer.from_pretrained(__A , config=__A ) self.assertIsInstance(__A , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def _a ( self ) -> Any: with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('./tests/fixtures/vocab.txt' , os.path.join(__A , 'vocab.txt' ) ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A , tokenizer_type='bert' , use_fast=__A ) self.assertIsInstance(__A , __A ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('./tests/fixtures/vocab.json' , os.path.join(__A , 'vocab.json' ) ) shutil.copy('./tests/fixtures/merges.txt' , os.path.join(__A , 'merges.txt' ) ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A , tokenizer_type='gpt2' , use_fast=__A ) self.assertIsInstance(__A , __A ) @require_tokenizers def _a ( self ) -> str: with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('./tests/fixtures/vocab.txt' , os.path.join(__A , 'vocab.txt' ) ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A , tokenizer_type='bert' ) self.assertIsInstance(__A , __A ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('./tests/fixtures/vocab.json' , os.path.join(__A , 'vocab.json' ) ) shutil.copy('./tests/fixtures/merges.txt' , os.path.join(__A , 'merges.txt' ) ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A , tokenizer_type='gpt2' ) self.assertIsInstance(__A , __A ) def _a ( self ) -> Any: with pytest.raises(__A ): AutoTokenizer.from_pretrained('./' , tokenizer_type='xxx' ) @require_tokenizers def _a ( self ) -> List[str]: for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: __UpperCamelCase =tokenizer_class.from_pretrained('wietsedv/bert-base-dutch-cased' ) self.assertIsInstance(__A , (BertTokenizer, BertTokenizerFast) ) if isinstance(__A , __A ): self.assertEqual(tokenizer.basic_tokenizer.do_lower_case , __A ) else: self.assertEqual(tokenizer.do_lower_case , __A ) self.assertEqual(tokenizer.model_max_length , 512 ) @require_tokenizers def _a ( self ) -> Optional[Any]: for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: with self.assertRaisesRegex( __A , 'julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier' , ): __UpperCamelCase =tokenizer_class.from_pretrained('julien-c/herlolip-not-exists' ) def _a ( self ) -> Any: # tests: https://github.com/huggingface/transformers/pull/13251 # 1. models with `-`, e.g. xlm-roberta -> xlm_roberta # 2. models that don't remap 1-1 from model-name to model file, e.g., openai-gpt -> openai __UpperCamelCase =TOKENIZER_MAPPING.values() __UpperCamelCase =[] for slow_tok, fast_tok in tokenizers: if slow_tok is not None: tokenizer_names.append(slow_tok.__name__ ) if fast_tok is not None: tokenizer_names.append(fast_tok.__name__ ) for tokenizer_name in tokenizer_names: # must find the right class tokenizer_class_from_name(__A ) @require_tokenizers def _a ( self ) -> Any: self.assertIsInstance(AutoTokenizer.from_pretrained('bert-base-cased' , use_fast=__A ) , __A ) self.assertIsInstance(AutoTokenizer.from_pretrained('bert-base-cased' ) , __A ) @require_tokenizers def _a ( self ) -> Any: __UpperCamelCase =AutoTokenizer.from_pretrained('distilbert-base-uncased' , do_lower_case=__A ) __UpperCamelCase ='Hello, world. How are you?' __UpperCamelCase =tokenizer.tokenize(__A ) self.assertEqual('[UNK]' , tokens[0] ) __UpperCamelCase =AutoTokenizer.from_pretrained('microsoft/mpnet-base' , do_lower_case=__A ) __UpperCamelCase =tokenizer.tokenize(__A ) self.assertEqual('[UNK]' , tokens[0] ) @require_tokenizers def _a ( self ) -> str: __UpperCamelCase =AutoTokenizer.from_pretrained('robot-test/dummy-tokenizer-fast-with-model-config' ) self.assertEqual(type(__A ) , __A ) self.assertEqual(tokenizer.model_max_length , 512 ) self.assertEqual(tokenizer.vocab_size , 30000 ) self.assertEqual(tokenizer.unk_token , '[UNK]' ) self.assertEqual(tokenizer.padding_side , 'right' ) self.assertEqual(tokenizer.truncation_side , 'right' ) def _a ( self ) -> int: __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) self.assertIsInstance(__A , (BertTokenizer, BertTokenizerFast) ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__A ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) self.assertIsInstance(__A , tokenizer.__class__ ) self.assertEqual(tokenizera.vocab_size , 12 ) def _a ( self ) -> List[str]: __UpperCamelCase =AutoTokenizer.from_pretrained('ctrl' ) # There is no fast CTRL so this always gives us a slow tokenizer. self.assertIsInstance(__A , __A ) def _a ( self ) -> Union[str, Any]: # Check we can load the tokenizer config of an online model. __UpperCamelCase =get_tokenizer_config('bert-base-cased' ) __UpperCamelCase =config.pop('_commit_hash' , __A ) # If we ever update bert-base-cased tokenizer config, this dict here will need to be updated. self.assertEqual(__A , {'do_lower_case': False} ) # This model does not have a tokenizer_config so we get back an empty dict. __UpperCamelCase =get_tokenizer_config(__A ) self.assertDictEqual(__A , {} ) # A tokenizer saved with `save_pretrained` always creates a tokenizer config. __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__A ) __UpperCamelCase =get_tokenizer_config(__A ) # Check the class of the tokenizer was properly saved (note that it always saves the slow class). self.assertEqual(config['tokenizer_class'] , 'BertTokenizer' ) def _a ( self ) -> str: try: AutoConfig.register('custom' , __A ) AutoTokenizer.register(__A , slow_tokenizer_class=__A ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__A ): AutoTokenizer.register(__A , slow_tokenizer_class=__A ) __UpperCamelCase =CustomTokenizer.from_pretrained(__A ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__A ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) self.assertIsInstance(__A , __A ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] @require_tokenizers def _a ( self ) -> int: try: AutoConfig.register('custom' , __A ) # Can register in two steps AutoTokenizer.register(__A , slow_tokenizer_class=__A ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, None) ) AutoTokenizer.register(__A , fast_tokenizer_class=__A ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) del TOKENIZER_MAPPING._extra_content[CustomConfig] # Can register in one step AutoTokenizer.register( __A , slow_tokenizer_class=__A , fast_tokenizer_class=__A ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__A ): AutoTokenizer.register(__A , fast_tokenizer_class=__A ) # We pass through a bert tokenizer fast cause there is no converter slow to fast for our new toknizer # and that model does not have a tokenizer.json with tempfile.TemporaryDirectory() as tmp_dir: __UpperCamelCase =BertTokenizerFast.from_pretrained(__A ) bert_tokenizer.save_pretrained(__A ) __UpperCamelCase =CustomTokenizerFast.from_pretrained(__A ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__A ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A , use_fast=__A ) self.assertIsInstance(__A , __A ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def _a ( self ) -> Any: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__A ): __UpperCamelCase =AutoTokenizer.from_pretrained('hf-internal-testing/test_dynamic_tokenizer' ) # If remote code is disabled, we can't load this config. with self.assertRaises(__A ): __UpperCamelCase =AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=__A ) __UpperCamelCase =AutoTokenizer.from_pretrained('hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=__A ) self.assertTrue(tokenizer.special_attribute_present ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__A ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A , trust_remote_code=__A ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version __UpperCamelCase =AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=__A , use_fast=__A ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__A ) __UpperCamelCase =AutoTokenizer.from_pretrained(__A , trust_remote_code=__A , use_fast=__A ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , 'NewTokenizer' ) @require_tokenizers def _a ( self ) -> Tuple: class UpperCAmelCase__ ( __lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Any = False class UpperCAmelCase__ ( __lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : str = NewTokenizer UpperCAmelCase__ : str = False try: AutoConfig.register('custom' , __A ) AutoTokenizer.register(__A , slow_tokenizer_class=__A ) AutoTokenizer.register(__A , fast_tokenizer_class=__A ) # If remote code is not set, the default is to use local __UpperCamelCase =AutoTokenizer.from_pretrained('hf-internal-testing/test_dynamic_tokenizer' ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) self.assertFalse(tokenizer.special_attribute_present ) __UpperCamelCase =AutoTokenizer.from_pretrained('hf-internal-testing/test_dynamic_tokenizer' , use_fast=__A ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertFalse(tokenizer.special_attribute_present ) # If remote code is disabled, we load the local one. __UpperCamelCase =AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=__A ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) self.assertFalse(tokenizer.special_attribute_present ) __UpperCamelCase =AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=__A , use_fast=__A ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertFalse(tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub __UpperCamelCase =AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=__A ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) self.assertTrue(tokenizer.special_attribute_present ) __UpperCamelCase =AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=__A , use_fast=__A ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertTrue(tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def _a ( self ) -> str: __UpperCamelCase =AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer_legacy' , trust_remote_code=__A ) self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version __UpperCamelCase =AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer_legacy' , trust_remote_code=__A , use_fast=__A ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) def _a ( self ) -> Dict: with self.assertRaisesRegex( __A , 'bert-base is not a local folder and is not a valid model identifier' ): __UpperCamelCase =AutoTokenizer.from_pretrained('bert-base' ) def _a ( self ) -> Optional[Any]: with self.assertRaisesRegex( __A , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): __UpperCamelCase =AutoTokenizer.from_pretrained(__A , revision='aaaaaa' ) def _a ( self ) -> Any: # Make sure we have cached the tokenizer. __UpperCamelCase =AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: __UpperCamelCase =AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
62
'''simple docstring''' from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Optional[int] , __A : list[list[int]] ): __UpperCamelCase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(__A ) != 0: __UpperCamelCase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__A ) != cols: raise error for value in row: if not isinstance(__A , (int, float) ): raise error __UpperCamelCase = rows else: __UpperCamelCase = [] def _lowerCamelCase ( self : int ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def _lowerCamelCase ( self : str ): return len(self.rows ) @property def _lowerCamelCase ( self : Any ): return len(self.rows[0] ) @property def _lowerCamelCase ( self : Optional[Any] ): return (self.num_rows, self.num_columns) @property def _lowerCamelCase ( self : Dict ): return self.order[0] == self.order[1] def _lowerCamelCase ( self : Any ): __UpperCamelCase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Any ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def _lowerCamelCase ( self : List[str] ): return bool(self.determinant() ) def _lowerCamelCase ( self : Dict , __A : int , __A : int ): __UpperCamelCase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__A ).determinant() def _lowerCamelCase ( self : Dict , __A : int , __A : int ): if (row + column) % 2 == 0: return self.get_minor(__A , __A ) return -1 * self.get_minor(__A , __A ) def _lowerCamelCase ( self : List[str] ): return Matrix( [ [self.get_minor(__A , __A ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def _lowerCamelCase ( self : Union[str, Any] ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : Union[str, Any] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(__A ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def _lowerCamelCase ( self : List[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in row: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(__A ) else: __UpperCamelCase = self.rows[0:position] + [row] + self.rows[position:] def _lowerCamelCase ( self : Optional[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in column: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: __UpperCamelCase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: __UpperCamelCase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : Tuple , __A : object ): if not isinstance(__A , __A ): return NotImplemented return self.rows == other.rows def __ne__( self : Any , __A : object ): return not self == other def __neg__( self : List[Any] ): return self * -1 def __add__( self : List[str] , __A : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : str , __A : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : str , __A : Matrix | int | float ): if isinstance(__A , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__A , __A ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(__A , __A ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Union[str, Any] , __A : int ): if not isinstance(__A , __A ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) __UpperCamelCase = self for _ in range(other - 1 ): result *= self return result @classmethod def _lowerCamelCase ( cls : Tuple , __A : list[int] , __A : list[int] ): return sum(row[i] * column[i] for i in range(len(__A ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
53
0
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Tuple , __a : Optional[int] , __a : str=13 , __a : Any=7 , __a : Optional[Any]=6 , __a : Any=17 , __a : int=23 , __a : Dict=11 , __a : str=True , ): _a = parent _a = batch_size _a = seq_length _a = act_dim _a = state_dim _a = hidden_size _a = max_length _a = is_training def UpperCamelCase__ ( self : List[Any] ): _a = floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) _a = floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) _a = floats_tensor((self.batch_size, self.seq_length, 1) ) _a = floats_tensor((self.batch_size, self.seq_length, 1) ) _a = ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) _a = random_attention_mask((self.batch_size, self.seq_length) ) _a = self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def UpperCamelCase__ ( self : int ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def UpperCamelCase__ ( self : Any , __a : Union[str, Any] , __a : Optional[int] , __a : List[Any] , __a : Union[str, Any] , __a : List[str] , __a : int , __a : str , ): _a = DecisionTransformerModel(config=__a ) model.to(__a ) model.eval() _a = model(__a , __a , __a , __a , __a , __a ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def UpperCamelCase__ ( self : Union[str, Any] ): _a = self.prepare_config_and_inputs() ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) = config_and_inputs _a = { "states": states, "actions": actions, "rewards": rewards, "returns_to_go": returns_to_go, "timesteps": timesteps, "attention_mask": attention_mask, } return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE (lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __a =(DecisionTransformerModel,) if is_torch_available() else () __a =() __a ={'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids __a =False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features __a =False __a =False __a =False __a =False __a =False __a =False __a =False __a =False __a =False def UpperCamelCase__ ( self : Optional[Any] ): _a = DecisionTransformerModelTester(self ) _a = ConfigTester(self , config_class=__a , hidden_size=37 ) def UpperCamelCase__ ( self : Tuple ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self : Tuple ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) @slow def UpperCamelCase__ ( self : Optional[int] ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a = DecisionTransformerModel.from_pretrained(__a ) self.assertIsNotNone(__a ) def UpperCamelCase__ ( self : Optional[int] ): _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = model_class(__a ) _a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a = [*signature.parameters.keys()] _a = [ "states", "actions", "rewards", "returns_to_go", "timesteps", "attention_mask", ] self.assertListEqual(arg_names[: len(__a )] , __a ) @require_torch class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" @slow def UpperCamelCase__ ( self : str ): _a = 2 # number of steps of autoregressive prediction we will perform _a = 10 # defined by the RL environment, may be normalized _a = DecisionTransformerModel.from_pretrained("edbeeching/decision-transformer-gym-hopper-expert" ) _a = model.to(__a ) _a = model.config torch.manual_seed(0 ) _a = torch.randn(1 , 1 , config.state_dim ).to(device=__a , dtype=torch.floataa ) # env.reset() _a = torch.tensor( [[0.242793, -0.28693074, 0.8742613], [0.67815274, -0.08101085, -0.12952147]] , device=__a ) _a = torch.tensor(__a , device=__a , dtype=torch.floataa ).reshape(1 , 1 , 1 ) _a = state _a = torch.zeros(1 , 0 , config.act_dim , device=__a , dtype=torch.floataa ) _a = torch.zeros(1 , 0 , device=__a , dtype=torch.floataa ) _a = torch.tensor(0 , device=__a , dtype=torch.long ).reshape(1 , 1 ) for step in range(__a ): _a = torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=__a )] , dim=1 ) _a = torch.cat([rewards, torch.zeros(1 , 1 , device=__a )] , dim=1 ) _a = torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): _a , _a , _a = model( states=__a , actions=__a , rewards=__a , returns_to_go=__a , timesteps=__a , attention_mask=__a , return_dict=__a , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1e-4 ) ) _a , _a , _a , _a = ( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=__a , dtype=torch.floataa ), 1.0, False, {}, ) _a = action_pred[0, -1] _a = torch.cat([states, state] , dim=1 ) _a = returns_to_go[0, -1] - reward _a = torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) _a = torch.cat( [timesteps, torch.ones((1, 1) , device=__a , dtype=torch.long ) * (step + 1)] , dim=1 )
364
'''simple docstring''' from random import randint, random def _lowerCamelCase ( lowercase : int , lowercase : int , lowercase : int , lowercase : bool = False , lowercase : bool = False , lowercase : int = 5 , ) -> list: _a = [[-1] * number_of_cells] # Create a highway without any car _a = 0 _a = max(lowercase , 0 ) while i < number_of_cells: _a = ( randint(0 , lowercase ) if random_speed else initial_speed ) # Place the cars i += ( randint(1 , max_speed * 2 ) if random_frequency else frequency ) # Arbitrary number, may need tuning return highway def _lowerCamelCase ( lowercase : list , lowercase : int ) -> int: _a = 0 _a = highway_now[car_index + 1 :] for cell in range(len(lowercase ) ): # May need a better name for this if cells[cell] != -1: # If the cell is not empty then return distance # we have the distance we wanted distance += 1 # Here if the car is near the end of the highway return distance + get_distance(lowercase , -1 ) def _lowerCamelCase ( lowercase : list , lowercase : float , lowercase : int ) -> list: _a = len(lowercase ) # Beforce calculations, the highway is empty _a = [-1] * number_of_cells for car_index in range(lowercase ): if highway_now[car_index] != -1: # Add 1 to the current speed of the car and cap the speed _a = min(highway_now[car_index] + 1 , lowercase ) # Number of empty cell before the next car _a = get_distance(lowercase , lowercase ) - 1 # We can't have the car causing an accident _a = min(next_highway[car_index] , lowercase ) if random() < probability: # Randomly, a driver will slow down _a = max(next_highway[car_index] - 1 , 0 ) return next_highway def _lowerCamelCase ( lowercase : list , lowercase : int , lowercase : float , lowercase : int ) -> list: _a = len(highway[0] ) for i in range(lowercase ): _a = update(highway[i] , lowercase , lowercase ) _a = [-1] * number_of_cells for car_index in range(lowercase ): _a = next_speeds_calculated[car_index] if speed != -1: # Change the position based on the speed (with % to create the loop) _a = (car_index + speed) % number_of_cells # Commit the change of position _a = speed highway.append(lowercase ) return highway if __name__ == "__main__": import doctest doctest.testmod()
346
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class __magic_name__ ( unittest.TestCase ): def __init__( self : Dict ,_UpperCAmelCase : Union[str, Any] ,_UpperCAmelCase : str=7 ,_UpperCAmelCase : int=3 ,_UpperCAmelCase : str=18 ,_UpperCAmelCase : List[Any]=30 ,_UpperCAmelCase : str=400 ,_UpperCAmelCase : Any=True ,_UpperCAmelCase : List[str]=32 ,_UpperCAmelCase : Dict=True ,): _a : str = parent _a : Union[str, Any] = batch_size _a : Tuple = num_channels _a : Optional[Any] = image_size _a : List[Any] = min_resolution _a : Dict = max_resolution _a : Optional[Any] = do_resize _a : Dict = size_divisor _a : Union[str, Any] = do_rescale def __lowercase ( self : Any ): return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class __magic_name__ ( _UpperCamelCase , unittest.TestCase ): lowerCAmelCase : str = GLPNImageProcessor if is_vision_available() else None def __lowercase ( self : List[str] ): _a : List[Any] = GLPNImageProcessingTester(self ) @property def __lowercase ( self : List[Any] ): return self.image_processor_tester.prepare_image_processor_dict() def __lowercase ( self : int ): _a : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_UpperCAmelCase ,'do_resize' ) ) self.assertTrue(hasattr(_UpperCAmelCase ,'size_divisor' ) ) self.assertTrue(hasattr(_UpperCAmelCase ,'resample' ) ) self.assertTrue(hasattr(_UpperCAmelCase ,'do_rescale' ) ) def __lowercase ( self : Any ): pass def __lowercase ( self : Union[str, Any] ): # Initialize image_processing _a : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _a : Union[str, Any] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase ,Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) _a : List[Any] = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def __lowercase ( self : Optional[Any] ): # Initialize image_processing _a : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _a : Any = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCAmelCase ,numpify=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase ,np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) _a : str = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def __lowercase ( self : Dict ): # Initialize image_processing _a : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _a : Any = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCAmelCase ,torchify=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase ,torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) _a : str = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
89
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCAmelCase = { '''configuration_lilt''': ['''LILT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LiltConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ '''LILT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LiltForQuestionAnswering''', '''LiltForSequenceClassification''', '''LiltForTokenClassification''', '''LiltModel''', '''LiltPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
89
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase = logging.get_logger(__name__) class _a ( UpperCamelCase__ ): _lowercase : Any = ['''pixel_values'''] def __init__( self: Optional[int] , UpperCamelCase_: bool = True , UpperCamelCase_: Dict[str, int] = None , UpperCamelCase_: PILImageResampling = PILImageResampling.BICUBIC , UpperCamelCase_: bool = True , UpperCamelCase_: Union[int, float] = 1 / 255 , UpperCamelCase_: bool = True , UpperCamelCase_: Optional[Union[float, List[float]]] = None , UpperCamelCase_: Optional[Union[float, List[float]]] = None , UpperCamelCase_: bool = True , **UpperCamelCase_: Optional[Any] , ) -> None: """simple docstring""" super().__init__(**UpperCamelCase_ ) lowercase__ = size if size is not None else {'''height''': 384, '''width''': 384} lowercase__ = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) lowercase__ = do_resize lowercase__ = size lowercase__ = resample lowercase__ = do_rescale lowercase__ = rescale_factor lowercase__ = do_normalize lowercase__ = image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowercase__ = image_std if image_std is not None else OPENAI_CLIP_STD lowercase__ = do_convert_rgb def lowerCamelCase_ ( self: int , UpperCamelCase_: np.ndarray , UpperCamelCase_: Dict[str, int] , UpperCamelCase_: PILImageResampling = PILImageResampling.BICUBIC , UpperCamelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_: Any , ) -> np.ndarray: """simple docstring""" lowercase__ = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) if "height" not in size or "width" not in size: raise ValueError(f'The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}' ) lowercase__ = (size['''height'''], size['''width''']) return resize(UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase_ ( self: str , UpperCamelCase_: np.ndarray , UpperCamelCase_: Union[int, float] , UpperCamelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_: Optional[int] , ) -> List[str]: """simple docstring""" return rescale(UpperCamelCase_ , scale=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase_: np.ndarray , UpperCamelCase_: Union[float, List[float]] , UpperCamelCase_: Union[float, List[float]] , UpperCamelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_: Optional[int] , ) -> np.ndarray: """simple docstring""" return normalize(UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase_: ImageInput , UpperCamelCase_: Optional[bool] = None , UpperCamelCase_: Optional[Dict[str, int]] = None , UpperCamelCase_: PILImageResampling = None , UpperCamelCase_: Optional[bool] = None , UpperCamelCase_: Optional[float] = None , UpperCamelCase_: Optional[bool] = None , UpperCamelCase_: Optional[Union[float, List[float]]] = None , UpperCamelCase_: Optional[Union[float, List[float]]] = None , UpperCamelCase_: Optional[Union[str, TensorType]] = None , UpperCamelCase_: bool = None , UpperCamelCase_: ChannelDimension = ChannelDimension.FIRST , **UpperCamelCase_: Union[str, Any] , ) -> PIL.Image.Image: """simple docstring""" lowercase__ = do_resize if do_resize is not None else self.do_resize lowercase__ = resample if resample is not None else self.resample lowercase__ = do_rescale if do_rescale is not None else self.do_rescale lowercase__ = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase__ = do_normalize if do_normalize is not None else self.do_normalize lowercase__ = image_mean if image_mean is not None else self.image_mean lowercase__ = image_std if image_std is not None else self.image_std lowercase__ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowercase__ = size if size is not None else self.size lowercase__ = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) lowercase__ = make_list_of_images(UpperCamelCase_ ) if not valid_images(UpperCamelCase_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowercase__ = [convert_to_rgb(UpperCamelCase_ ) for image in images] # All transformations expect numpy arrays. lowercase__ = [to_numpy_array(UpperCamelCase_ ) for image in images] if do_resize: lowercase__ = [self.resize(image=UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ ) for image in images] if do_rescale: lowercase__ = [self.rescale(image=UpperCamelCase_ , scale=UpperCamelCase_ ) for image in images] if do_normalize: lowercase__ = [self.normalize(image=UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ ) for image in images] lowercase__ = [to_channel_dimension_format(UpperCamelCase_ , UpperCamelCase_ ) for image in images] lowercase__ = BatchFeature(data={'''pixel_values''': images} , tensor_type=UpperCamelCase_ ) return encoded_outputs
368
import gc import random import unittest import numpy as np import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImgaImgPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import floats_tensor, load_image, load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class _a ( UpperCamelCase__ , unittest.TestCase ): _lowercase : Union[str, Any] = ShapEImgaImgPipeline _lowercase : Optional[Any] = ['''image'''] _lowercase : Optional[int] = ['''image'''] _lowercase : Optional[int] = [ '''num_images_per_prompt''', '''num_inference_steps''', '''generator''', '''latents''', '''guidance_scale''', '''frame_size''', '''output_type''', '''return_dict''', ] _lowercase : Tuple = False @property def lowerCamelCase_ ( self: List[Any] ) -> Union[str, Any]: """simple docstring""" return 32 @property def lowerCamelCase_ ( self: Union[str, Any] ) -> int: """simple docstring""" return 32 @property def lowerCamelCase_ ( self: str ) -> List[str]: """simple docstring""" return self.time_input_dim * 4 @property def lowerCamelCase_ ( self: List[Any] ) -> str: """simple docstring""" return 8 @property def lowerCamelCase_ ( self: int ) -> Dict: """simple docstring""" torch.manual_seed(0 ) lowercase__ = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=64 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=1 , ) lowercase__ = CLIPVisionModel(UpperCamelCase_ ) return model @property def lowerCamelCase_ ( self: Dict ) -> List[Any]: """simple docstring""" lowercase__ = CLIPImageProcessor( crop_size=224 , do_center_crop=UpperCamelCase_ , do_normalize=UpperCamelCase_ , do_resize=UpperCamelCase_ , image_mean=[0.48145466, 0.4578275, 0.40821073] , image_std=[0.26862954, 0.26130258, 0.27577711] , resample=3 , size=224 , ) return image_processor @property def lowerCamelCase_ ( self: List[str] ) -> str: """simple docstring""" torch.manual_seed(0 ) lowercase__ = { '''num_attention_heads''': 2, '''attention_head_dim''': 16, '''embedding_dim''': self.time_input_dim, '''num_embeddings''': 32, '''embedding_proj_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''num_layers''': 1, '''clip_embed_dim''': self.time_input_dim * 2, '''additional_embeddings''': 0, '''time_embed_act_fn''': '''gelu''', '''norm_in_type''': '''layer''', '''embedding_proj_norm_type''': '''layer''', '''encoder_hid_proj_type''': None, '''added_emb_type''': None, } lowercase__ = PriorTransformer(**UpperCamelCase_ ) return model @property def lowerCamelCase_ ( self: Dict ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) lowercase__ = { '''param_shapes''': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), '''d_latent''': self.time_input_dim, '''d_hidden''': self.renderer_dim, '''n_output''': 12, '''background''': ( 0.1, 0.1, 0.1, ), } lowercase__ = ShapERenderer(**UpperCamelCase_ ) return model def lowerCamelCase_ ( self: str ) -> Any: """simple docstring""" lowercase__ = self.dummy_prior lowercase__ = self.dummy_image_encoder lowercase__ = self.dummy_image_processor lowercase__ = self.dummy_renderer lowercase__ = HeunDiscreteScheduler( beta_schedule='''exp''' , num_train_timesteps=1_024 , prediction_type='''sample''' , use_karras_sigmas=UpperCamelCase_ , clip_sample=UpperCamelCase_ , clip_sample_range=1.0 , ) lowercase__ = { '''prior''': prior, '''image_encoder''': image_encoder, '''image_processor''': image_processor, '''renderer''': renderer, '''scheduler''': scheduler, } return components def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase_: Tuple , UpperCamelCase_: Optional[int]=0 ) -> Tuple: """simple docstring""" lowercase__ = floats_tensor((1, 3, 64, 64) , rng=random.Random(UpperCamelCase_ ) ).to(UpperCamelCase_ ) if str(UpperCamelCase_ ).startswith('''mps''' ): lowercase__ = torch.manual_seed(UpperCamelCase_ ) else: lowercase__ = torch.Generator(device=UpperCamelCase_ ).manual_seed(UpperCamelCase_ ) lowercase__ = { '''image''': input_image, '''generator''': generator, '''num_inference_steps''': 1, '''frame_size''': 32, '''output_type''': '''np''', } return inputs def lowerCamelCase_ ( self: int ) -> str: """simple docstring""" lowercase__ = '''cpu''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**UpperCamelCase_ ) lowercase__ = pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase__ = pipe(**self.get_dummy_inputs(UpperCamelCase_ ) ) lowercase__ = output.images[0] lowercase__ = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) lowercase__ = np.array( [ 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase_ ( self: int ) -> int: """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def lowerCamelCase_ ( self: List[str] ) -> List[Any]: """simple docstring""" lowercase__ = torch_device == '''cpu''' lowercase__ = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=UpperCamelCase_ , relax_max_difference=UpperCamelCase_ , ) def lowerCamelCase_ ( self: List[str] ) -> str: """simple docstring""" lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**UpperCamelCase_ ) lowercase__ = pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase__ = 1 lowercase__ = 2 lowercase__ = self.get_dummy_inputs(UpperCamelCase_ ) for key in inputs.keys(): if key in self.batch_params: lowercase__ = batch_size * [inputs[key]] lowercase__ = pipe(**UpperCamelCase_ , num_images_per_prompt=UpperCamelCase_ )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class _a ( unittest.TestCase ): def lowerCamelCase_ ( self: Optional[int] ) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self: str ) -> str: """simple docstring""" lowercase__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/corgi.png''' ) lowercase__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/test_shap_e_img2img_out.npy''' ) lowercase__ = ShapEImgaImgPipeline.from_pretrained('''openai/shap-e-img2img''' ) lowercase__ = pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowercase__ = torch.Generator(device=UpperCamelCase_ ).manual_seed(0 ) lowercase__ = pipe( UpperCamelCase_ , generator=UpperCamelCase_ , guidance_scale=3.0 , num_inference_steps=64 , frame_size=64 , output_type='''np''' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(UpperCamelCase_ , UpperCamelCase_ )
93
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _A = { '''configuration_rag''': ['''RagConfig'''], '''retrieval_rag''': ['''RagRetriever'''], '''tokenization_rag''': ['''RagTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''RagModel''', '''RagPreTrainedModel''', '''RagSequenceForGeneration''', '''RagTokenForGeneration''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''TFRagModel''', '''TFRagPreTrainedModel''', '''TFRagSequenceForGeneration''', '''TFRagTokenForGeneration''', ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys _A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
122
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
"""simple docstring""" def lowerCAmelCase__ ( _UpperCamelCase : Union[str, Any] ) -> str: """simple docstring""" snake_case = [] snake_case = set({'(', '[', '{'} ) snake_case = set({')', ']', '}'} ) snake_case = {'{': '}', '[': ']', '(': ')'} for i in range(len(_UpperCamelCase ) ): if s[i] in open_brackets: stack.append(s[i] ) elif s[i] in closed_brackets and ( len(_UpperCamelCase ) == 0 or (len(_UpperCamelCase ) > 0 and open_to_closed[stack.pop()] != s[i]) ): return False return len(_UpperCamelCase ) == 0 def lowerCAmelCase__ ( ) -> List[str]: """simple docstring""" snake_case = input('Enter sequence of brackets: ' ) if is_balanced(_UpperCamelCase ): print(_UpperCamelCase , 'is balanced' ) else: print(_UpperCamelCase , 'is not balanced' ) if __name__ == "__main__": main()
149
"""simple docstring""" SCREAMING_SNAKE_CASE__ = {str(digit): digit**5 for digit in range(10)} def lowerCAmelCase__ ( _UpperCamelCase : int ) -> int: """simple docstring""" return sum(DIGITS_FIFTH_POWER[digit] for digit in str(_UpperCamelCase ) ) def lowerCAmelCase__ ( ) -> int: """simple docstring""" return sum( number for number in range(1_0_0_0 , 1_0_0_0_0_0_0 ) if number == digits_fifth_powers_sum(_UpperCamelCase ) ) if __name__ == "__main__": print(solution())
149
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase ): """simple docstring""" @slow def _lowerCAmelCase ( self : int ) -> int: """simple docstring""" _UpperCAmelCase : List[Any] = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) _UpperCAmelCase : Dict = { "input_ids": tf.convert_to_tensor([[0, 2_6_4_6, 1_0_2_6_9, 8_3, 9_9_9_4_2, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } _UpperCAmelCase : Tuple = model(lowerCAmelCase__ )["last_hidden_state"] _UpperCAmelCase : Optional[int] = tf.TensorShape((1, 6, 7_6_8) ) self.assertEqual(output.shape , lowerCAmelCase__ ) # compare the actual values for a slice. _UpperCAmelCase : Dict = tf.convert_to_tensor( [ [ [0.068_1762, 0.1089_4451, 0.0677_2504], [-0.0642_3668, 0.0236_6615, 0.0432_9344], [-0.0605_7295, 0.0997_4135, -0.0007_0584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
145
'''simple docstring''' import math def __UpperCAmelCase ( a_: int ): return math.sqrt(a_ ) * math.sqrt(a_ ) == num def __UpperCAmelCase ( a_: int ): _UpperCAmelCase : Dict = 0 _UpperCAmelCase : List[str] = n while left <= right: _UpperCAmelCase : Dict = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: _UpperCAmelCase : int = mid - 1 else: _UpperCAmelCase : Tuple = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
145
1
'''simple docstring''' import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowercase_ : """simple docstring""" def __init__( self : Dict ,lowercase__ : Any ,lowercase__ : Dict=1_3 ,lowercase__ : List[Any]=3_0 ,lowercase__ : Optional[Any]=2 ,lowercase__ : List[str]=3 ,lowercase__ : Union[str, Any]=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=3_2 ,lowercase__ : Optional[Any]=5 ,lowercase__ : List[Any]=4 ,lowercase__ : str=3_7 ,lowercase__ : Optional[int]="gelu" ,lowercase__ : Any=0.1 ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : Tuple=1_0 ,lowercase__ : Tuple=0.0_2 ,lowercase__ : Tuple=None ,lowercase__ : Union[str, Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = scope __lowercase = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowercase = (image_size // patch_size) ** 2 __lowercase = num_patches + 1 def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : List[Any] ): return ViTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowercase__ ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Tuple ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = ViTModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ): __lowercase = ViTForMaskedImageModeling(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) self.parent.assertEqual( result.reconstruction.shape ,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __lowercase = 1 __lowercase = ViTForMaskedImageModeling(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.reconstruction.shape ,(self.batch_size, 1, self.image_size, self.image_size) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ,lowercase__ : List[Any] ,lowercase__ : List[Any] ): __lowercase = self.type_sequence_label_size __lowercase = ViTForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowercase = 1 __lowercase = ViTForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : int = ( {'feature-extraction': ViTModel, 'image-classification': ViTForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = True SCREAMING_SNAKE_CASE : Any = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = ViTModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): pass def SCREAMING_SNAKE_CASE ( self : int ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase__ ,nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ViTModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return ViTImageProcessor.from_pretrained('''google/vit-base-patch16-224''' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = ViTForImageClassification.from_pretrained('''google/vit-base-patch16-224''' ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self : Dict ): # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. __lowercase = ViTModel.from_pretrained('''facebook/dino-vits8''' ).to(lowercase__ ) __lowercase = ViTImageProcessor.from_pretrained('''facebook/dino-vits8''' ,size=4_8_0 ) __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ) __lowercase = inputs.pixel_values.to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(lowercase__ ,interpolate_pos_encoding=lowercase__ ) # verify the logits __lowercase = torch.Size((1, 3_6_0_1, 3_8_4) ) self.assertEqual(outputs.last_hidden_state.shape ,lowercase__ ) __lowercase = torch.tensor( [[4.2_3_4_0, 4.3_9_0_6, -6.6_6_9_2], [4.5_4_6_3, 1.8_9_2_8, -6.7_2_5_7], [4.4_4_2_9, 0.8_4_9_6, -5.8_5_8_5]] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] ,lowercase__ ,atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = ViTModel.from_pretrained('''facebook/dino-vits8''' ,torch_dtype=torch.floataa ,device_map='''auto''' ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ) __lowercase = inputs.pixel_values.to(lowercase__ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __lowercase = model(lowercase__ )
366
'''simple docstring''' import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = OpenAIGPTTokenizer SCREAMING_SNAKE_CASE : str = OpenAIGPTTokenizerFast SCREAMING_SNAKE_CASE : Optional[int] = True SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : int ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowercase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] __lowercase = dict(zip(lowercase__ ,range(len(lowercase__ ) ) ) ) __lowercase = ['''#version: 0.2''', '''l o''', '''lo w''', '''e r</w>''', ''''''] __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''vocab_file'''] ) __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file ,'''w''' ) as fp: fp.write(json.dumps(lowercase__ ) ) with open(self.merges_file ,'''w''' ) as fp: fp.write('''\n'''.join(lowercase__ ) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Optional[Any] ): return "lower newer", "lower newer" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = OpenAIGPTTokenizer(self.vocab_file ,self.merges_file ) __lowercase = '''lower''' __lowercase = ['''low''', '''er</w>'''] __lowercase = tokenizer.tokenize(lowercase__ ) self.assertListEqual(lowercase__ ,lowercase__ ) __lowercase = tokens + ['''<unk>'''] __lowercase = [1_4, 1_5, 2_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase__ ) ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Dict=1_5 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = self.rust_tokenizer_class.from_pretrained(lowercase__ ,**lowercase__ ) # Simple input __lowercase = '''This is a simple input''' __lowercase = ['''This is a simple input 1''', '''This is a simple input 2'''] __lowercase = ('''This is a simple input''', '''This is a pair''') __lowercase = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(lowercase__ ,tokenizer_r.encode ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ) # Simple input self.assertRaises(lowercase__ ,tokenizer_r.encode_plus ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ) # Simple input self.assertRaises( lowercase__ ,tokenizer_r.batch_encode_plus ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ,) # Pair input self.assertRaises(lowercase__ ,tokenizer_r.encode ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ) # Pair input self.assertRaises(lowercase__ ,tokenizer_r.encode_plus ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ) # Pair input self.assertRaises( lowercase__ ,tokenizer_r.batch_encode_plus ,lowercase__ ,max_length=lowercase__ ,padding='''max_length''' ,) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): pass @require_ftfy @require_spacy @require_tokenizers class lowercase_ (lowerCamelCase__ ): """simple docstring""" pass
52
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase_ : List[Any] = logging.get_logger(__name__) lowerCAmelCase_ : Dict = { 'google/mobilenet_v2_1.4_224': 'https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json', 'google/mobilenet_v2_1.0_224': 'https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json', 'google/mobilenet_v2_0.75_160': 'https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json', 'google/mobilenet_v2_0.35_96': 'https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class __SCREAMING_SNAKE_CASE (__a ): """simple docstring""" __a ="""mobilenet_v2""" def __init__( self : List[Any] , __a : Any=3 , __a : Tuple=2_24 , __a : Dict=1.0 , __a : Union[str, Any]=8 , __a : str=8 , __a : Dict=6 , __a : Dict=32 , __a : Optional[int]=True , __a : Any=True , __a : List[Any]="relu6" , __a : Tuple=True , __a : Dict=0.8 , __a : int=0.02 , __a : List[Any]=0.001 , __a : Union[str, Any]=2_55 , **__a : List[str] , ): super().__init__(**__a ) if depth_multiplier <= 0: raise ValueError("depth_multiplier must be greater than zero." ) _a = num_channels _a = image_size _a = depth_multiplier _a = depth_divisible_by _a = min_depth _a = expand_ratio _a = output_stride _a = first_layer_is_expansion _a = finegrained_output _a = hidden_act _a = tf_padding _a = classifier_dropout_prob _a = initializer_range _a = layer_norm_eps _a = semantic_loss_ignore_index class __SCREAMING_SNAKE_CASE (__a ): """simple docstring""" __a =version.parse('1.11' ) @property def UpperCamelCase__ ( self : str ): return OrderedDict([("pixel_values", {0: "batch"})] ) @property def UpperCamelCase__ ( self : Dict ): if self.task == "image-classification": return OrderedDict([("logits", {0: "batch"})] ) else: return OrderedDict([("last_hidden_state", {0: "batch"}), ("pooler_output", {0: "batch"})] ) @property def UpperCamelCase__ ( self : str ): return 1e-4
63
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _a ( __a ): __a : int = ["""image_processor""", """tokenizer"""] __a : Union[str, Any] = """ChineseCLIPImageProcessor""" __a : List[Any] = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self : Dict , lowercase : Union[str, Any]=None , lowercase : Dict=None , **lowercase : Optional[Any] ): '''simple docstring''' UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , lowercase , ) UpperCAmelCase = kwargs.pop('''feature_extractor''' ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(lowercase , lowercase ) UpperCAmelCase = self.image_processor def __call__( self : Tuple , lowercase : Optional[Any]=None , lowercase : Union[str, Any]=None , lowercase : int=None , **lowercase : Dict ): '''simple docstring''' if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: UpperCAmelCase = self.tokenizer(lowercase , return_tensors=lowercase , **lowercase ) if images is not None: UpperCAmelCase = self.image_processor(lowercase , return_tensors=lowercase , **lowercase ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowercase ) , tensor_type=lowercase ) def A ( self : int , *lowercase : Tuple , **lowercase : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*lowercase , **lowercase ) def A ( self : Optional[Any] , *lowercase : int , **lowercase : Optional[int] ): '''simple docstring''' return self.tokenizer.decode(*lowercase , **lowercase ) @property def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def A ( self : List[Any] ): '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , lowercase , ) return self.image_processor_class
34
0
import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin lowercase_ = get_tests_dir("""fixtures/test_sentencepiece.model""") lowercase_ = get_tests_dir("""fixtures/test_sentencepiece_bpe.model""") lowercase_ = 'pt' if is_torch_available() else 'tf' @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE (__SCREAMING_SNAKE_CASE , unittest.TestCase ): _UpperCamelCase : Union[str, Any] = CamembertTokenizer _UpperCamelCase : List[str] = CamembertTokenizerFast _UpperCamelCase : Any = True _UpperCamelCase : Union[str, Any] = True def SCREAMING_SNAKE_CASE_ ( self : List[str] )-> List[str]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing lowercase__ = CamembertTokenizer(_SCREAMING_SNAKE_CASE ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE_ ( self : Dict )-> Tuple: """simple docstring""" lowercase__ = "<pad>" lowercase__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] )-> Tuple: """simple docstring""" lowercase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>NOTUSED' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , 1_004 ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] )-> Dict: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_005 ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] )-> Dict: """simple docstring""" lowercase__ = CamembertTokenizer(_SCREAMING_SNAKE_CASE ) tokenizer.save_pretrained(self.tmpdirname ) lowercase__ = CamembertTokenizerFast.from_pretrained(self.tmpdirname ) lowercase__ = "I was born in 92000, and this is falsé." lowercase__ = tokenizer.encode(_SCREAMING_SNAKE_CASE ) lowercase__ = rust_tokenizer.encode(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowercase__ = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE ) lowercase__ = rust_tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) lowercase__ = tokenizer.convert_ids_to_tokens(_SCREAMING_SNAKE_CASE ) lowercase__ = rust_tokenizer.tokenize(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE_ ( self : Any )-> int: """simple docstring""" if not self.test_rust_tokenizer: return lowercase__ = self.get_tokenizer() lowercase__ = self.get_rust_tokenizer() lowercase__ = "I was born in 92000, and this is falsé." lowercase__ = tokenizer.tokenize(_SCREAMING_SNAKE_CASE ) lowercase__ = rust_tokenizer.tokenize(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowercase__ = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE ) lowercase__ = rust_tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowercase__ = self.get_rust_tokenizer() lowercase__ = tokenizer.encode(_SCREAMING_SNAKE_CASE ) lowercase__ = rust_tokenizer.encode(_SCREAMING_SNAKE_CASE ) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @slow def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] )-> List[str]: """simple docstring""" lowercase__ = {"input_ids": [[5, 54, 7_196, 297, 30, 23, 776, 18, 11, 3_215, 3_705, 8_252, 22, 3_164, 1_181, 2_116, 29, 16, 813, 25, 791, 3_314, 20, 3_446, 38, 27_575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9_088, 20, 1_517, 8, 22_804, 18_818, 10, 38, 629, 607, 607, 142, 19, 7_196, 867, 56, 10_326, 24, 2_267, 20, 416, 5_072, 15_612, 233, 734, 7, 2_399, 27, 16, 3_015, 1_649, 7, 24, 20, 4_338, 2_399, 27, 13, 3_400, 14, 13, 6_189, 8, 930, 9, 6]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. lowercase__ = [ "Le transformeur est un modèle d'apprentissage profond introduit en 2017, " "utilisé principalement dans le domaine du traitement automatique des langues (TAL).", "À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus " "pour gérer des données séquentielles, telles que le langage naturel, pour des tâches " "telles que la traduction et la synthèse de texte.", ] self.tokenizer_integration_test_util( expected_encoding=_SCREAMING_SNAKE_CASE , model_name='camembert-base' , revision='3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf' , sequences=_SCREAMING_SNAKE_CASE , )
364
import sys lowercase_ = ( """73167176531330624919225119674426574742355349194934""" """96983520312774506326239578318016984801869478851843""" """85861560789112949495459501737958331952853208805511""" """12540698747158523863050715693290963295227443043557""" """66896648950445244523161731856403098711121722383113""" """62229893423380308135336276614282806444486645238749""" """30358907296290491560440772390713810515859307960866""" """70172427121883998797908792274921901699720888093776""" """65727333001053367881220235421809751254540594752243""" """52584907711670556013604839586446706324415722155397""" """53697817977846174064955149290862569321978468622482""" """83972241375657056057490261407972968652414535100474""" """82166370484403199890008895243450658541227588666881""" """16427171479924442928230863465674813919123162824586""" """17866458359124566529476545682848912883142607690042""" """24219022671055626321111109370544217506941658960408""" """07198403850962455444362981230987879927244284909188""" """84580156166097919133875499200524063689912560717606""" """05886116467109405077541002256983155200055935729725""" """71636269561882670428252483600823257530420752963450""" ) def __UpperCamelCase (_SCREAMING_SNAKE_CASE = N ) -> int: lowercase__ = -sys.maxsize - 1 for i in range(len(_SCREAMING_SNAKE_CASE ) - 12 ): lowercase__ = 1 for j in range(13 ): product *= int(n[i + j] ) if product > largest_product: lowercase__ = product return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
269
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Tuple = logging.get_logger(__name__) A : Union[str, Any] = { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/config.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/config.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/config.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/config.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json", "roberta-large-openai-detector": "https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json", } class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = '''roberta''' def __init__( self : Dict , __magic_name__ : Tuple=50_265 , __magic_name__ : Optional[Any]=768 , __magic_name__ : List[str]=12 , __magic_name__ : Union[str, Any]=12 , __magic_name__ : Any=3_072 , __magic_name__ : Any="gelu" , __magic_name__ : Optional[Any]=0.1 , __magic_name__ : Optional[int]=0.1 , __magic_name__ : str=512 , __magic_name__ : Optional[int]=2 , __magic_name__ : Dict=0.02 , __magic_name__ : str=1e-12 , __magic_name__ : Optional[int]=1 , __magic_name__ : Any=0 , __magic_name__ : List[Any]=2 , __magic_name__ : Any="absolute" , __magic_name__ : Dict=True , __magic_name__ : Optional[Any]=None , **__magic_name__ : Optional[int] , ) -> Dict: super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_dropout_prob SCREAMING_SNAKE_CASE_ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ = max_position_embeddings SCREAMING_SNAKE_CASE_ = type_vocab_size SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = layer_norm_eps SCREAMING_SNAKE_CASE_ = position_embedding_type SCREAMING_SNAKE_CASE_ = use_cache SCREAMING_SNAKE_CASE_ = classifier_dropout class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def __A ( self : Optional[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": SCREAMING_SNAKE_CASE_ = {0: "batch", 1: "choice", 2: "sequence"} else: SCREAMING_SNAKE_CASE_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
118
from functools import lru_cache @lru_cache def a__ ( __UpperCamelCase ): if num < 0: raise ValueError("Number should not be negative." ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
118
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class lowercase__ : '''simple docstring''' def __init__( self : Any , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any]=2 , _UpperCAmelCase : Any=True , _UpperCAmelCase : int=False , _UpperCAmelCase : str=10 , _UpperCAmelCase : int=3 , _UpperCAmelCase : Optional[int]=32 * 8 , _UpperCAmelCase : str=32 * 8 , _UpperCAmelCase : Any=4 , _UpperCAmelCase : List[Any]=64 , ) -> str: '''simple docstring''' UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = is_training UpperCAmelCase_ = use_auxiliary_loss UpperCAmelCase_ = num_queries UpperCAmelCase_ = num_channels UpperCAmelCase_ = min_size UpperCAmelCase_ = max_size UpperCAmelCase_ = num_labels UpperCAmelCase_ = hidden_dim UpperCAmelCase_ = hidden_dim def lowercase__ ( self : int ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _UpperCAmelCase ) UpperCAmelCase_ = torch.ones([self.batch_size, self.min_size, self.max_size] , device=_UpperCAmelCase ) UpperCAmelCase_ = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=_UpperCAmelCase ) > 0.5 ).float() UpperCAmelCase_ = (torch.rand((self.batch_size, self.num_labels) , device=_UpperCAmelCase ) > 0.5).long() UpperCAmelCase_ = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowercase__ ( self : Any ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = MaskaFormerConfig( hidden_size=self.hidden_dim , ) UpperCAmelCase_ = self.num_queries UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = [1, 1, 1, 1] UpperCAmelCase_ = self.num_channels UpperCAmelCase_ = 64 UpperCAmelCase_ = 128 UpperCAmelCase_ = self.hidden_dim UpperCAmelCase_ = self.hidden_dim UpperCAmelCase_ = self.hidden_dim return config def lowercase__ ( self : Dict ) -> Dict: '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ = {"pixel_values": pixel_values, "pixel_mask": pixel_mask} return config, inputs_dict def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any] ) -> str: '''simple docstring''' UpperCAmelCase_ = output.encoder_hidden_states UpperCAmelCase_ = output.pixel_decoder_hidden_states UpperCAmelCase_ = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_UpperCAmelCase ) , config.decoder_layers ) def lowercase__ ( self : Tuple , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[int]=False ) -> str: '''simple docstring''' with torch.no_grad(): UpperCAmelCase_ = MaskaFormerModel(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model(pixel_values=_UpperCAmelCase , pixel_mask=_UpperCAmelCase ) UpperCAmelCase_ = model(_UpperCAmelCase , output_hidden_states=_UpperCAmelCase ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_UpperCAmelCase , _UpperCAmelCase ) def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = MaskaFormerForUniversalSegmentation(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() def comm_check_on_output(_UpperCAmelCase : List[Any] ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): UpperCAmelCase_ = model(pixel_values=_UpperCAmelCase , pixel_mask=_UpperCAmelCase ) UpperCAmelCase_ = model(_UpperCAmelCase ) comm_check_on_output(_UpperCAmelCase ) UpperCAmelCase_ = model( pixel_values=_UpperCAmelCase , pixel_mask=_UpperCAmelCase , mask_labels=_UpperCAmelCase , class_labels=_UpperCAmelCase ) comm_check_on_output(_UpperCAmelCase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class lowercase__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' UpperCamelCase = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () UpperCamelCase = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def lowercase__ ( self : Tuple ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = MaskaFormerModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=_UpperCAmelCase , has_text_modality=_UpperCAmelCase ) def lowercase__ ( self : List[str] ) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() def lowercase__ ( self : Dict ) -> str: '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_UpperCAmelCase , **_UpperCAmelCase , output_hidden_states=_UpperCAmelCase ) def lowercase__ ( self : Optional[int] ) -> Dict: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_UpperCAmelCase ) @unittest.skip(reason="Mask2Former does not use inputs_embeds" ) def lowercase__ ( self : Tuple ) -> Any: '''simple docstring''' pass @unittest.skip(reason="Mask2Former does not have a get_input_embeddings method" ) def lowercase__ ( self : Tuple ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="Mask2Former is not a generative model" ) def lowercase__ ( self : str ) -> Dict: '''simple docstring''' pass @unittest.skip(reason="Mask2Former does not use token embeddings" ) def lowercase__ ( self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason="Mask2Former has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def lowercase__ ( self : List[Any] ) -> Dict: '''simple docstring''' pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def lowercase__ ( self : List[str] ) -> List[str]: '''simple docstring''' pass def lowercase__ ( self : Tuple ) -> str: '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(_UpperCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _UpperCAmelCase ) @slow def lowercase__ ( self : List[Any] ) -> Dict: '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: UpperCAmelCase_ = MaskaFormerModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) def lowercase__ ( self : List[str] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = (self.model_tester.min_size,) * 2 UpperCAmelCase_ = { "pixel_values": torch.randn((2, 3, *size) , device=_UpperCAmelCase ), "mask_labels": torch.randn((2, 10, *size) , device=_UpperCAmelCase ), "class_labels": torch.zeros(2 , 10 , device=_UpperCAmelCase ).long(), } UpperCAmelCase_ = self.model_tester.get_config() UpperCAmelCase_ = MaskaFormerForUniversalSegmentation(_UpperCAmelCase ).to(_UpperCAmelCase ) UpperCAmelCase_ = model(**_UpperCAmelCase ) self.assertTrue(outputs.loss is not None ) def lowercase__ ( self : Union[str, Any] ) -> str: '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_UpperCAmelCase , **_UpperCAmelCase , output_hidden_states=_UpperCAmelCase ) def lowercase__ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(_UpperCAmelCase ).to(_UpperCAmelCase ) UpperCAmelCase_ = model(**_UpperCAmelCase , output_attentions=_UpperCAmelCase ) self.assertTrue(outputs.attentions is not None ) def lowercase__ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' if not self.model_tester.is_training: return UpperCAmelCase_ = self.all_model_classes[1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() UpperCAmelCase_ = model_class(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.train() UpperCAmelCase_ = model(_UpperCAmelCase , mask_labels=_UpperCAmelCase , class_labels=_UpperCAmelCase ).loss loss.backward() def lowercase__ ( self : str ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.all_model_classes[1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() UpperCAmelCase_ = True UpperCAmelCase_ = True UpperCAmelCase_ = model_class(_UpperCAmelCase ).to(_UpperCAmelCase ) model.train() UpperCAmelCase_ = model(_UpperCAmelCase , mask_labels=_UpperCAmelCase , class_labels=_UpperCAmelCase ) UpperCAmelCase_ = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() UpperCAmelCase_ = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() UpperCAmelCase_ = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() UpperCAmelCase_ = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_UpperCAmelCase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) lowerCamelCase = 1e-4 def a__ ( ): UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @slow class lowercase__ ( unittest.TestCase ): '''simple docstring''' @cached_property def lowercase__ ( self : Optional[Any] ) -> int: '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def lowercase__ ( self : Dict ) -> Optional[int]: '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_UpperCAmelCase ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(_UpperCAmelCase , return_tensors="pt" ).to(_UpperCAmelCase ) UpperCAmelCase_ = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_UpperCAmelCase , (1, 3, 384, 384) ) with torch.no_grad(): UpperCAmelCase_ = model(**_UpperCAmelCase ) UpperCAmelCase_ = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) UpperCAmelCase_ = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) UpperCAmelCase_ = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) def lowercase__ ( self : int ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_UpperCAmelCase ).eval() UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(_UpperCAmelCase , return_tensors="pt" ).to(_UpperCAmelCase ) UpperCAmelCase_ = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_UpperCAmelCase , (1, 3, 384, 384) ) with torch.no_grad(): UpperCAmelCase_ = model(**_UpperCAmelCase ) # masks_queries_logits UpperCAmelCase_ = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) UpperCAmelCase_ = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] UpperCAmelCase_ = torch.tensor(_UpperCAmelCase ).to(_UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) # class_queries_logits UpperCAmelCase_ = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) UpperCAmelCase_ = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _UpperCAmelCase , atol=_UpperCAmelCase ) ) def lowercase__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_UpperCAmelCase ).eval() UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="pt" , ) UpperCAmelCase_ = inputs["pixel_values"].to(_UpperCAmelCase ) UpperCAmelCase_ = [el.to(_UpperCAmelCase ) for el in inputs["mask_labels"]] UpperCAmelCase_ = [el.to(_UpperCAmelCase ) for el in inputs["class_labels"]] with torch.no_grad(): UpperCAmelCase_ = model(**_UpperCAmelCase ) self.assertTrue(outputs.loss is not None )
241
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py lowerCamelCase = """src/diffusers""" lowerCamelCase = """.""" # This is to make sure the diffusers module imported is the one in the repo. lowerCamelCase = importlib.util.spec_from_file_location( """diffusers""", os.path.join(DIFFUSERS_PATH, """__init__.py"""), submodule_search_locations=[DIFFUSERS_PATH], ) lowerCamelCase = spec.loader.load_module() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): return line.startswith(lowerCAmelCase__ ) or len(lowerCAmelCase__ ) <= 1 or re.search(r"^\s*\)(\s*->.*:|:)\s*$" , lowerCAmelCase__ ) is not None def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = object_name.split("." ) UpperCAmelCase_ = 0 # First let's find the module where our object lives. UpperCAmelCase_ = parts[i] while i < len(lowerCAmelCase__ ) and not os.path.isfile(os.path.join(lowerCAmelCase__ , f"""{module}.py""" ) ): i += 1 if i < len(lowerCAmelCase__ ): UpperCAmelCase_ = os.path.join(lowerCAmelCase__ , parts[i] ) if i >= len(lowerCAmelCase__ ): raise ValueError(f"""`object_name` should begin with the name of a module of diffusers but got {object_name}.""" ) with open(os.path.join(lowerCAmelCase__ , f"""{module}.py""" ) , "r" , encoding="utf-8" , newline="\n" ) as f: UpperCAmelCase_ = f.readlines() # Now let's find the class / func in the code! UpperCAmelCase_ = "" UpperCAmelCase_ = 0 for name in parts[i + 1 :]: while ( line_index < len(lowerCAmelCase__ ) and re.search(rf"""^{indent}(class|def)\s+{name}(\(|\:)""" , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowerCAmelCase__ ): raise ValueError(f""" {object_name} does not match any function or class in {module}.""" ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). UpperCAmelCase_ = line_index while line_index < len(lowerCAmelCase__ ) and _should_continue(lines[line_index] , lowerCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 UpperCAmelCase_ = lines[start_index:line_index] return "".join(lowerCAmelCase__ ) lowerCamelCase = re.compile(r"""^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)""") lowerCamelCase = re.compile(r"""^\s*(\S+)->(\S+)(\s+.*|$)""") lowerCamelCase = re.compile(r"""<FILL\s+[^>]*>""") def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = code.split("\n" ) UpperCAmelCase_ = 0 while idx < len(lowerCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowerCAmelCase__ ): return re.search(r"^(\s*)\S" , lines[idx] ).groups()[0] return "" def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = len(get_indent(lowerCAmelCase__ ) ) > 0 if has_indent: UpperCAmelCase_ = f"""class Bla:\n{code}""" UpperCAmelCase_ = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=lowerCAmelCase__ ) UpperCAmelCase_ = black.format_str(lowerCAmelCase__ , mode=lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = style_docstrings_in_code(lowerCAmelCase__ ) return result[len("class Bla:\n" ) :] if has_indent else result def a__ ( lowerCAmelCase__ , lowerCAmelCase__=False ): with open(lowerCAmelCase__ , "r" , encoding="utf-8" , newline="\n" ) as f: UpperCAmelCase_ = f.readlines() UpperCAmelCase_ = [] UpperCAmelCase_ = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowerCAmelCase__ ): UpperCAmelCase_ = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = search.groups() UpperCAmelCase_ = find_code_in_diffusers(lowerCAmelCase__ ) UpperCAmelCase_ = get_indent(lowerCAmelCase__ ) UpperCAmelCase_ = line_index + 1 if indent == theoretical_indent else line_index + 2 UpperCAmelCase_ = theoretical_indent UpperCAmelCase_ = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. UpperCAmelCase_ = True while line_index < len(lowerCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(lowerCAmelCase__ ): break UpperCAmelCase_ = lines[line_index] UpperCAmelCase_ = _should_continue(lowerCAmelCase__ , lowerCAmelCase__ ) and re.search(f"""^{indent}# End copy""" , lowerCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 UpperCAmelCase_ = lines[start_index:line_index] UpperCAmelCase_ = "".join(lowerCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies UpperCAmelCase_ = [line for line in theoretical_code.split("\n" ) if _re_copy_warning.search(lowerCAmelCase__ ) is None] UpperCAmelCase_ = "\n".join(lowerCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(lowerCAmelCase__ ) > 0: UpperCAmelCase_ = replace_pattern.replace("with" , "" ).split("," ) UpperCAmelCase_ = [_re_replace_pattern.search(lowerCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = pattern.groups() UpperCAmelCase_ = re.sub(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if option.strip() == "all-casing": UpperCAmelCase_ = re.sub(obja.lower() , obja.lower() , lowerCAmelCase__ ) UpperCAmelCase_ = re.sub(obja.upper() , obja.upper() , lowerCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line UpperCAmelCase_ = blackify(lines[start_index - 1] + theoretical_code ) UpperCAmelCase_ = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: UpperCAmelCase_ = lines[:start_index] + [theoretical_code] + lines[line_index:] UpperCAmelCase_ = start_index + 1 if overwrite and len(lowerCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f"""Detected changes, rewriting {filename}.""" ) with open(lowerCAmelCase__ , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(lowerCAmelCase__ ) return diffs def a__ ( lowerCAmelCase__ = False ): UpperCAmelCase_ = glob.glob(os.path.join(lowerCAmelCase__ , "**/*.py" ) , recursive=lowerCAmelCase__ ) UpperCAmelCase_ = [] for filename in all_files: UpperCAmelCase_ = is_copy_consistent(lowerCAmelCase__ , lowerCAmelCase__ ) diffs += [f"""- {filename}: copy does not match {d[0]} at line {d[1]}""" for d in new_diffs] if not overwrite and len(lowerCAmelCase__ ) > 0: UpperCAmelCase_ = "\n".join(lowerCAmelCase__ ) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them." ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") lowerCamelCase = parser.parse_args() check_copies(args.fix_and_overwrite)
241
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/config.json', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/config.json', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-german-cased': 'https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json', 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json' ), 'distilbert-base-uncased-finetuned-sst-2-english': ( 'https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json' ), } class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Optional[Any] = '''distilbert''' _snake_case : Dict = { '''hidden_size''': '''dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', } def __init__( self , _UpperCamelCase=3_0_5_2_2 , _UpperCamelCase=5_1_2 , _UpperCamelCase=False , _UpperCamelCase=6 , _UpperCamelCase=1_2 , _UpperCamelCase=7_6_8 , _UpperCamelCase=4 * 7_6_8 , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase="gelu" , _UpperCamelCase=0.02 , _UpperCamelCase=0.1 , _UpperCamelCase=0.2 , _UpperCamelCase=0 , **_UpperCamelCase , ) -> Any: UpperCAmelCase_ : Union[str, Any] = vocab_size UpperCAmelCase_ : Optional[int] = max_position_embeddings UpperCAmelCase_ : Tuple = sinusoidal_pos_embds UpperCAmelCase_ : Tuple = n_layers UpperCAmelCase_ : Optional[int] = n_heads UpperCAmelCase_ : Optional[int] = dim UpperCAmelCase_ : str = hidden_dim UpperCAmelCase_ : Tuple = dropout UpperCAmelCase_ : Optional[int] = attention_dropout UpperCAmelCase_ : Optional[Any] = activation UpperCAmelCase_ : int = initializer_range UpperCAmelCase_ : Tuple = qa_dropout UpperCAmelCase_ : List[str] = seq_classif_dropout super().__init__(**_UpperCamelCase , pad_token_id=_UpperCamelCase ) class lowerCamelCase (_snake_case ): '''simple docstring''' @property def __UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": UpperCAmelCase_ : List[str] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: UpperCAmelCase_ : Optional[int] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
29
'''simple docstring''' import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE :int = get_tests_dir('''fixtures/test_sentencepiece_no_bos.model''') @require_sentencepiece @require_tokenizers class A_ ( lowerCAmelCase_ , unittest.TestCase ): _lowerCamelCase : List[str] = PegasusTokenizer _lowerCamelCase : int = PegasusTokenizerFast _lowerCamelCase : Union[str, Any] = True _lowerCamelCase : List[str] = True def lowercase ( self : Optional[int] ): super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase = PegasusTokenizer(snake_case_ ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowercase ( self : Tuple ): return PegasusTokenizer.from_pretrained("google/pegasus-large" ) def lowercase ( self : Union[str, Any] , **snake_case_ : Union[str, Any] ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **snake_case_ ) def lowercase ( self : Tuple , snake_case_ : Any ): return ("This is a test", "This is a test") def lowercase ( self : Optional[int] ): _UpperCAmelCase = "</s>" _UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def lowercase ( self : Union[str, Any] ): _UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<pad>" ) self.assertEqual(vocab_keys[1] , "</s>" ) self.assertEqual(vocab_keys[-1] , "v" ) self.assertEqual(len(snake_case_ ) , 1_1_0_3 ) def lowercase ( self : Any ): self.assertEqual(self.get_tokenizer().vocab_size , 1_1_0_3 ) def lowercase ( self : List[Any] ): _UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) _UpperCAmelCase = self.tokenizer_class.from_pretrained(self.tmpdirname ) _UpperCAmelCase = ( "Let's see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important" " </s> <pad> <pad> <pad>" ) _UpperCAmelCase = rust_tokenizer([raw_input_str] , return_tensors=snake_case_ , add_special_tokens=snake_case_ ).input_ids[0] _UpperCAmelCase = py_tokenizer([raw_input_str] , return_tensors=snake_case_ , add_special_tokens=snake_case_ ).input_ids[0] self.assertListEqual(snake_case_ , snake_case_ ) def lowercase ( self : Tuple ): _UpperCAmelCase = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word _UpperCAmelCase = "<mask_1> To ensure a <mask_2> flow of bank resolutions." _UpperCAmelCase = [2, 4_1_3, 6_1_5, 1_1_4, 3, 1_9_7_1, 1_1_3, 1_6_7_9, 1_0_7_1_0, 1_0_7, 1] _UpperCAmelCase = tokenizer([raw_input_str] , return_tensors=snake_case_ ).input_ids[0] self.assertListEqual(snake_case_ , snake_case_ ) def lowercase ( self : Union[str, Any] ): _UpperCAmelCase = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 9_6_1_0_3 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 1_0_3 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 1_0_5 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 1_0_2_4 _UpperCAmelCase = "To ensure a smooth flow of bank resolutions." _UpperCAmelCase = [4_1_3, 6_1_5, 1_1_4, 2_2_9_1, 1_9_7_1, 1_1_3, 1_6_7_9, 1_0_7_1_0, 1_0_7, 1] _UpperCAmelCase = tokenizer([raw_input_str] , return_tensors=snake_case_ ).input_ids[0] self.assertListEqual(snake_case_ , snake_case_ ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def lowercase ( self : int ): _UpperCAmelCase = ["This is going to be way too long." * 1_5_0, "short example"] _UpperCAmelCase = ["not super long but more than 5 tokens", "tiny"] _UpperCAmelCase = self._large_tokenizer(snake_case_ , padding=snake_case_ , truncation=snake_case_ , return_tensors="pt" ) _UpperCAmelCase = self._large_tokenizer( text_target=snake_case_ , max_length=5 , padding=snake_case_ , truncation=snake_case_ , return_tensors="pt" ) assert batch.input_ids.shape == (2, 1_0_2_4) assert batch.attention_mask.shape == (2, 1_0_2_4) assert targets["input_ids"].shape == (2, 5) assert len(snake_case_ ) == 2 # input_ids, attention_mask. @slow def lowercase ( self : Dict ): # fmt: off _UpperCAmelCase = {"input_ids": [[3_8_9_7_9, 1_4_3, 1_8_4_8_5, 6_0_6, 1_3_0, 2_6_6_6_9, 8_7_6_8_6, 1_2_1, 5_4_1_8_9, 1_1_2_9, 1_1_1, 2_6_6_6_9, 8_7_6_8_6, 1_2_1, 9_1_1_4, 1_4_7_8_7, 1_2_1, 1_3_2_4_9, 1_5_8, 5_9_2, 9_5_6, 1_2_1, 1_4_6_2_1, 3_1_5_7_6, 1_4_3, 6_2_6_1_3, 1_0_8, 9_6_8_8, 9_3_0, 4_3_4_3_0, 1_1_5_6_2, 6_2_6_1_3, 3_0_4, 1_0_8, 1_1_4_4_3, 8_9_7, 1_0_8, 9_3_1_4, 1_7_4_1_5, 6_3_3_9_9, 1_0_8, 1_1_4_4_3, 7_6_1_4, 1_8_3_1_6, 1_1_8, 4_2_8_4, 7_1_4_8, 1_2_4_3_0, 1_4_3, 1_4_0_0, 2_5_7_0_3, 1_5_8, 1_1_1, 4_2_8_4, 7_1_4_8, 1_1_7_7_2, 1_4_3, 2_1_2_9_7, 1_0_6_4, 1_5_8, 1_2_2, 2_0_4, 3_5_0_6, 1_7_5_4, 1_1_3_3, 1_4_7_8_7, 1_5_8_1, 1_1_5, 3_3_2_2_4, 4_4_8_2, 1_1_1, 1_3_5_5, 1_1_0, 2_9_1_7_3, 3_1_7, 5_0_8_3_3, 1_0_8, 2_0_1_4_7, 9_4_6_6_5, 1_1_1, 7_7_1_9_8, 1_0_7, 1], [1_1_0, 6_2_6_1_3, 1_1_7, 6_3_8, 1_1_2, 1_1_3_3, 1_2_1, 2_0_0_9_8, 1_3_5_5, 7_9_0_5_0, 1_3_8_7_2, 1_3_5, 1_5_9_6, 5_3_5_4_1, 1_3_5_2, 1_4_1, 1_3_0_3_9, 5_5_4_2, 1_2_4, 3_0_2, 5_1_8, 1_1_1, 2_6_8, 2_9_5_6, 1_1_5, 1_4_9, 4_4_2_7, 1_0_7, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_3_9, 1_2_3_5, 2_7_9_9, 1_8_2_8_9, 1_7_7_8_0, 2_0_4, 1_0_9, 9_4_7_4, 1_2_9_6, 1_0_7, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name="google/bigbird-pegasus-large-arxiv" , revision="ba85d0851d708441f91440d509690f1ab6353415" , ) @require_sentencepiece @require_tokenizers class A_ ( lowerCAmelCase_ , unittest.TestCase ): _lowerCamelCase : List[str] = PegasusTokenizer _lowerCamelCase : List[Any] = PegasusTokenizerFast _lowerCamelCase : int = True _lowerCamelCase : Union[str, Any] = True def lowercase ( self : Any ): super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase = PegasusTokenizer(snake_case_ , offset=0 , mask_token_sent=snake_case_ , mask_token="[MASK]" ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowercase ( self : Tuple ): return PegasusTokenizer.from_pretrained("google/bigbird-pegasus-large-arxiv" ) def lowercase ( self : Optional[Any] , **snake_case_ : Dict ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **snake_case_ ) def lowercase ( self : Union[str, Any] , snake_case_ : str ): return ("This is a test", "This is a test") def lowercase ( self : List[str] ): _UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) _UpperCAmelCase = self.tokenizer_class.from_pretrained(self.tmpdirname ) _UpperCAmelCase = ( "Let's see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>" " <pad> <pad> <pad>" ) _UpperCAmelCase = rust_tokenizer([raw_input_str] , return_tensors=snake_case_ , add_special_tokens=snake_case_ ).input_ids[0] _UpperCAmelCase = py_tokenizer([raw_input_str] , return_tensors=snake_case_ , add_special_tokens=snake_case_ ).input_ids[0] self.assertListEqual(snake_case_ , snake_case_ ) @require_torch def lowercase ( self : Tuple ): _UpperCAmelCase = ["This is going to be way too long." * 1_0_0_0, "short example"] _UpperCAmelCase = ["not super long but more than 5 tokens", "tiny"] _UpperCAmelCase = self._large_tokenizer(snake_case_ , padding=snake_case_ , truncation=snake_case_ , return_tensors="pt" ) _UpperCAmelCase = self._large_tokenizer( text_target=snake_case_ , max_length=5 , padding=snake_case_ , truncation=snake_case_ , return_tensors="pt" ) assert batch.input_ids.shape == (2, 4_0_9_6) assert batch.attention_mask.shape == (2, 4_0_9_6) assert targets["input_ids"].shape == (2, 5) assert len(snake_case_ ) == 2 # input_ids, attention_mask. def lowercase ( self : Union[str, Any] ): _UpperCAmelCase = ( "This is an example string that is used to test the original TF implementation against the HF" " implementation" ) _UpperCAmelCase = self._large_tokenizer(snake_case_ ).input_ids self.assertListEqual( snake_case_ , [1_8_2, 1_1_7, 1_4_2, 5_8_7, 4_2_1_1, 1_2_0, 1_1_7, 2_6_3, 1_1_2, 8_0_4, 1_0_9, 8_5_6, 2_5_0_1_6, 3_1_3_7, 4_6_4, 1_0_9, 2_6_9_5_5, 3_1_3_7, 1] , )
22
0
from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__ ) -> List[str]: A_ = k_size // 2 A_ , A_ = mgrid[0 - center : k_size - center, 0 - center : k_size - center] A_ = 1 / (2 * pi * sigma) * exp(-(square(UpperCAmelCase__ ) + square(UpperCAmelCase__ )) / (2 * square(UpperCAmelCase__ )) ) return g def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) -> Optional[Any]: A_ , A_ = image.shape[0], image.shape[1] # dst image height and width A_ = height - k_size + 1 A_ = width - k_size + 1 # im2col, turn the k_size*k_size pixels into a row and np.vstack all rows A_ = zeros((dst_height * dst_width, k_size * k_size) ) A_ = 0 for i, j in product(range(UpperCAmelCase__ ), range(UpperCAmelCase__ ) ): A_ = ravel(image[i : i + k_size, j : j + k_size] ) A_ = window row += 1 # turn the kernel into shape(k*k, 1) A_ = gen_gaussian_kernel(UpperCAmelCase__, UpperCAmelCase__ ) A_ = ravel(UpperCAmelCase__ ) # reshape and get the dst image A_ = dot(UpperCAmelCase__, UpperCAmelCase__ ).reshape(UpperCAmelCase__, UpperCAmelCase__ ).astype(UpperCAmelCase__ ) return dst if __name__ == "__main__": # read original image __lowerCamelCase = imread(r'''../image_data/lena.jpg''') # turn image in gray scale value __lowerCamelCase = cvtColor(img, COLOR_BGR2GRAY) # get values with two different mask size __lowerCamelCase = gaussian_filter(gray, 3, sigma=1) __lowerCamelCase = gaussian_filter(gray, 5, sigma=0.8) # show result images imshow('''gaussian filter with 3x3 mask''', gaussianaxa) imshow('''gaussian filter with 5x5 mask''', gaussianaxa) waitKey()
358
'''simple docstring''' def UpperCAmelCase__ ( UpperCAmelCase__ = 1_00 ) -> int: A_ = n * (n + 1) * (2 * n + 1) / 6 A_ = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f"""{solution() = }""")
101
0
import requests A_ : str = '' # <-- Put your OpenWeatherMap appid here! A_ : List[str] = 'https://api.openweathermap.org/data/2.5/' def UpperCamelCase (lowercase_: str = "Chicago" , lowercase_: str = APPID ) -> dict: return requests.get(URL_BASE + """weather""" , params=locals() ).json() def UpperCamelCase (lowercase_: str = "Kolkata, India" , lowercase_: str = APPID ) -> dict: return requests.get(URL_BASE + """forecast""" , params=locals() ).json() def UpperCamelCase (lowercase_: float = 55.68 , lowercase_: float = 12.57 , lowercase_: str = APPID ) -> dict: return requests.get(URL_BASE + """onecall""" , params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: A_ : List[str] = input('Enter a location:').strip() if location: pprint(current_weather(location)) else: break
192
import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer A_ : List[str] = logging.get_logger(__name__) A_ : Tuple = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} A_ : List[Any] = { 'vocab_file': { 'junnyu/roformer_chinese_small': 'https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt', 'junnyu/roformer_chinese_base': 'https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt', 'junnyu/roformer_chinese_char_small': ( 'https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt' ), 'junnyu/roformer_chinese_char_base': ( 'https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt' ), 'junnyu/roformer_small_discriminator': ( 'https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt' ), 'junnyu/roformer_small_generator': ( 'https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt' ), } } A_ : Tuple = { 'junnyu/roformer_chinese_small': 1536, 'junnyu/roformer_chinese_base': 1536, 'junnyu/roformer_chinese_char_small': 512, 'junnyu/roformer_chinese_char_base': 512, 'junnyu/roformer_small_discriminator': 128, 'junnyu/roformer_small_generator': 128, } A_ : Union[str, Any] = { 'junnyu/roformer_chinese_small': {'do_lower_case': True}, 'junnyu/roformer_chinese_base': {'do_lower_case': True}, 'junnyu/roformer_chinese_char_small': {'do_lower_case': True}, 'junnyu/roformer_chinese_char_base': {'do_lower_case': True}, 'junnyu/roformer_small_discriminator': {'do_lower_case': True}, 'junnyu/roformer_small_generator': {'do_lower_case': True}, } class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: Dict = VOCAB_FILES_NAMES UpperCAmelCase__: Tuple = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__: List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__: List[Any] = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__: Optional[int] = RoFormerTokenizer def __init__( self , A__=None , A__=None , A__=True , A__="[UNK]" , A__="[SEP]" , A__="[PAD]" , A__="[CLS]" , A__="[MASK]" , A__=True , A__=None , **A__ , ): super().__init__( A__ , tokenizer_file=A__ , do_lower_case=A__ , unk_token=A__ , sep_token=A__ , pad_token=A__ , cls_token=A__ , mask_token=A__ , tokenize_chinese_chars=A__ , strip_accents=A__ , **A__ , ) A__ : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get("""lowercase""" , A__ ) != do_lower_case or pre_tok_state.get("""strip_accents""" , A__ ) != strip_accents ): A__ : List[Any] = getattr(A__ , pre_tok_state.pop("""type""" ) ) A__ : Optional[int] = do_lower_case A__ : int = strip_accents A__ : str = pre_tok_class(**A__ ) A__ : Any = do_lower_case def __getstate__( self ): A__ : int = self.__dict__.copy() A__ : Union[str, Any] = BertPreTokenizer() return state def __setstate__( self , A__ ): A__ : Union[str, Any] = d A__ : Union[str, Any] = self.__dict__["""_tokenizer"""].get_vocab() A__ : Dict = PreTokenizer.custom(JiebaPreTokenizer(A__ ) ) def __A ( self , A__ , A__=None ): A__ : Optional[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __A ( self , A__ , A__ = None ): A__ : Tuple = [self.sep_token_id] A__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __A ( self , A__ , A__ = None ): A__ : Any = self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ ) def __A ( self , A__ , A__=None , A__=None , A__=False , **A__ , ): A__ : str = BertPreTokenizer() return super().save_pretrained(A__ , A__ , A__ , A__ , **A__ )
192
1
"""simple docstring""" __A : Tuple = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} __A : List[Any] = ["a", "b", "c", "d", "e"] def lowercase ( _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' _UpperCAmelCase = start # add current to visited visited.append(_snake_case ) _UpperCAmelCase = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: _UpperCAmelCase = topological_sort(_snake_case , _snake_case , _snake_case ) # if all neighbors visited add current to sort sort.append(_snake_case ) # if all vertices haven't been visited select a new one to visit if len(_snake_case ) != len(_snake_case ): for vertice in vertices: if vertice not in visited: _UpperCAmelCase = topological_sort(_snake_case , _snake_case , _snake_case ) # return sort return sort if __name__ == "__main__": __A : str = topological_sort("a", [], []) print(sort)
369
"""simple docstring""" import logging import os from .state import PartialState class _a ( logging.LoggerAdapter): """simple docstring""" @staticmethod def lowercase__ ( __UpperCamelCase : Optional[Any] )->List[Any]: _UpperCAmelCase = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def lowercase__ ( self : List[Any] , __UpperCamelCase : List[Any] , __UpperCamelCase : Tuple , *__UpperCamelCase : Optional[Any] , **__UpperCamelCase : Union[str, Any] )->int: if PartialState._shared_state == {}: raise RuntimeError( '''You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.''' ) _UpperCAmelCase = kwargs.pop('''main_process_only''' , __UpperCamelCase ) _UpperCAmelCase = kwargs.pop('''in_order''' , __UpperCamelCase ) if self.isEnabledFor(__UpperCamelCase ): if self._should_log(__UpperCamelCase ): _UpperCAmelCase , _UpperCAmelCase = self.process(__UpperCamelCase , __UpperCamelCase ) self.logger.log(__UpperCamelCase , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ) elif in_order: _UpperCAmelCase = PartialState() for i in range(state.num_processes ): if i == state.process_index: _UpperCAmelCase , _UpperCAmelCase = self.process(__UpperCamelCase , __UpperCamelCase ) self.logger.log(__UpperCamelCase , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ) state.wait_for_everyone() def lowercase ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str = None ): '''simple docstring''' if log_level is None: _UpperCAmelCase = os.environ.get('''ACCELERATE_LOG_LEVEL''' , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = logging.getLogger(_SCREAMING_SNAKE_CASE ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(_SCREAMING_SNAKE_CASE , {} )
326
0
"""simple docstring""" import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline _a = { """n_samples""": 64, """horizon""": 32, """num_inference_steps""": 20, """n_guide_steps""": 2, # can set to 0 for faster sampling, does not use value network """scale_grad_by_std""": True, """scale""": 0.1, """eta""": 0.0, """t_grad_cutoff""": 2, """device""": """cpu""", } if __name__ == "__main__": _a = """hopper-medium-v2""" _a = gym.make(env_name) _a = ValueGuidedRLPipeline.from_pretrained( """bglick13/hopper-medium-v2-value-function-hor32""", env=env, ) env.seed(0) _a = env.reset() _a = 0 _a = 0 _a = 1000 _a = [obs.copy()] try: for t in tqdm.tqdm(range(T)): # call the policy _a = pipeline(obs, planning_horizon=32) # execute action in environment _a , _a , _a , _a = env.step(denorm_actions) _a = env.get_normalized_score(total_reward) # update return total_reward += reward total_score += score print( F"""Step: {t}, Reward: {reward}, Total Reward: {total_reward}, Score: {score}, Total Score:""" F""" {total_score}""" ) # save observations for rendering rollout.append(next_observation.copy()) _a = next_observation except KeyboardInterrupt: pass print(F"""Total reward: {total_reward}""")
194
"""simple docstring""" def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" return credit_card_number.startswith(('''34''', '''35''', '''37''', '''4''', '''5''', '''6''') ) def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" _UpperCamelCase = credit_card_number _UpperCamelCase = 0 _UpperCamelCase = len(__snake_case ) - 2 for i in range(__snake_case, -1, -2 ): # double the value of every second digit _UpperCamelCase = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 _UpperCamelCase = cc_number[:i] + str(__snake_case ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(__snake_case ) - 1, -1, -2 ): total += int(cc_number[i] ) return total % 10 == 0 def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" _UpperCamelCase = F'''{credit_card_number} is an invalid credit card number because''' if not credit_card_number.isdigit(): print(F'''{error_message} it has nonnumerical characters.''' ) return False if not 13 <= len(__snake_case ) <= 16: print(F'''{error_message} of its length.''' ) return False if not validate_initial_digits(__snake_case ): print(F'''{error_message} of its first two digits.''' ) return False if not luhn_validation(__snake_case ): print(F'''{error_message} it fails the Luhn check.''' ) return False print(F'''{credit_card_number} is a valid credit card number.''' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number("""4111111111111111""") validate_credit_card_number("""32323""")
194
1
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class UpperCAmelCase_ ( unittest.TestCase): def __init__( self, __a, __a=7, __a=3, __a=30, __a=400, __a=True, __a=None, __a=True, __a=1 / 255, __a=True, __a=[0.5, 0.5, 0.5], __a=[0.5, 0.5, 0.5], __a=True, ): '''simple docstring''' _lowerCAmelCase : Tuple = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} _lowerCAmelCase : Dict = parent _lowerCAmelCase : Dict = batch_size _lowerCAmelCase : List[str] = num_channels _lowerCAmelCase : Optional[Any] = min_resolution _lowerCAmelCase : Tuple = max_resolution _lowerCAmelCase : Union[str, Any] = do_resize _lowerCAmelCase : List[str] = size _lowerCAmelCase : Optional[int] = do_rescale _lowerCAmelCase : Tuple = rescale_factor _lowerCAmelCase : Union[str, Any] = do_normalize _lowerCAmelCase : List[Any] = image_mean _lowerCAmelCase : Optional[Any] = image_std _lowerCAmelCase : str = do_pad def snake_case__ ( self): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def snake_case__ ( self, __a, __a=False): '''simple docstring''' if not batched: _lowerCAmelCase : List[Any] = image_inputs[0] if isinstance(__UpperCAmelCase, Image.Image): _lowerCAmelCase , _lowerCAmelCase : int = image.size else: _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = image.shape[1], image.shape[2] if w < h: _lowerCAmelCase : Any = int(self.size["shortest_edge"] * h / w) _lowerCAmelCase : Any = self.size["shortest_edge"] elif w > h: _lowerCAmelCase : Dict = self.size["shortest_edge"] _lowerCAmelCase : str = int(self.size["shortest_edge"] * w / h) else: _lowerCAmelCase : Dict = self.size["shortest_edge"] _lowerCAmelCase : Union[str, Any] = self.size["shortest_edge"] else: _lowerCAmelCase : List[str] = [] for image in image_inputs: _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) _lowerCAmelCase : int = max(__UpperCAmelCase, key=lambda __a: item[0])[0] _lowerCAmelCase : Optional[int] = max(__UpperCAmelCase, key=lambda __a: item[1])[1] return expected_height, expected_width @require_torch @require_vision class UpperCAmelCase_ ( a , unittest.TestCase): lowerCamelCase__ = DetrImageProcessor if is_vision_available() else None def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : List[Any] = DetrImageProcessingTester(self) @property def snake_case__ ( self): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : List[Any] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__UpperCAmelCase, "image_mean")) self.assertTrue(hasattr(__UpperCAmelCase, "image_std")) self.assertTrue(hasattr(__UpperCAmelCase, "do_normalize")) self.assertTrue(hasattr(__UpperCAmelCase, "do_rescale")) self.assertTrue(hasattr(__UpperCAmelCase, "rescale_factor")) self.assertTrue(hasattr(__UpperCAmelCase, "do_resize")) self.assertTrue(hasattr(__UpperCAmelCase, "size")) self.assertTrue(hasattr(__UpperCAmelCase, "do_pad")) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Any = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size, {"shortest_edge": 18, "longest_edge": 1333}) self.assertEqual(image_processor.do_pad, __UpperCAmelCase) _lowerCAmelCase : str = self.image_processing_class.from_dict( self.image_processor_dict, size=42, max_size=84, pad_and_return_pixel_mask=__UpperCAmelCase) self.assertEqual(image_processor.size, {"shortest_edge": 42, "longest_edge": 84}) self.assertEqual(image_processor.do_pad, __UpperCAmelCase) def snake_case__ ( self): '''simple docstring''' pass def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict) # create random PIL images _lowerCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester, equal_resolution=__UpperCAmelCase) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase, Image.Image) # Test not batched input _lowerCAmelCase : Dict = image_processing(image_inputs[0], return_tensors="pt").pixel_values _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.image_processor_tester.get_expected_values(__UpperCAmelCase) self.assertEqual( encoded_images.shape, (1, self.image_processor_tester.num_channels, expected_height, expected_width), ) # Test batched _lowerCAmelCase , _lowerCAmelCase : int = self.image_processor_tester.get_expected_values(__UpperCAmelCase, batched=__UpperCAmelCase) _lowerCAmelCase : List[Any] = image_processing(__UpperCAmelCase, return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ), ) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : List[str] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _lowerCAmelCase : Optional[Any] = prepare_image_inputs(self.image_processor_tester, equal_resolution=__UpperCAmelCase, numpify=__UpperCAmelCase) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase, np.ndarray) # Test not batched input _lowerCAmelCase : str = image_processing(image_inputs[0], return_tensors="pt").pixel_values _lowerCAmelCase , _lowerCAmelCase : int = self.image_processor_tester.get_expected_values(__UpperCAmelCase) self.assertEqual( encoded_images.shape, (1, self.image_processor_tester.num_channels, expected_height, expected_width), ) # Test batched _lowerCAmelCase : Any = image_processing(__UpperCAmelCase, return_tensors="pt").pixel_values _lowerCAmelCase , _lowerCAmelCase : Tuple = self.image_processor_tester.get_expected_values(__UpperCAmelCase, batched=__UpperCAmelCase) self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ), ) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _lowerCAmelCase : int = prepare_image_inputs(self.image_processor_tester, equal_resolution=__UpperCAmelCase, torchify=__UpperCAmelCase) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase, torch.Tensor) # Test not batched input _lowerCAmelCase : Tuple = image_processing(image_inputs[0], return_tensors="pt").pixel_values _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = self.image_processor_tester.get_expected_values(__UpperCAmelCase) self.assertEqual( encoded_images.shape, (1, self.image_processor_tester.num_channels, expected_height, expected_width), ) # Test batched _lowerCAmelCase : List[Any] = image_processing(__UpperCAmelCase, return_tensors="pt").pixel_values _lowerCAmelCase , _lowerCAmelCase : Union[str, Any] = self.image_processor_tester.get_expected_values(__UpperCAmelCase, batched=__UpperCAmelCase) self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ), ) @slow def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt", "r") as f: _lowerCAmelCase : Union[str, Any] = json.loads(f.read()) _lowerCAmelCase : str = {"image_id": 3_9769, "annotations": target} # encode them _lowerCAmelCase : Optional[int] = DetrImageProcessor.from_pretrained("facebook/detr-resnet-50") _lowerCAmelCase : Any = image_processing(images=__UpperCAmelCase, annotations=__UpperCAmelCase, return_tensors="pt") # verify pixel values _lowerCAmelCase : int = torch.Size([1, 3, 800, 1066]) self.assertEqual(encoding["pixel_values"].shape, __UpperCAmelCase) _lowerCAmelCase : List[str] = torch.tensor([0.2_796, 0.3_138, 0.3_481]) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3], __UpperCAmelCase, atol=1E-4)) # verify area _lowerCAmelCase : Optional[int] = torch.tensor([5_887.9_600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438]) self.assertTrue(torch.allclose(encoding["labels"][0]["area"], __UpperCAmelCase)) # verify boxes _lowerCAmelCase : List[str] = torch.Size([6, 4]) self.assertEqual(encoding["labels"][0]["boxes"].shape, __UpperCAmelCase) _lowerCAmelCase : Union[str, Any] = torch.tensor([0.5_503, 0.2_765, 0.0_604, 0.2_215]) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0], __UpperCAmelCase, atol=1E-3)) # verify image_id _lowerCAmelCase : Optional[int] = torch.tensor([3_9769]) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"], __UpperCAmelCase)) # verify is_crowd _lowerCAmelCase : Optional[Any] = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"], __UpperCAmelCase)) # verify class_labels _lowerCAmelCase : str = torch.tensor([75, 75, 63, 65, 17, 17]) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"], __UpperCAmelCase)) # verify orig_size _lowerCAmelCase : Dict = torch.tensor([480, 640]) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"], __UpperCAmelCase)) # verify size _lowerCAmelCase : Union[str, Any] = torch.tensor([800, 1066]) self.assertTrue(torch.allclose(encoding["labels"][0]["size"], __UpperCAmelCase)) @slow def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt", "r") as f: _lowerCAmelCase : str = json.loads(f.read()) _lowerCAmelCase : Union[str, Any] = {"file_name": "000000039769.png", "image_id": 3_9769, "segments_info": target} _lowerCAmelCase : List[Any] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic") # encode them _lowerCAmelCase : int = DetrImageProcessor.from_pretrained("facebook/detr-resnet-50-panoptic") _lowerCAmelCase : str = image_processing(images=__UpperCAmelCase, annotations=__UpperCAmelCase, masks_path=__UpperCAmelCase, return_tensors="pt") # verify pixel values _lowerCAmelCase : Any = torch.Size([1, 3, 800, 1066]) self.assertEqual(encoding["pixel_values"].shape, __UpperCAmelCase) _lowerCAmelCase : Any = torch.tensor([0.2_796, 0.3_138, 0.3_481]) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3], __UpperCAmelCase, atol=1E-4)) # verify area _lowerCAmelCase : Optional[Any] = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5_879.6_562, 7_634.1_147]) self.assertTrue(torch.allclose(encoding["labels"][0]["area"], __UpperCAmelCase)) # verify boxes _lowerCAmelCase : Optional[int] = torch.Size([6, 4]) self.assertEqual(encoding["labels"][0]["boxes"].shape, __UpperCAmelCase) _lowerCAmelCase : Optional[int] = torch.tensor([0.2_625, 0.5_437, 0.4_688, 0.8_625]) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0], __UpperCAmelCase, atol=1E-3)) # verify image_id _lowerCAmelCase : List[Any] = torch.tensor([3_9769]) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"], __UpperCAmelCase)) # verify is_crowd _lowerCAmelCase : Any = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"], __UpperCAmelCase)) # verify class_labels _lowerCAmelCase : Optional[Any] = torch.tensor([17, 17, 63, 75, 75, 93]) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"], __UpperCAmelCase)) # verify masks _lowerCAmelCase : Optional[int] = 82_2873 self.assertEqual(encoding["labels"][0]["masks"].sum().item(), __UpperCAmelCase) # verify orig_size _lowerCAmelCase : int = torch.tensor([480, 640]) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"], __UpperCAmelCase)) # verify size _lowerCAmelCase : str = torch.tensor([800, 1066]) self.assertTrue(torch.allclose(encoding["labels"][0]["size"], __UpperCAmelCase))
354
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class UpperCAmelCase_ ( a): lowerCamelCase__ = 42 lowerCamelCase__ = 42 class UpperCAmelCase_ ( a , a): lowerCamelCase__ = 1 @register_to_config def __init__( self, __a = 2000, __a = 0.15, __a = 0.01, __a = 1_348.0, __a = 1E-5, __a = 1, ): '''simple docstring''' _lowerCAmelCase : Dict = sigma_max # setable values _lowerCAmelCase : str = None self.set_sigmas(__a, __a, __a, __a) def snake_case__ ( self, __a, __a = None): '''simple docstring''' return sample def snake_case__ ( self, __a, __a = None, __a = None): '''simple docstring''' _lowerCAmelCase : int = sampling_eps if sampling_eps is not None else self.config.sampling_eps _lowerCAmelCase : Dict = torch.linspace(1, __a, __a, device=__a) def snake_case__ ( self, __a, __a = None, __a = None, __a = None): '''simple docstring''' _lowerCAmelCase : List[str] = sigma_min if sigma_min is not None else self.config.sigma_min _lowerCAmelCase : Tuple = sigma_max if sigma_max is not None else self.config.sigma_max _lowerCAmelCase : str = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(__a, __a) _lowerCAmelCase : int = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) _lowerCAmelCase : Any = torch.exp(torch.linspace(math.log(__a), math.log(__a), __a)) _lowerCAmelCase : int = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps]) def snake_case__ ( self, __a, __a): '''simple docstring''' return torch.where( timesteps == 0, torch.zeros_like(t.to(timesteps.device)), self.discrete_sigmas[timesteps - 1].to(timesteps.device), ) def snake_case__ ( self, __a, __a, __a, __a = None, __a = True, ): '''simple docstring''' if self.timesteps is None: raise ValueError( "`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler") _lowerCAmelCase : Dict = timestep * torch.ones( sample.shape[0], device=sample.device) # torch.repeat_interleave(timestep, sample.shape[0]) _lowerCAmelCase : Dict = (timestep * (len(self.timesteps) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda _lowerCAmelCase : Union[str, Any] = timesteps.to(self.discrete_sigmas.device) _lowerCAmelCase : Any = self.discrete_sigmas[timesteps].to(sample.device) _lowerCAmelCase : List[Any] = self.get_adjacent_sigma(__a, __a).to(sample.device) _lowerCAmelCase : List[str] = torch.zeros_like(__a) _lowerCAmelCase : Union[str, Any] = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods _lowerCAmelCase : Union[str, Any] = diffusion.flatten() while len(diffusion.shape) < len(sample.shape): _lowerCAmelCase : Optional[int] = diffusion.unsqueeze(-1) _lowerCAmelCase : Dict = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of _lowerCAmelCase : Optional[Any] = randn_tensor( sample.shape, layout=sample.layout, generator=__a, device=sample.device, dtype=sample.dtype) _lowerCAmelCase : int = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? _lowerCAmelCase : Tuple = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=__a, prev_sample_mean=__a) def snake_case__ ( self, __a, __a, __a = None, __a = True, ): '''simple docstring''' if self.timesteps is None: raise ValueError( "`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler") # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction _lowerCAmelCase : Union[str, Any] = randn_tensor(sample.shape, layout=sample.layout, generator=__a).to(sample.device) # compute step size from the model_output, the noise, and the snr _lowerCAmelCase : Any = torch.norm(model_output.reshape(model_output.shape[0], -1), dim=-1).mean() _lowerCAmelCase : Dict = torch.norm(noise.reshape(noise.shape[0], -1), dim=-1).mean() _lowerCAmelCase : Optional[Any] = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 _lowerCAmelCase : Dict = step_size * torch.ones(sample.shape[0]).to(sample.device) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term _lowerCAmelCase : List[Any] = step_size.flatten() while len(step_size.shape) < len(sample.shape): _lowerCAmelCase : int = step_size.unsqueeze(-1) _lowerCAmelCase : List[Any] = sample + step_size * model_output _lowerCAmelCase : Tuple = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__a) def snake_case__ ( self, __a, __a, __a, ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = timesteps.to(original_samples.device) _lowerCAmelCase : Union[str, Any] = self.discrete_sigmas.to(original_samples.device)[timesteps] _lowerCAmelCase : Any = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(__a) * sigmas[:, None, None, None] ) _lowerCAmelCase : int = noise + original_samples return noisy_samples def __len__( self): '''simple docstring''' return self.config.num_train_timesteps
300
0
def lowerCAmelCase ( _lowerCAmelCase : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : Union[str, Any] ): """simple docstring""" UpperCAmelCase__ = [False] * len(_lowerCAmelCase ) UpperCAmelCase__ = [] queue.append(_lowerCAmelCase ) UpperCAmelCase__ = True while queue: UpperCAmelCase__ = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(_lowerCAmelCase ) UpperCAmelCase__ = True UpperCAmelCase__ = u return visited[t] def lowerCAmelCase ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ): """simple docstring""" UpperCAmelCase__ = [-1] * (len(_lowerCAmelCase )) UpperCAmelCase__ = 0 while bfs(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ = float("Inf" ) UpperCAmelCase__ = sink while s != source: # Find the minimum value in select path UpperCAmelCase__ = min(_lowerCAmelCase , graph[parent[s]][s] ) UpperCAmelCase__ = parent[s] max_flow += path_flow UpperCAmelCase__ = sink while v != source: UpperCAmelCase__ = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow UpperCAmelCase__ = parent[v] return max_flow _lowerCAmelCase : Union[str, Any] = [ [0, 1_6, 1_3, 0, 0, 0], [0, 0, 1_0, 1_2, 0, 0], [0, 4, 0, 0, 1_4, 0], [0, 0, 9, 0, 0, 2_0], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] _lowerCAmelCase, _lowerCAmelCase : List[Any] = 0, 5 print(ford_fulkerson(graph, source, sink))
169
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging lowerCAmelCase = logging.get_logger(__name__) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" try: with open(SCREAMING_SNAKE_CASE , '''rb''' ) as flax_state_f: lowercase__ = from_bytes(SCREAMING_SNAKE_CASE , flax_state_f.read() ) except UnpicklingError as e: try: with open(SCREAMING_SNAKE_CASE ) as f: if f.read().startswith('''version''' ): raise OSError( '''You seem to have cloned a repository without having git-lfs installed. Please''' ''' install git-lfs and run `git lfs install` followed by `git lfs pull` in the''' ''' folder you cloned.''' ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(f'Unable to convert {model_file} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( '''Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see''' ''' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation''' ''' instructions.''' ) raise # check if we have bf16 weights lowercase__ = flatten_dict(jax.tree_util.tree_map(lambda SCREAMING_SNAKE_CASE : x.dtype == jnp.bfloataa , SCREAMING_SNAKE_CASE ) ).values() if any(SCREAMING_SNAKE_CASE ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( '''Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ''' '''before loading those in PyTorch model.''' ) lowercase__ = jax.tree_util.tree_map( lambda SCREAMING_SNAKE_CASE : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , SCREAMING_SNAKE_CASE ) lowercase__ = '''''' lowercase__ = flatten_dict(SCREAMING_SNAKE_CASE , sep='''.''' ) lowercase__ = pt_model.state_dict() # keep track of unexpected & missing keys lowercase__ = [] lowercase__ = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): lowercase__ = flax_key_tuple.split('''.''' ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] lowercase__ = jnp.transpose(SCREAMING_SNAKE_CASE , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] lowercase__ = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(SCREAMING_SNAKE_CASE ): lowercase__ = ( flax_key_tuple_string.replace('''_0''' , '''.0''' ) .replace('''_1''' , '''.1''' ) .replace('''_2''' , '''.2''' ) .replace('''_3''' , '''.3''' ) .replace('''_4''' , '''.4''' ) .replace('''_5''' , '''.5''' ) .replace('''_6''' , '''.6''' ) .replace('''_7''' , '''.7''' ) .replace('''_8''' , '''.8''' ) .replace('''_9''' , '''.9''' ) ) lowercase__ = '''.'''.join(SCREAMING_SNAKE_CASE ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( f'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' f'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict lowercase__ = np.asarray(SCREAMING_SNAKE_CASE ) if not isinstance(SCREAMING_SNAKE_CASE , np.ndarray ) else flax_tensor lowercase__ = torch.from_numpy(SCREAMING_SNAKE_CASE ) # remove from missing keys missing_keys.remove(SCREAMING_SNAKE_CASE ) else: # weight is not expected by PyTorch model unexpected_keys.append(SCREAMING_SNAKE_CASE ) pt_model.load_state_dict(SCREAMING_SNAKE_CASE ) # re-transform missing_keys to list lowercase__ = list(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0: logger.warning( '''Some weights of the Flax model were not used when initializing the PyTorch model''' f' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' f' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' ''' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This''' f' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' ''' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a''' ''' FlaxBertForSequenceClassification model).''' ) if len(SCREAMING_SNAKE_CASE ) > 0: logger.warning( f'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' f' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' ''' use it for predictions and inference.''' ) return pt_model
110
0
from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 lowercase_ = { # 1536-bit 5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 2048-bit 1_4: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AACAA68FFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 3072-bit 1_5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 4096-bit 1_6: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199' + 'FFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 6144-bit 1_7: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08' + '8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B' + '302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9' + 'A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6' + '49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8' + 'FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C' + '180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718' + '3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D' + '04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D' + 'B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226' + '1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC' + 'E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26' + '99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB' + '04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2' + '233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127' + 'D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406' + 'AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918' + 'DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151' + '2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03' + 'F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F' + 'BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B' + 'B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632' + '387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E' + '6DCC4024FFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 8192-bit 1_8: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD' + 'F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831' + '179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B' + 'DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF' + '5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6' + 'D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3' + '23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328' + '06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C' + 'DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE' + '12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4' + '38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300' + '741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568' + '3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9' + '22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B' + '4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A' + '062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36' + '4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1' + 'B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92' + '4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47' + '9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71' + '60C980DD98EDD3DFFFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, } class A_ : '''simple docstring''' def __init__( self: List[Any] , a: int = 14 ): if group not in primes: raise ValueError('Unsupported Group' ) __lowerCamelCase : Dict = primes[group]['prime'] __lowerCamelCase : Dict = primes[group]['generator'] __lowerCamelCase : str = int(hexlify(urandom(32 ) ) , base=16 ) def _snake_case ( self: Tuple ): return hex(self.__private_key )[2:] def _snake_case ( self: Optional[int] ): __lowerCamelCase : Dict = pow(self.generator , self.__private_key , self.prime ) return hex(a )[2:] def _snake_case ( self: Optional[Any] , a: int ): # check if the other public key is valid based on NIST SP800-56 return ( 2 <= key <= self.prime - 2 and pow(a , (self.prime - 1) // 2 , self.prime ) == 1 ) def _snake_case ( self: int , a: str ): __lowerCamelCase : List[Any] = int(a , base=16 ) if not self.is_valid_public_key(a ): raise ValueError('Invalid public key' ) __lowerCamelCase : int = pow(a , self.__private_key , self.prime ) return shaaaa(str(a ).encode() ).hexdigest() @staticmethod def _snake_case ( a: int , a: int ): # check if the other public key is valid based on NIST SP800-56 return ( 2 <= remote_public_key_str <= prime - 2 and pow(a , (prime - 1) // 2 , a ) == 1 ) @staticmethod def _snake_case ( a: str , a: str , a: int = 14 ): __lowerCamelCase : Tuple = int(a , base=16 ) __lowerCamelCase : Tuple = int(a , base=16 ) __lowerCamelCase : Union[str, Any] = primes[group]['prime'] if not DiffieHellman.is_valid_public_key_static(a , a ): raise ValueError('Invalid public key' ) __lowerCamelCase : Optional[Any] = pow(a , a , a ) return shaaaa(str(a ).encode() ).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
365
from math import pow def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ): if current_sum == needed_sum: # If the sum of the powers is equal to needed_sum, then we have a solution. solutions_count += 1 return current_sum, solutions_count __lowerCamelCase : Optional[Any] = int(pow(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) if current_sum + i_to_n <= needed_sum: # If the sum of the powers is less than needed_sum, then continue adding powers. current_sum += i_to_n __lowerCamelCase , __lowerCamelCase : Optional[Any] = backtrack( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , current_number + 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) current_sum -= i_to_n if i_to_n < needed_sum: # If the power of i is less than needed_sum, then try with the next power. __lowerCamelCase , __lowerCamelCase : Dict = backtrack( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , current_number + 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return current_sum, solutions_count def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): if not (1 <= needed_sum <= 1_000 and 2 <= power <= 10): raise ValueError( 'Invalid input\n' 'needed_sum must be between 1 and 1000, power between 2 and 10.' ) return backtrack(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 , 0 , 0 )[1] # Return the solutions_count if __name__ == "__main__": import doctest doctest.testmod()
194
0
'''simple docstring''' import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder __lowerCAmelCase = '__DUMMY_TRANSFORMERS_USER__' __lowerCAmelCase = 'Dummy User' __lowerCAmelCase = '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' __lowerCAmelCase = 'https://hub-ci.huggingface.co' __lowerCAmelCase = CI_HUB_ENDPOINT + '/datasets/{repo_id}/resolve/{revision}/{path}' __lowerCAmelCase = CI_HUB_ENDPOINT + '/{repo_id}/resolve/{revision}/{filename}' __lowerCAmelCase = Path('~/.huggingface/hub_ci_token').expanduser() @pytest.fixture def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): monkeypatch.setattr( """huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE""" , _SCREAMING_SNAKE_CASE ) @pytest.fixture def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): monkeypatch.setattr("""datasets.config.HF_ENDPOINT""" , _SCREAMING_SNAKE_CASE ) monkeypatch.setattr("""datasets.config.HUB_DATASETS_URL""" , _SCREAMING_SNAKE_CASE ) @pytest.fixture def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): monkeypatch.setattr("""huggingface_hub.hf_api.HfFolder.path_token""" , _SCREAMING_SNAKE_CASE ) @pytest.fixture def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): HfFolder.save_token(_SCREAMING_SNAKE_CASE ) yield HfFolder.delete_token() @pytest.fixture(scope="""session""" ) def __SCREAMING_SNAKE_CASE ( ): return HfApi(endpoint=_SCREAMING_SNAKE_CASE ) @pytest.fixture(scope="""session""" ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = HfFolder.get_token() HfFolder.save_token(_SCREAMING_SNAKE_CASE ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(_SCREAMING_SNAKE_CASE ) @pytest.fixture def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): def _cleanup_repo(_SCREAMING_SNAKE_CASE ): hf_api.delete_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" ) return _cleanup_repo @pytest.fixture def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): @contextmanager def _temporary_repo(_SCREAMING_SNAKE_CASE ): try: yield repo_id finally: cleanup_repo(_SCREAMING_SNAKE_CASE ) return _temporary_repo @pytest.fixture(scope="""session""" ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""repo_txt_data-{int(time.time() * 10E3 )}""" _snake_case = f"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" , private=_SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=_SCREAMING_SNAKE_CASE , path_or_fileobj=str(_SCREAMING_SNAKE_CASE ) , path_in_repo="""data/text_data.txt""" , repo_id=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="""session""" ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""repo_zipped_txt_data-{int(time.time() * 10E3 )}""" _snake_case = f"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" , private=_SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=_SCREAMING_SNAKE_CASE , path_or_fileobj=str(_SCREAMING_SNAKE_CASE ) , path_in_repo="""data.zip""" , repo_id=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="""session""" ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""repo_zipped_img_data-{int(time.time() * 10E3 )}""" _snake_case = f"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" , private=_SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=_SCREAMING_SNAKE_CASE , path_or_fileobj=str(_SCREAMING_SNAKE_CASE ) , path_in_repo="""data.zip""" , repo_id=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return hf_private_dataset_repo_zipped_img_data_
341
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union __lowerCAmelCase = TypeVar('T') __lowerCAmelCase = Union[List[T], Tuple[T, ...]] __lowerCAmelCase = Union[T, List[T], Dict[str, T]] __lowerCAmelCase = Union[str, bytes, os.PathLike]
341
1
import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from tensorflow.python.eager import context from tensorflow.python.framework import ops from transformers import GradientAccumulator, create_optimizer @require_tf class lowerCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Dict: self.assertEqual(len(SCREAMING_SNAKE_CASE_ ), len(SCREAMING_SNAKE_CASE_ ) ) for a, b in zip(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): self.assertAlmostEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, delta=SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self ) -> Dict: UpperCamelCase : Any = GradientAccumulator() accumulator([tf.constant([1.0, 2.0] )] ) accumulator([tf.constant([-2.0, 1.0] )] ) accumulator([tf.constant([-1.0, 2.0] )] ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): accumulator([tf.constant([1.0, 1.0] ), tf.constant([2.0, 2.0] )] ) self.assertEqual(accumulator.step, 3 ) self.assertEqual(len(accumulator.gradients ), 1 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist(), [-2.0, 5.0], tol=1e-2 ) accumulator.reset() self.assertEqual(accumulator.step, 0 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist(), [0.0, 0.0], tol=1e-2 ) def snake_case_ ( self ) -> Optional[Any]: UpperCamelCase : Optional[int] = None ops.enable_eager_execution_internal() UpperCamelCase : str = tf.config.list_physical_devices('CPU' ) if len(SCREAMING_SNAKE_CASE_ ) == 1: tf.config.set_logical_device_configuration( physical_devices[0], [tf.config.LogicalDeviceConfiguration(), tf.config.LogicalDeviceConfiguration()] ) UpperCamelCase : Optional[int] = tf.config.list_logical_devices(device_type='CPU' ) UpperCamelCase : Dict = tf.distribute.MirroredStrategy(devices=devices[:2] ) with strategy.scope(): UpperCamelCase : Any = GradientAccumulator() UpperCamelCase : Any = tf.Variable([4.0, 3.0] ) UpperCamelCase , UpperCamelCase : Optional[Any] = create_optimizer(5e-5, 10, 5 ) UpperCamelCase : Optional[int] = tf.Variable([0.0, 0.0], trainable=SCREAMING_SNAKE_CASE_ ) def accumulate_on_replica(SCREAMING_SNAKE_CASE_ ): accumulator([gradient] ) def apply_on_replica(): optimizer.apply_gradients(list(zip(accumulator.gradients, [variable] ) ) ) @tf.function def accumulate(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): with strategy.scope(): UpperCamelCase : List[str] = strategy.experimental_local_results(SCREAMING_SNAKE_CASE_ ) local_variables[0].assign(SCREAMING_SNAKE_CASE_ ) local_variables[1].assign(SCREAMING_SNAKE_CASE_ ) strategy.run(SCREAMING_SNAKE_CASE_, args=(gradient_placeholder,) ) @tf.function def apply_grad(): with strategy.scope(): strategy.run(SCREAMING_SNAKE_CASE_ ) def _check_local_values(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): UpperCamelCase : str = strategy.experimental_local_results(accumulator._gradients[0] ) self.assertListAlmostEqual(values[0].value(), SCREAMING_SNAKE_CASE_, tol=1e-2 ) self.assertListAlmostEqual(values[1].value(), SCREAMING_SNAKE_CASE_, tol=1e-2 ) accumulate([1.0, 2.0], [-1.0, 1.0] ) accumulate([3.0, -1.0], [-1.0, -1.0] ) accumulate([-2.0, 2.0], [3.0, -2.0] ) self.assertEqual(accumulator.step, 3 ) _check_local_values([2.0, 3.0], [1.0, -2.0] ) apply_grad() self.assertListAlmostEqual(variable.value(), [4.0, 3.0], tol=1e-2 ) accumulator.reset() self.assertEqual(accumulator.step, 0 ) _check_local_values([0.0, 0.0], [0.0, 0.0] )
103
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class lowerCAmelCase_ ( a__ ): def __init__( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = False, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = "arrow", **SCREAMING_SNAKE_CASE_, ) -> int: super().__init__( split=SCREAMING_SNAKE_CASE_, features=SCREAMING_SNAKE_CASE_, cache_dir=SCREAMING_SNAKE_CASE_, keep_in_memory=SCREAMING_SNAKE_CASE_, streaming=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_, ) UpperCamelCase : Any = load_from_cache_file UpperCamelCase : Any = file_format UpperCamelCase : str = Spark( df=SCREAMING_SNAKE_CASE_, features=SCREAMING_SNAKE_CASE_, cache_dir=SCREAMING_SNAKE_CASE_, working_dir=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_, ) def snake_case_ ( self ) -> Tuple: if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) UpperCamelCase : Dict = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=SCREAMING_SNAKE_CASE_, file_format=self._file_format, ) return self.builder.as_dataset(split=self.split )
103
1
'''simple docstring''' import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def _A (lowerCAmelCase__ :List[str] , lowerCAmelCase__ :List[str] ) -> List[str]: '''simple docstring''' if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer _a = flax_key_tuple[:-1] + ('''weight''',) _a = torch.permute(__SCREAMING_SNAKE_CASE , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(__SCREAMING_SNAKE_CASE ): # linear layer _a = flax_key_tuple[:-1] + ('''weight''',) _a = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: _a = flax_key_tuple[:-1] + ('''weight''',) return flax_key_tuple, flax_tensor def _A (lowerCAmelCase__ :List[str] , lowerCAmelCase__ :Any , lowerCAmelCase__ :Optional[Any] ) -> Optional[Any]: '''simple docstring''' if "metadata" in layer: _a = layer.split('metadata' ) _a = ''''''.join(split_layer[0] )[:-1] _a = [tuple(('metadata' + split_layer[1]).split('/' ) )] elif "kvstore" in layer: _a = layer.split('kvstore' ) _a = ''''''.join(split_layer[0] )[:-1] _a = [tuple(('kvstore' + split_layer[1]).split('/' ) )] else: _a = layer.split('/' ) _a = '''/'''.join(split_layer[:-1] ) _a = (split_layer[-1],) if "kvstore/path" in layer: _a = f'{switch_checkpoint_path}/{checkpoint_info[layer]}' elif "kvstore/driver" in layer: _a = '''file''' else: _a = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def _A (lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Union[str, Any] ) -> Any: '''simple docstring''' _a = rename_keys(__SCREAMING_SNAKE_CASE ) _a = {} for k, v in current_block.items(): _a = v _a = new_current_block torch.save(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _A (lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :str , lowerCAmelCase__ :str = WEIGHTS_NAME ) -> List[str]: '''simple docstring''' _a = convert_file_size_to_int(__SCREAMING_SNAKE_CASE ) _a = [] _a = {} _a = 0 _a = 0 os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) with gfile.GFile(switch_checkpoint_path + '/checkpoint' , 'rb' ) as fp: _a = serialization.msgpack_restore(fp.read() )['''optimizer''']['''target'''] _a = flatten_dict(__SCREAMING_SNAKE_CASE , sep='/' ) _a = {} for layer in checkpoint_info.keys(): _a = get_key_and_tensorstore_dict( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if curr_real_layer_name in all_layers: _a = content else: _a = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file _a = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() _a = torch.tensor(__SCREAMING_SNAKE_CASE ) _a = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts _a = rename_base_flax_keys(tuple(key.split('/' ) ) , __SCREAMING_SNAKE_CASE ) _a = '''/'''.join(__SCREAMING_SNAKE_CASE ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: _a = os.path.join( __SCREAMING_SNAKE_CASE , weights_name.replace('.bin' , f'-{len(__SCREAMING_SNAKE_CASE )+1:05d}-of-???.bin' ) ) rename_and_save_block(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) sharded_state_dicts.append(current_block.keys() ) del current_block _a = {} _a = 0 _a = raw_weights.to(getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) current_block_size += weight_size total_size += weight_size # Add the last block _a = os.path.join(__SCREAMING_SNAKE_CASE , weights_name.replace('.bin' , f'-{len(__SCREAMING_SNAKE_CASE )+1:05d}-of-???.bin' ) ) rename_and_save_block(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(__SCREAMING_SNAKE_CASE ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index _a = {} _a = {} for idx, shard in enumerate(__SCREAMING_SNAKE_CASE ): _a = weights_name.replace( '.bin' , f'-{idx+1:05d}-of-{len(__SCREAMING_SNAKE_CASE ):05d}.bin' ) # len(sharded_state_dicts):05d} _a = os.path.join(__SCREAMING_SNAKE_CASE , weights_name.replace('.bin' , f'-{idx+1:05d}-of-???.bin' ) ) os.rename(__SCREAMING_SNAKE_CASE , os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) _a = shard for key in shard: _a = shard_file # Add the metadata _a = {'''total_size''': total_size} _a = {'''metadata''': metadata, '''weight_map''': weight_map} with open(os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , 'w' , encoding='utf-8' ) as f: _a = json.dumps(__SCREAMING_SNAKE_CASE , indent=2 , sort_keys=__SCREAMING_SNAKE_CASE ) + '''\n''' f.write(__SCREAMING_SNAKE_CASE ) return metadata, index if __name__ == "__main__": a_ : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--switch_t5x_checkpoint_path", default="/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600", type=str, required=False, help="Path to a directory containing a folder per layer. Follows the original Google format.", ) parser.add_argument("--max_shard_size", default="10GB", required=False, help="Max shard size") parser.add_argument("--dtype", default="bfloat16", type=str, required=False, help="dtype of the saved model") parser.add_argument( "--pytorch_dump_folder_path", default="/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted", type=str, required=False, help="Path to the output pytorch model.", ) a_ : Optional[Any] = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def _A () -> str: '''simple docstring''' from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer _a = SwitchTransformersConfig.from_pretrained('google/switch-base-8' ) config.save_pretrained('/home/arthur_huggingface_co/transformers/switch_converted' ) _a = SwitchTransformersForConditionalGeneration.from_pretrained( '/home/arthur_huggingface_co/transformers/switch_converted' , device_map='auto' ) _a = TaTokenizer.from_pretrained('t5-small' ) _a = '''A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.''' _a = tokenizer(__SCREAMING_SNAKE_CASE , return_tensors='pt' ).input_ids _a = model.generate(__SCREAMING_SNAKE_CASE , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
168
'''simple docstring''' import torch from transformers import AutoModel class lowerCAmelCase__ ( torch.nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE="sayef/fsner-bert-base-uncased" ): """simple docstring""" super(__SCREAMING_SNAKE_CASE , self ).__init__() lowercase_ : Tuple = AutoModel.from_pretrained(__SCREAMING_SNAKE_CASE , return_dict=__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = torch.nn.CosineSimilarity(3 , 1E-0_8 ) lowercase_ : Optional[Any] = torch.nn.Softmax(dim=1 ) def _snake_case ( self , **__SCREAMING_SNAKE_CASE ): """simple docstring""" return self.bert(**__SCREAMING_SNAKE_CASE ).last_hidden_state def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" return token_embeddings.sum(2 , keepdim=__SCREAMING_SNAKE_CASE ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=1 ): """simple docstring""" return self.softmax(T * self.cos(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Optional[Any] = W_supports['''sizes'''].tolist() lowercase_ : Dict = W_supports['''start_token_id'''].item() lowercase_ : List[Any] = W_supports['''end_token_id'''].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] lowercase_ : List[str] = self.BERT(**__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = self.BERT(**__SCREAMING_SNAKE_CASE ) lowercase_ : str = None lowercase_ : Dict = None lowercase_ : Tuple = W_supports['''input_ids'''] == start_token_id lowercase_ : Any = W_supports['''input_ids'''] == end_token_id for i, size in enumerate(__SCREAMING_SNAKE_CASE ): if i == 0: lowercase_ : List[str] = 0 else: lowercase_ : List[Any] = support_sizes[i - 1] lowercase_ : str = S[s : s + size][start_token_masks[s : s + size]] lowercase_ : Optional[int] = S[s : s + size][end_token_masks[s : s + size]] lowercase_ : List[str] = torch.matmul(q[i] , s_start.T ).sum(1 ).softmax(0 ) lowercase_ : List[str] = torch.matmul(q[i] , s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: lowercase_ : Tuple = torch.vstack((p_starts, p_start) ) lowercase_ : Optional[Any] = torch.vstack((p_ends, p_end) ) else: lowercase_ : str = p_start lowercase_ : int = p_end return p_starts, p_ends
93
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging a_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( a__ ): snake_case_ = ["""pixel_values"""] def __init__( self : int , __lowercase : bool = True , __lowercase : int = 32 , __lowercase : List[str]=PILImageResampling.BILINEAR , __lowercase : bool = True , **__lowercase : Union[str, Any] , ) -> List[str]: SCREAMING_SNAKE_CASE__ : Optional[Any] =do_resize SCREAMING_SNAKE_CASE__ : List[str] =do_rescale SCREAMING_SNAKE_CASE__ : Union[str, Any] =size_divisor SCREAMING_SNAKE_CASE__ : Any =resample super().__init__(**_lowerCamelCase ) def __magic_name__ ( self : Optional[int] , __lowercase : np.ndarray , __lowercase : int , __lowercase : str , __lowercase : Optional[ChannelDimension] = None , **__lowercase : Optional[Any] ) -> int: SCREAMING_SNAKE_CASE__ : Tuple =get_image_size(_lowerCamelCase ) # Rounds the height and width down to the closest multiple of size_divisor SCREAMING_SNAKE_CASE__ : Union[str, Any] =height // size_divisor * size_divisor SCREAMING_SNAKE_CASE__ : Dict =width // size_divisor * size_divisor SCREAMING_SNAKE_CASE__ : Any =resize(_lowerCamelCase , (new_h, new_w) , resample=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) return image def __magic_name__ ( self : List[str] , __lowercase : np.ndarray , __lowercase : float , __lowercase : Optional[ChannelDimension] = None , **__lowercase : Optional[int] ) -> Tuple: return rescale(image=_lowerCamelCase , scale=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) def __magic_name__ ( self : Optional[Any] , __lowercase : Union["PIL.Image.Image", TensorType, List["PIL.Image.Image"], List[TensorType]] , __lowercase : Optional[bool] = None , __lowercase : Optional[int] = None , __lowercase : str=None , __lowercase : Optional[bool] = None , __lowercase : Optional[Union[TensorType, str]] = None , __lowercase : ChannelDimension = ChannelDimension.FIRST , **__lowercase : Any , ) -> Tuple: SCREAMING_SNAKE_CASE__ : Dict =do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE__ : List[str] =do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE__ : str =size_divisor if size_divisor is not None else self.size_divisor SCREAMING_SNAKE_CASE__ : Dict =resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('''size_divisor is required for resizing''' ) SCREAMING_SNAKE_CASE__ : Tuple =make_list_of_images(_lowerCamelCase ) if not valid_images(_lowerCamelCase ): raise ValueError('''Invalid image(s)''' ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE__ : Dict =[to_numpy_array(_lowerCamelCase ) for img in images] if do_resize: SCREAMING_SNAKE_CASE__ : Tuple =[self.resize(_lowerCamelCase , size_divisor=_lowerCamelCase , resample=_lowerCamelCase ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE__ : str =[self.rescale(_lowerCamelCase , scale=1 / 2_55 ) for image in images] SCREAMING_SNAKE_CASE__ : List[str] =[to_channel_dimension_format(_lowerCamelCase , _lowerCamelCase ) for image in images] SCREAMING_SNAKE_CASE__ : Union[str, Any] ={'''pixel_values''': images} return BatchFeature(data=_lowerCamelCase , tensor_type=_lowerCamelCase )
354
'''simple docstring''' from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class __SCREAMING_SNAKE_CASE : snake_case_ = PegasusConfig snake_case_ = {} snake_case_ = """gelu""" def __init__( self : int , __lowercase : Optional[Any] , __lowercase : int=13 , __lowercase : List[str]=7 , __lowercase : Dict=True , __lowercase : Tuple=False , __lowercase : Optional[Any]=99 , __lowercase : str=32 , __lowercase : List[str]=2 , __lowercase : str=4 , __lowercase : Optional[int]=37 , __lowercase : List[Any]=0.1 , __lowercase : List[Any]=0.1 , __lowercase : List[Any]=40 , __lowercase : str=2 , __lowercase : List[Any]=1 , __lowercase : Optional[Any]=0 , ) -> Tuple: SCREAMING_SNAKE_CASE__ : Optional[Any] =parent SCREAMING_SNAKE_CASE__ : List[Any] =batch_size SCREAMING_SNAKE_CASE__ : Optional[int] =seq_length SCREAMING_SNAKE_CASE__ : Optional[Any] =is_training SCREAMING_SNAKE_CASE__ : Union[str, Any] =use_labels SCREAMING_SNAKE_CASE__ : str =vocab_size SCREAMING_SNAKE_CASE__ : Optional[Any] =hidden_size SCREAMING_SNAKE_CASE__ : List[str] =num_hidden_layers SCREAMING_SNAKE_CASE__ : int =num_attention_heads SCREAMING_SNAKE_CASE__ : List[str] =intermediate_size SCREAMING_SNAKE_CASE__ : List[Any] =hidden_dropout_prob SCREAMING_SNAKE_CASE__ : List[Any] =attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : List[Any] =max_position_embeddings SCREAMING_SNAKE_CASE__ : Optional[Any] =eos_token_id SCREAMING_SNAKE_CASE__ : Any =pad_token_id SCREAMING_SNAKE_CASE__ : Union[str, Any] =bos_token_id def __magic_name__ ( self : Any ) -> str: SCREAMING_SNAKE_CASE__ : List[str] =ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) SCREAMING_SNAKE_CASE__ : Optional[int] =tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) SCREAMING_SNAKE_CASE__ : Any =tf.concat([input_ids, eos_tensor] , axis=1 ) SCREAMING_SNAKE_CASE__ : Dict =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ : Optional[int] =self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) SCREAMING_SNAKE_CASE__ : Union[str, Any] =prepare_pegasus_inputs_dict(__lowercase , __lowercase , __lowercase ) return config, inputs_dict def __magic_name__ ( self : Optional[int] , __lowercase : List[str] , __lowercase : Optional[int] ) -> Tuple: SCREAMING_SNAKE_CASE__ : List[Any] =TFPegasusModel(config=__lowercase ).get_decoder() SCREAMING_SNAKE_CASE__ : List[str] =inputs_dict['''input_ids'''] SCREAMING_SNAKE_CASE__ : Tuple =input_ids[:1, :] SCREAMING_SNAKE_CASE__ : Tuple =inputs_dict['''attention_mask'''][:1, :] SCREAMING_SNAKE_CASE__ : Tuple =inputs_dict['''head_mask'''] SCREAMING_SNAKE_CASE__ : List[str] =1 # first forward pass SCREAMING_SNAKE_CASE__ : Any =model(__lowercase , attention_mask=__lowercase , head_mask=__lowercase , use_cache=__lowercase ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Tuple =outputs.to_tuple() # create hypothetical next token and extent to next_input_ids SCREAMING_SNAKE_CASE__ : str =ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE__ : List[Any] =tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and SCREAMING_SNAKE_CASE__ : Tuple =tf.concat([input_ids, next_tokens] , axis=-1 ) SCREAMING_SNAKE_CASE__ : Optional[Any] =tf.concat([attention_mask, next_attn_mask] , axis=-1 ) SCREAMING_SNAKE_CASE__ : int =model(__lowercase , attention_mask=__lowercase )[0] SCREAMING_SNAKE_CASE__ : Any =model(__lowercase , attention_mask=__lowercase , past_key_values=__lowercase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice SCREAMING_SNAKE_CASE__ : Optional[Any] =int(ids_tensor((1,) , output_from_past.shape[-1] ) ) SCREAMING_SNAKE_CASE__ : Any =output_from_no_past[:, -3:, random_slice_idx] SCREAMING_SNAKE_CASE__ : List[str] =output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__lowercase , __lowercase , rtol=1e-3 ) def _a( UpperCamelCase__ : Any, UpperCamelCase__ : Union[str, Any], UpperCamelCase__ : Union[str, Any], UpperCamelCase__ : Optional[Any]=None, UpperCamelCase__ : Optional[Any]=None, UpperCamelCase__ : Union[str, Any]=None, UpperCamelCase__ : Any=None, UpperCamelCase__ : Optional[Any]=None, ): '''simple docstring''' if attention_mask is None: SCREAMING_SNAKE_CASE__ : str =tf.cast(tf.math.not_equal(UpperCamelCase__, config.pad_token_id ), tf.inta ) if decoder_attention_mask is None: SCREAMING_SNAKE_CASE__ : Any =tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape, dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:], config.pad_token_id ), tf.inta ), ], axis=-1, ) if head_mask is None: SCREAMING_SNAKE_CASE__ : int =tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: SCREAMING_SNAKE_CASE__ : List[Any] =tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: SCREAMING_SNAKE_CASE__ : List[str] =tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __SCREAMING_SNAKE_CASE ( lowerCamelCase , lowerCamelCase , unittest.TestCase ): snake_case_ = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () snake_case_ = (TFPegasusForConditionalGeneration,) if is_tf_available() else () snake_case_ = ( { """conversational""": TFPegasusForConditionalGeneration, """feature-extraction""": TFPegasusModel, """summarization""": TFPegasusForConditionalGeneration, """text2text-generation""": TFPegasusForConditionalGeneration, """translation""": TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) snake_case_ = True snake_case_ = False snake_case_ = False def __magic_name__ ( self : Union[str, Any] ) -> str: SCREAMING_SNAKE_CASE__ : List[Any] =TFPegasusModelTester(self ) SCREAMING_SNAKE_CASE__ : Dict =ConfigTester(self , config_class=__lowercase ) def __magic_name__ ( self : int ) -> Any: self.config_tester.run_common_tests() def __magic_name__ ( self : Optional[Any] ) -> List[Any]: SCREAMING_SNAKE_CASE__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__lowercase ) @require_sentencepiece @require_tokenizers @require_tf class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): snake_case_ = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] snake_case_ = [ """California's largest electricity provider has cut power to hundreds of thousands of customers in an effort to""" """ reduce the risk of wildfires.""", """N-Dubz have revealed they\'re \"grateful\" to have been nominated for four Mobo Awards.""", ] # differs slightly from pytorch, likely due to numerical differences in linear layers snake_case_ = """google/pegasus-xsum""" @cached_property def __magic_name__ ( self : Optional[int] ) -> Tuple: return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def __magic_name__ ( self : List[Any] ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ : Optional[int] =TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def __magic_name__ ( self : List[str] , **__lowercase : Any ) -> Optional[int]: SCREAMING_SNAKE_CASE__ : Union[str, Any] =self.translate_src_text(**__lowercase ) assert self.expected_text == generated_words def __magic_name__ ( self : Optional[Any] , **__lowercase : List[str] ) -> Optional[int]: SCREAMING_SNAKE_CASE__ : Union[str, Any] =self.tokenizer(self.src_text , **__lowercase , padding=__lowercase , return_tensors='''tf''' ) SCREAMING_SNAKE_CASE__ : Union[str, Any] =self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=__lowercase , ) SCREAMING_SNAKE_CASE__ : Any =self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=__lowercase ) return generated_words @slow def __magic_name__ ( self : Optional[Any] ) -> Optional[int]: self._assert_generated_batch_equal_expected()
222
0
import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class _a : """simple docstring""" def __init__( self: int , __lowerCamelCase: Tuple , __lowerCamelCase: Tuple=2 , __lowerCamelCase: Dict=3 , __lowerCamelCase: List[str]=4 , __lowerCamelCase: Tuple=2 , __lowerCamelCase: int=7 , __lowerCamelCase: Any=True , __lowerCamelCase: Any=True , __lowerCamelCase: Optional[Any]=True , __lowerCamelCase: int=True , __lowerCamelCase: str=99 , __lowerCamelCase: Any=36 , __lowerCamelCase: int=3 , __lowerCamelCase: str=4 , __lowerCamelCase: Any=37 , __lowerCamelCase: Optional[Any]="gelu" , __lowerCamelCase: Any=0.1 , __lowerCamelCase: Dict=0.1 , __lowerCamelCase: List[Any]=512 , __lowerCamelCase: Any=16 , __lowerCamelCase: int=2 , __lowerCamelCase: str=0.02 , __lowerCamelCase: Optional[Any]=6 , __lowerCamelCase: Optional[int]=6 , __lowerCamelCase: str=3 , __lowerCamelCase: Optional[int]=4 , __lowerCamelCase: Optional[Any]=None , __lowerCamelCase: Any=1000 , ): '''simple docstring''' UpperCamelCase__: Any = parent UpperCamelCase__: List[Any] = batch_size UpperCamelCase__: Optional[int] = num_channels UpperCamelCase__: Union[str, Any] = image_size UpperCamelCase__: Dict = patch_size UpperCamelCase__: List[Any] = text_seq_length UpperCamelCase__: Optional[Any] = is_training UpperCamelCase__: str = use_input_mask UpperCamelCase__: Dict = use_token_type_ids UpperCamelCase__: Optional[Any] = use_labels UpperCamelCase__: Optional[int] = vocab_size UpperCamelCase__: Dict = hidden_size UpperCamelCase__: Optional[int] = num_hidden_layers UpperCamelCase__: int = num_attention_heads UpperCamelCase__: int = intermediate_size UpperCamelCase__: Dict = hidden_act UpperCamelCase__: Any = hidden_dropout_prob UpperCamelCase__: Union[str, Any] = attention_probs_dropout_prob UpperCamelCase__: int = max_position_embeddings UpperCamelCase__: Any = type_vocab_size UpperCamelCase__: int = type_sequence_label_size UpperCamelCase__: Dict = initializer_range UpperCamelCase__: str = coordinate_size UpperCamelCase__: int = shape_size UpperCamelCase__: List[Any] = num_labels UpperCamelCase__: List[Any] = num_choices UpperCamelCase__: List[Any] = scope UpperCamelCase__: Any = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) UpperCamelCase__: Optional[int] = text_seq_length UpperCamelCase__: Optional[int] = (image_size // patch_size) ** 2 + 1 UpperCamelCase__: List[Any] = self.text_seq_length + self.image_seq_length def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' UpperCamelCase__: List[Any] = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) UpperCamelCase__: Optional[Any] = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: UpperCamelCase__: List[Any] = bbox[i, j, 3] UpperCamelCase__: Any = bbox[i, j, 1] UpperCamelCase__: Optional[int] = t if bbox[i, j, 2] < bbox[i, j, 0]: UpperCamelCase__: Tuple = bbox[i, j, 2] UpperCamelCase__: int = bbox[i, j, 0] UpperCamelCase__: List[Any] = t UpperCamelCase__: Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase__: int = None if self.use_input_mask: UpperCamelCase__: Any = random_attention_mask([self.batch_size, self.text_seq_length] ) UpperCamelCase__: List[str] = None if self.use_token_type_ids: UpperCamelCase__: List[str] = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) UpperCamelCase__: Union[str, Any] = None UpperCamelCase__: Union[str, Any] = None if self.use_labels: UpperCamelCase__: Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase__: List[Any] = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) UpperCamelCase__: Any = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def UpperCAmelCase_ ( self: Optional[int] , __lowerCamelCase: Optional[int] , __lowerCamelCase: str , __lowerCamelCase: Any , __lowerCamelCase: Optional[int] , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Optional[int] , __lowerCamelCase: Tuple , __lowerCamelCase: Any ): '''simple docstring''' UpperCamelCase__: Optional[Any] = LayoutLMvaModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() # text + image UpperCamelCase__: str = model(__lowerCamelCase , pixel_values=__lowerCamelCase ) UpperCamelCase__: List[str] = model( __lowerCamelCase , bbox=__lowerCamelCase , pixel_values=__lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase ) UpperCamelCase__: Dict = model(__lowerCamelCase , bbox=__lowerCamelCase , pixel_values=__lowerCamelCase , token_type_ids=__lowerCamelCase ) UpperCamelCase__: Tuple = model(__lowerCamelCase , bbox=__lowerCamelCase , pixel_values=__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only UpperCamelCase__: Optional[int] = model(__lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only UpperCamelCase__: Tuple = model(pixel_values=__lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self: Any , __lowerCamelCase: Dict , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: List[str] , __lowerCamelCase: int , __lowerCamelCase: Any , __lowerCamelCase: Tuple , __lowerCamelCase: int , __lowerCamelCase: Any ): '''simple docstring''' UpperCamelCase__: Optional[Any] = self.num_labels UpperCamelCase__: Union[str, Any] = LayoutLMvaForSequenceClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase__: Optional[int] = model( __lowerCamelCase , bbox=__lowerCamelCase , pixel_values=__lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self: str , __lowerCamelCase: int , __lowerCamelCase: List[Any] , __lowerCamelCase: Dict , __lowerCamelCase: List[str] , __lowerCamelCase: Optional[int] , __lowerCamelCase: Dict , __lowerCamelCase: List[Any] , __lowerCamelCase: str ): '''simple docstring''' UpperCamelCase__: Tuple = self.num_labels UpperCamelCase__: Dict = LayoutLMvaForTokenClassification(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase__: List[str] = model( __lowerCamelCase , bbox=__lowerCamelCase , pixel_values=__lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def UpperCAmelCase_ ( self: Optional[Any] , __lowerCamelCase: str , __lowerCamelCase: Dict , __lowerCamelCase: List[Any] , __lowerCamelCase: Any , __lowerCamelCase: Dict , __lowerCamelCase: Dict , __lowerCamelCase: Dict , __lowerCamelCase: Any ): '''simple docstring''' UpperCamelCase__: int = LayoutLMvaForQuestionAnswering(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase__: List[str] = model( __lowerCamelCase , bbox=__lowerCamelCase , pixel_values=__lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' UpperCamelCase__: Tuple = self.prepare_config_and_inputs() ( ( UpperCamelCase__ ) , ( UpperCamelCase__ ) , ( UpperCamelCase__ ) , ( UpperCamelCase__ ) , ( UpperCamelCase__ ) , ( UpperCamelCase__ ) , ( UpperCamelCase__ ) , ( UpperCamelCase__ ) , ): Dict = config_and_inputs UpperCamelCase__: Union[str, Any] = { "input_ids": input_ids, "bbox": bbox, "pixel_values": pixel_values, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_torch class _a ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) UpperCamelCase__ = ( {"""document-question-answering""": LayoutLMvaForQuestionAnswering, """feature-extraction""": LayoutLMvaModel} if is_torch_available() else {} ) def UpperCAmelCase_ ( self: List[Any] , __lowerCamelCase: Any , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Tuple , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: List[Any] ): '''simple docstring''' return True def UpperCAmelCase_ ( self: Union[str, Any] ): '''simple docstring''' UpperCamelCase__: List[str] = LayoutLMvaModelTester(self ) UpperCamelCase__: List[str] = ConfigTester(self , config_class=__lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self: Union[str, Any] , __lowerCamelCase: Optional[int] , __lowerCamelCase: List[Any] , __lowerCamelCase: List[str]=False ): '''simple docstring''' UpperCamelCase__: Dict = copy.deepcopy(__lowerCamelCase ) if model_class in get_values(__lowerCamelCase ): UpperCamelCase__: List[Any] = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(__lowerCamelCase , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(__lowerCamelCase ): UpperCamelCase__: int = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) elif model_class in get_values(__lowerCamelCase ): UpperCamelCase__: Optional[int] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) UpperCamelCase__: str = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) elif model_class in [ *get_values(__lowerCamelCase ), ]: UpperCamelCase__: List[str] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) elif model_class in [ *get_values(__lowerCamelCase ), ]: UpperCamelCase__: Any = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=__lowerCamelCase , ) return inputs_dict def UpperCAmelCase_ ( self: Optional[int] ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self: str ): '''simple docstring''' UpperCamelCase__: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def UpperCAmelCase_ ( self: List[Any] ): '''simple docstring''' UpperCamelCase__: Dict = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCamelCase__: str = type self.model_tester.create_and_check_model(*__lowerCamelCase ) def UpperCAmelCase_ ( self: str ): '''simple docstring''' UpperCamelCase__: List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__lowerCamelCase ) def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' UpperCamelCase__: Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowerCamelCase ) def UpperCAmelCase_ ( self: Any ): '''simple docstring''' UpperCamelCase__: int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__lowerCamelCase ) @slow def UpperCAmelCase_ ( self: int ): '''simple docstring''' for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase__: Union[str, Any] = LayoutLMvaModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def lowerCAmelCase_ ( ): UpperCamelCase__: str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") return image @require_torch class _a ( unittest.TestCase): """simple docstring""" @cached_property def UpperCAmelCase_ ( self: Dict ): '''simple docstring''' return LayoutLMvaImageProcessor(apply_ocr=__lowerCamelCase ) if is_vision_available() else None @slow def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' UpperCamelCase__: Optional[Any] = LayoutLMvaModel.from_pretrained("microsoft/layoutlmv3-base" ).to(__lowerCamelCase ) UpperCamelCase__: Dict = self.default_image_processor UpperCamelCase__: Optional[int] = prepare_img() UpperCamelCase__: Optional[Any] = image_processor(images=__lowerCamelCase , return_tensors="pt" ).pixel_values.to(__lowerCamelCase ) UpperCamelCase__: Optional[int] = torch.tensor([[1, 2]] ) UpperCamelCase__: List[str] = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass UpperCamelCase__: Optional[int] = model( input_ids=input_ids.to(__lowerCamelCase ) , bbox=bbox.to(__lowerCamelCase ) , pixel_values=pixel_values.to(__lowerCamelCase ) , ) # verify the logits UpperCamelCase__: List[str] = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , __lowerCamelCase ) UpperCamelCase__: Union[str, Any] = torch.tensor( [[-0.0_529, 0.3_618, 0.1_632], [-0.1_587, -0.1_667, -0.0_400], [-0.1_557, -0.1_671, -0.0_505]] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowerCamelCase , atol=1e-4 ) )
149
import logging import os from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional from tqdm import auto as tqdm_lib A__: Optional[int] = { '''debug''': logging.DEBUG, '''info''': logging.INFO, '''warning''': logging.WARNING, '''error''': logging.ERROR, '''critical''': logging.CRITICAL, } A__: int = logging.WARNING def lowerCAmelCase_ ( ): UpperCamelCase__: Optional[int] = os.getenv("DATASETS_VERBOSITY" ,A_) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( F"Unknown option DATASETS_VERBOSITY={env_level_str}, " F"has to be one of: { ', '.join(log_levels.keys()) }") return _default_log_level def lowerCAmelCase_ ( ): return __name__.split(".")[0] def lowerCAmelCase_ ( ): return logging.getLogger(_get_library_name()) def lowerCAmelCase_ ( ): # Apply our default configuration to the library root logger. UpperCamelCase__: Tuple = _get_library_root_logger() library_root_logger.setLevel(_get_default_logging_level()) def lowerCAmelCase_ ( ): UpperCamelCase__: Tuple = _get_library_root_logger() library_root_logger.setLevel(logging.NOTSET) def lowerCAmelCase_ ( A_ = None): if name is None: UpperCamelCase__: Optional[Any] = _get_library_name() return logging.getLogger(A_) def lowerCAmelCase_ ( ): return _get_library_root_logger().getEffectiveLevel() def lowerCAmelCase_ ( A_): _get_library_root_logger().setLevel(A_) def lowerCAmelCase_ ( ): return set_verbosity(A_) def lowerCAmelCase_ ( ): return set_verbosity(A_) def lowerCAmelCase_ ( ): return set_verbosity(A_) def lowerCAmelCase_ ( ): return set_verbosity(A_) def lowerCAmelCase_ ( ): UpperCamelCase__: List[Any] = False def lowerCAmelCase_ ( ): UpperCamelCase__: List[str] = True # Configure the library root logger at the module level (singleton-like) _configure_library_root_logger() class _a : """simple docstring""" def __init__( self: int , *__lowerCamelCase: Tuple , **__lowerCamelCase: str ): # pylint: disable=unused-argument '''simple docstring''' UpperCamelCase__: int = args[0] if args else None def __iter__( self: Optional[Any] ): '''simple docstring''' return iter(self._iterator ) def __getattr__( self: Dict , __lowerCamelCase: Any ): '''simple docstring''' def empty_fn(*__lowerCamelCase: Any , **__lowerCamelCase: Optional[Any] ): # pylint: disable=unused-argument return return empty_fn def __enter__( self: str ): '''simple docstring''' return self def __exit__( self: Tuple , __lowerCamelCase: List[str] , __lowerCamelCase: Any , __lowerCamelCase: List[Any] ): '''simple docstring''' return A__: Tuple = True class _a : """simple docstring""" def __call__( self: Any , *__lowerCamelCase: List[str] , __lowerCamelCase: List[Any]=False , **__lowerCamelCase: Union[str, Any] ): '''simple docstring''' if _tqdm_active and not disable: return tqdm_lib.tqdm(*__lowerCamelCase , **__lowerCamelCase ) else: return EmptyTqdm(*__lowerCamelCase , **__lowerCamelCase ) def UpperCAmelCase_ ( self: List[str] , *__lowerCamelCase: List[str] , **__lowerCamelCase: Tuple ): '''simple docstring''' UpperCamelCase__: Optional[int] = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*__lowerCamelCase , **__lowerCamelCase ) def UpperCAmelCase_ ( self: int ): '''simple docstring''' if _tqdm_active: return tqdm_lib.tqdm.get_lock() A__: Optional[Any] = _tqdm_cls() def lowerCAmelCase_ ( ): global _tqdm_active return bool(_tqdm_active) def lowerCAmelCase_ ( ): global _tqdm_active UpperCamelCase__: int = True def lowerCAmelCase_ ( ): global _tqdm_active UpperCamelCase__: str = False
149
1
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=99 , __UpperCamelCase=13 , __UpperCamelCase=16 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=2 , __UpperCamelCase=32 , __UpperCamelCase=4 , __UpperCamelCase=4 , __UpperCamelCase=30 , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=2 , __UpperCamelCase=None , ) -> List[str]: '''simple docstring''' __UpperCamelCase : List[Any] = parent __UpperCamelCase : Optional[Any] = batch_size __UpperCamelCase : Any = decoder_seq_length # For common tests __UpperCamelCase : int = self.decoder_seq_length __UpperCamelCase : List[Any] = is_training __UpperCamelCase : Any = use_attention_mask __UpperCamelCase : Optional[int] = use_labels __UpperCamelCase : int = vocab_size __UpperCamelCase : Dict = d_model __UpperCamelCase : List[Any] = d_model __UpperCamelCase : Dict = decoder_layers __UpperCamelCase : Optional[Any] = decoder_layers __UpperCamelCase : Dict = decoder_ffn_dim __UpperCamelCase : Optional[int] = decoder_attention_heads __UpperCamelCase : int = decoder_attention_heads __UpperCamelCase : Any = eos_token_id __UpperCamelCase : List[str] = bos_token_id __UpperCamelCase : Optional[Any] = pad_token_id __UpperCamelCase : Optional[int] = decoder_start_token_id __UpperCamelCase : Optional[int] = use_cache __UpperCamelCase : Any = max_position_embeddings __UpperCamelCase : Any = None __UpperCamelCase : Dict = decoder_seq_length __UpperCamelCase : Any = 2 __UpperCamelCase : int = 1 def __lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' __UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) __UpperCamelCase : Optional[int] = None if self.use_attention_mask: __UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) __UpperCamelCase : Optional[int] = None if self.use_labels: __UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) __UpperCamelCase : Optional[int] = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ) -> Dict: '''simple docstring''' __UpperCamelCase : Tuple = True __UpperCamelCase : Dict = TrOCRDecoder(config=__UpperCamelCase ).to(__UpperCamelCase ).eval() __UpperCamelCase : List[str] = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass __UpperCamelCase : List[str] = model(__UpperCamelCase , use_cache=__UpperCamelCase ) __UpperCamelCase : List[Any] = model(__UpperCamelCase ) __UpperCamelCase : str = model(__UpperCamelCase , use_cache=__UpperCamelCase ) self.parent.assertTrue(len(__UpperCamelCase ) == len(__UpperCamelCase ) ) self.parent.assertTrue(len(__UpperCamelCase ) == len(__UpperCamelCase ) + 1 ) __UpperCamelCase : Optional[int] = outputs["past_key_values"] # create hypothetical next token and extent to next_input_ids __UpperCamelCase : str = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and __UpperCamelCase : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) __UpperCamelCase : Optional[Any] = model(__UpperCamelCase )["last_hidden_state"] __UpperCamelCase : Any = model(__UpperCamelCase , past_key_values=__UpperCamelCase )["last_hidden_state"] # select random slice __UpperCamelCase : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() __UpperCamelCase : List[str] = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() __UpperCamelCase : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1E-3 ) def __lowerCamelCase ( self ) -> Any: '''simple docstring''' __UpperCamelCase : int = self.prepare_config_and_inputs() __UpperCamelCase : Tuple = config_and_inputs __UpperCamelCase : Optional[Any] = {"input_ids": input_ids, "attention_mask": attention_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" lowercase : Any = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase : int = (TrOCRForCausalLM,) if is_torch_available() else () lowercase : Union[str, Any] = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase : Union[str, Any] = True lowercase : Optional[int] = False def __lowerCamelCase ( self ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase : Optional[int] = TrOCRStandaloneDecoderModelTester(self , is_training=__UpperCamelCase ) __UpperCamelCase : List[Any] = ConfigTester(self , config_class=__UpperCamelCase ) def __lowerCamelCase ( self ) -> Dict: '''simple docstring''' pass def __lowerCamelCase ( self ) -> Tuple: '''simple docstring''' pass def __lowerCamelCase ( self ) -> Any: '''simple docstring''' pass def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def __lowerCamelCase ( self ) -> Dict: '''simple docstring''' __UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*__UpperCamelCase ) def __lowerCamelCase ( self ) -> Dict: '''simple docstring''' return @unittest.skip("The model doesn't support left padding" ) # and it's not used enough to be worth fixing :) def __lowerCamelCase ( self ) -> Dict: '''simple docstring''' pass
361
from math import sqrt def UpperCAmelCase_ (_lowerCAmelCase : int = 1_00_00_00 ): __UpperCamelCase : int = 0 __UpperCamelCase : int = 0 __UpperCamelCase : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_lowerCAmelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
171
0
"""simple docstring""" from __future__ import annotations from functools import lru_cache from math import ceil _snake_case : str = 100 _snake_case : Any = set(range(3, NUM_PRIMES, 2)) primes.add(2) _snake_case : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100 ) def A__ ( UpperCamelCase ): if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} A = set() A = 42 A = 42 for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def A__ ( UpperCamelCase = 5_000 ): for number_to_partition in range(1 , _lowerCAmelCase ): if len(partition(_lowerCAmelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(F"""{solution() = }""")
292
import pickle import numpy as np from matplotlib import pyplot as plt class A__ : def __init__( self , A_ , A_ , A_ , A_ , A_ , A_=0.2 , A_=0.2 ): '''simple docstring''' UpperCamelCase : int = bp_numa UpperCamelCase : int = bp_numa UpperCamelCase : List[Any] = bp_numa UpperCamelCase : Optional[int] = conva_get[:2] UpperCamelCase : Optional[Any] = conva_get[2] UpperCamelCase : Dict = size_pa UpperCamelCase : Union[str, Any] = rate_w UpperCamelCase : Dict = rate_t UpperCamelCase : Union[str, Any] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] UpperCamelCase : Any = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCamelCase : List[Any] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCamelCase : Optional[Any] = -2 * np.random.rand(self.conva[1] ) + 1 UpperCamelCase : Any = -2 * np.random.rand(self.num_bpa ) + 1 UpperCamelCase : int = -2 * np.random.rand(self.num_bpa ) + 1 def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(A_ , "wb" ) as f: pickle.dump(A_ , A_ ) print(F"""Model saved: {save_path}""" ) @classmethod def __UpperCamelCase( cls , A_ ): '''simple docstring''' with open(A_ , "rb" ) as f: UpperCamelCase : Optional[Any] = pickle.load(A_ ) # noqa: S301 UpperCamelCase : List[Any] = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) UpperCamelCase : Union[str, Any] = model_dic.get("size_pooling1" ) UpperCamelCase : List[Any] = model_dic.get("num_bp1" ) UpperCamelCase : Dict = model_dic.get("num_bp2" ) UpperCamelCase : Dict = model_dic.get("num_bp3" ) UpperCamelCase : Dict = model_dic.get("rate_weight" ) UpperCamelCase : str = model_dic.get("rate_thre" ) # create model instance UpperCamelCase : Any = CNN(A_ , A_ , A_ , A_ , A_ , A_ , A_ ) # modify model parameter UpperCamelCase : str = model_dic.get("w_conv1" ) UpperCamelCase : Optional[Any] = model_dic.get("wkj" ) UpperCamelCase : int = model_dic.get("vji" ) UpperCamelCase : Any = model_dic.get("thre_conv1" ) UpperCamelCase : Optional[int] = model_dic.get("thre_bp2" ) UpperCamelCase : Union[str, Any] = model_dic.get("thre_bp3" ) return conv_ins def __UpperCamelCase( self , A_ ): '''simple docstring''' return 1 / (1 + np.exp(-1 * x )) def __UpperCamelCase( self , A_ ): '''simple docstring''' return round(A_ , 3 ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = convs[0] UpperCamelCase : Optional[Any] = convs[1] UpperCamelCase : Optional[Any] = np.shape(A_ )[0] # get the data slice of original image data, data_focus UpperCamelCase : List[str] = [] for i_focus in range(0 , size_data - size_conv + 1 , A_ ): for j_focus in range(0 , size_data - size_conv + 1 , A_ ): UpperCamelCase : Union[str, Any] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(A_ ) # calculate the feature map of every single kernel, and saved as list of matrix UpperCamelCase : int = [] UpperCamelCase : Optional[Any] = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(A_ ): UpperCamelCase : str = [] for i_focus in range(len(A_ ) ): UpperCamelCase : List[Any] = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(A_ ) ) UpperCamelCase : Optional[int] = np.asmatrix(A_ ).reshape( A_ , A_ ) data_featuremap.append(A_ ) # expanding the data slice to One dimenssion UpperCamelCase : List[Any] = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(A_ ) ) UpperCamelCase : Tuple = np.asarray(A_ ) return focus_list, data_featuremap def __UpperCamelCase( self , A_ , A_ , A_="average_pool" ): '''simple docstring''' UpperCamelCase : Any = len(featuremaps[0] ) UpperCamelCase : str = int(size_map / size_pooling ) UpperCamelCase : Optional[int] = [] for i_map in range(len(A_ ) ): UpperCamelCase : Tuple = featuremaps[i_map] UpperCamelCase : Any = [] for i_focus in range(0 , A_ , A_ ): for j_focus in range(0 , A_ , A_ ): UpperCamelCase : int = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(A_ ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(A_ ) ) UpperCamelCase : Optional[Any] = np.asmatrix(A_ ).reshape(A_ , A_ ) featuremap_pooled.append(A_ ) return featuremap_pooled def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = [] for i in range(len(A_ ) ): UpperCamelCase : List[Any] = np.shape(data[i] ) UpperCamelCase : str = data[i].reshape(1 , shapes[0] * shapes[1] ) UpperCamelCase : Optional[int] = data_listed.getA().tolist()[0] data_expanded.extend(A_ ) UpperCamelCase : Any = np.asarray(A_ ) return data_expanded def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = np.asarray(A_ ) UpperCamelCase : List[Any] = np.shape(A_ ) UpperCamelCase : Any = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : int = [] UpperCamelCase : Optional[int] = 0 for i_map in range(A_ ): UpperCamelCase : int = np.ones((size_map, size_map) ) for i in range(0 , A_ , A_ ): for j in range(0 , A_ , A_ ): UpperCamelCase : str = pd_pool[ i_pool ] UpperCamelCase : str = i_pool + 1 UpperCamelCase : str = np.multiply( A_ , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(A_ ) return pd_all def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_=bool ): '''simple docstring''' print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(A_ )) ) print((" - - Shape: Teach_Data ", np.shape(A_ )) ) UpperCamelCase : List[str] = 0 UpperCamelCase : Union[str, Any] = [] UpperCamelCase : int = 1_0000 while rp < n_repeat and mse >= error_accuracy: UpperCamelCase : Tuple = 0 print(F"""-------------Learning Time {rp}--------------""" ) for p in range(len(A_ ) ): # print('------------Learning Image: %d--------------'%p) UpperCamelCase : Any = np.asmatrix(datas_train[p] ) UpperCamelCase : List[str] = np.asarray(datas_teach[p] ) UpperCamelCase , UpperCamelCase : Dict = self.convolute( A_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase : Tuple = self.pooling(A_ , self.size_poolinga ) UpperCamelCase : int = np.shape(A_ ) UpperCamelCase : List[str] = self._expand(A_ ) UpperCamelCase : Optional[int] = data_bp_input UpperCamelCase : str = np.dot(A_ , self.vji.T ) - self.thre_bpa UpperCamelCase : Optional[int] = self.sig(A_ ) UpperCamelCase : List[Any] = np.dot(A_ , self.wkj.T ) - self.thre_bpa UpperCamelCase : Dict = self.sig(A_ ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- UpperCamelCase : List[Any] = np.multiply( (data_teach - bp_outa) , np.multiply(A_ , (1 - bp_outa) ) ) UpperCamelCase : str = np.multiply( np.dot(A_ , self.wkj ) , np.multiply(A_ , (1 - bp_outa) ) ) UpperCamelCase : Any = np.dot(A_ , self.vji ) UpperCamelCase : Dict = pd_i_all / (self.size_poolinga * self.size_poolinga) UpperCamelCase : List[Any] = pd_conva_pooled.T.getA().tolist() UpperCamelCase : List[Any] = self._calculate_gradient_from_pool( A_ , A_ , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): UpperCamelCase : List[Any] = self._expand_mat(pd_conva_all[k_conv] ) UpperCamelCase : List[Any] = self.rate_weight * np.dot(A_ , A_ ) UpperCamelCase : str = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) UpperCamelCase : Dict = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer UpperCamelCase : Optional[Any] = self.wkj + pd_k_all.T * bp_outa * self.rate_weight UpperCamelCase : List[Any] = self.vji + pd_j_all.T * bp_outa * self.rate_weight UpperCamelCase : Optional[Any] = self.thre_bpa - pd_k_all * self.rate_thre UpperCamelCase : List[str] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image UpperCamelCase : List[Any] = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) UpperCamelCase : Any = rp + 1 UpperCamelCase : Union[str, Any] = error_count / patterns all_mse.append(A_ ) def draw_error(): UpperCamelCase : Tuple = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(A_ , "+-" ) plt.plot(A_ , "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(A_ , alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, F""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(A_ )) ) for p in range(len(A_ ) ): UpperCamelCase : int = np.asmatrix(datas_test[p] ) UpperCamelCase , UpperCamelCase : Any = self.convolute( A_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase : List[str] = self.pooling(A_ , self.size_poolinga ) UpperCamelCase : Dict = self._expand(A_ ) UpperCamelCase : List[Any] = data_bp_input UpperCamelCase : Any = bp_outa * self.vji.T - self.thre_bpa UpperCamelCase : List[Any] = self.sig(A_ ) UpperCamelCase : int = bp_outa * self.wkj.T - self.thre_bpa UpperCamelCase : Optional[int] = self.sig(A_ ) produce_out.extend(bp_outa.getA().tolist() ) UpperCamelCase : List[str] = [list(map(self.do_round , A_ ) ) for each in produce_out] return np.asarray(A_ ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = np.asmatrix(A_ ) UpperCamelCase , UpperCamelCase : List[Any] = self.convolute( A_ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase : str = self.pooling(A_ , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
52
0
"""simple docstring""" from math import ceil def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase = 1001 ): _lowercase : Union[str, Any] = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): _lowercase : Optional[int] = 2 * i + 1 _lowercase : Tuple = 2 * i _lowercase : Tuple = total + 4 * odd**2 - 6 * even return total if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution()) else: try: UpperCAmelCase: str = int(sys.argv[1]) print(solution(n)) except ValueError: print("""Invalid entry - please enter a number""")
336
"""simple docstring""" from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __SCREAMING_SNAKE_CASE ( ): _lowercase : Dict = [randint(-1000 , 1000 ) for i in range(10 )] _lowercase : Tuple = randint(-5000 , 5000 ) return (arr, r) UpperCAmelCase: int = make_dataset() def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase , __UpperCAmelCase ): for triplet in permutations(__UpperCAmelCase , 3 ): if sum(__UpperCAmelCase ) == target: return tuple(sorted(__UpperCAmelCase ) ) return (0, 0, 0) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase , __UpperCAmelCase ): arr.sort() _lowercase : Optional[Any] = len(__UpperCAmelCase ) for i in range(n - 1 ): _lowercase , _lowercase : str = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __SCREAMING_SNAKE_CASE ( ): _lowercase : Tuple = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ _lowercase : Union[str, Any] = """ triplet_sum1(*dataset) """ _lowercase : Union[str, Any] = """ triplet_sum2(*dataset) """ _lowercase : Dict = repeat(setup=__UpperCAmelCase , stmt=__UpperCAmelCase , repeat=5 , number=10000 ) _lowercase : Any = repeat(setup=__UpperCAmelCase , stmt=__UpperCAmelCase , repeat=5 , number=10000 ) return (min(__UpperCAmelCase ), min(__UpperCAmelCase )) if __name__ == "__main__": from doctest import testmod testmod() UpperCAmelCase: Any = solution_times() print(F'The time for naive implementation is {times[0]}.') print(F'The time for optimized implementation is {times[1]}.')
336
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device snake_case_ : Optional[Any] = False class __snake_case ( unittest.TestCase ): pass @nightly @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase ( self : int): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''') # remove text_unet pipe.remove_unused_weights() pipe.to(_snake_case) pipe.set_progress_bar_config(disable=_snake_case) UpperCAmelCase_ = '''A painting of a squirrel eating a burger ''' UpperCAmelCase_ = torch.manual_seed(0) UpperCAmelCase_ = pipe( prompt=_snake_case , generator=_snake_case , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''').images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_snake_case) UpperCAmelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained(_snake_case) pipe.to(_snake_case) pipe.set_progress_bar_config(disable=_snake_case) UpperCAmelCase_ = generator.manual_seed(0) UpperCAmelCase_ = pipe( prompt=_snake_case , generator=_snake_case , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''').images assert np.abs(image - new_image).sum() < 1e-5, "Models don't have the same forward pass" def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa) pipe.to(_snake_case) pipe.set_progress_bar_config(disable=_snake_case) UpperCAmelCase_ = '''A painting of a squirrel eating a burger ''' UpperCAmelCase_ = torch.manual_seed(0) UpperCAmelCase_ = pipe( prompt=_snake_case , generator=_snake_case , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''').images UpperCAmelCase_ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase_ = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
51
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> Tuple: """simple docstring""" if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=_SCREAMING_SNAKE_CASE , ) assert hasattr(self , "env") def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any=1) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=_SCREAMING_SNAKE_CASE , instance_type=self.instance_type , debugger_hook_config=_SCREAMING_SNAKE_CASE , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[Any]) -> Optional[Any]: """simple docstring""" TrainingJobAnalytics(_SCREAMING_SNAKE_CASE).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""") def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Tuple = self.create_estimator() # run training estimator.fit() # result dataframe __lowerCAmelCase : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis __lowerCAmelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) __lowerCAmelCase : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase : Tuple = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 99_9999) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , _SCREAMING_SNAKE_CASE)
269
0
"""simple docstring""" import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _a (unittest.TestCase ): '''simple docstring''' def __init__( self , A__ , A__=7 , A__=3 , A__=18 , A__=30 , A__=400 , A__=True , A__=None , A__=True , A__=None , A__=True , A__=[0.5, 0.5, 0.5] , A__=[0.5, 0.5, 0.5] , A__=False , ): A__ : int = size if size is not None else {'height': 20, 'width': 20} A__ : int = crop_size if crop_size is not None else {'height': 18, 'width': 18} A__ : Union[str, Any] = parent A__ : Any = batch_size A__ : List[str] = num_channels A__ : List[str] = image_size A__ : Optional[Any] = min_resolution A__ : Optional[int] = max_resolution A__ : Union[str, Any] = do_resize A__ : Any = size A__ : int = do_center_crop A__ : int = crop_size A__ : Union[str, Any] = do_normalize A__ : Optional[int] = image_mean A__ : int = image_std A__ : Tuple = do_reduce_labels def __A ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def UpperCamelCase () -> List[Any]: A__ : Union[str, Any] = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A__ : Union[str, Any] = Image.open(dataset[0]["""file"""] ) A__ : Any = Image.open(dataset[1]["""file"""] ) return image, map def UpperCamelCase () -> Union[str, Any]: A__ : int = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A__ : str = Image.open(ds[0]["""file"""] ) A__ : List[str] = Image.open(ds[1]["""file"""] ) A__ : int = Image.open(ds[2]["""file"""] ) A__ : Dict = Image.open(ds[3]["""file"""] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class _a (SCREAMING_SNAKE_CASE_ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: Union[str, Any] = BeitImageProcessor if is_vision_available() else None def __A ( self ): A__ : str = BeitImageProcessingTester(self ) @property def __A ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __A ( self ): A__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case__ , """size""" ) ) self.assertTrue(hasattr(snake_case__ , """do_center_crop""" ) ) self.assertTrue(hasattr(snake_case__ , """center_crop""" ) ) self.assertTrue(hasattr(snake_case__ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case__ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case__ , """image_std""" ) ) def __A ( self ): A__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 20, """width""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) self.assertEqual(image_processor.do_reduce_labels , snake_case__ ) A__ : List[str] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=snake_case__ ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) self.assertEqual(image_processor.do_reduce_labels , snake_case__ ) def __A ( self ): pass def __A ( self ): A__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A__ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input A__ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : Tuple = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __A ( self ): A__ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A__ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , numpify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , np.ndarray ) # Test not batched input A__ : str = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : Union[str, Any] = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __A ( self ): A__ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A__ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , torchify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) # Test not batched input A__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : Union[str, Any] = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __A ( self ): A__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A__ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , torchify=snake_case__ ) A__ : List[Any] = [] for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input A__ : List[Any] = image_processing(image_inputs[0] , maps[0] , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched A__ : List[Any] = image_processing(snake_case__ , snake_case__ , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test not batched input (PIL images) A__ : List[Any] = prepare_semantic_single_inputs() A__ : Any = image_processing(snake_case__ , snake_case__ , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched input (PIL images) A__ : Optional[int] = prepare_semantic_batch_inputs() A__ : Dict = image_processing(snake_case__ , snake_case__ , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 2, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) def __A ( self ): A__ : str = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 A__ : Optional[int] = prepare_semantic_single_inputs() A__ : int = image_processing(snake_case__ , snake_case__ , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 150 ) A__ : Union[str, Any] = True A__ : Any = image_processing(snake_case__ , snake_case__ , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 )
353
from queue import PriorityQueue from typing import Any import numpy as np def UpperCamelCase (lowercase_: dict , lowercase_: str , lowercase_: set , lowercase_: set , lowercase_: dict , lowercase_: dict , lowercase_: PriorityQueue , lowercase_: dict , lowercase_: float | int , ) -> float | int: for nxt, d in graph[v]: if nxt in visited_forward: continue A__ : Any = cst_fwd.get(lowercase_ , np.inf ) A__ : List[Any] = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) A__ : Tuple = new_cost_f A__ : Any = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: A__ : Optional[int] = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def UpperCamelCase (lowercase_: str , lowercase_: str , lowercase_: dict , lowercase_: dict ) -> int: A__ : Dict = -1 A__ : List[Any] = set() A__ : Union[str, Any] = set() A__ : Optional[Any] = {source: 0} A__ : int = {destination: 0} A__ : Optional[Any] = {source: None} A__ : Union[str, Any] = {destination: None} A__ : PriorityQueue[Any] = PriorityQueue() A__ : PriorityQueue[Any] = PriorityQueue() A__ : List[Any] = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): A__ , A__ : Tuple = queue_forward.get() visited_forward.add(lowercase_ ) A__ , A__ : Optional[Any] = queue_backward.get() visited_backward.add(lowercase_ ) A__ : List[Any] = pass_and_relaxation( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) A__ : List[Any] = pass_and_relaxation( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: A__ : int = shortest_distance return shortest_path_distance A_ : List[Any] = { 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } A_ : Optional[int] = { 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
141
0
"""simple docstring""" import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt"""} lowercase__ = { """vocab_file""": { """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/vocab.json""", """allenai/longformer-large-4096""": ( """https://huggingface.co/allenai/longformer-large-4096/resolve/main/vocab.json""" ), """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/vocab.json""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/vocab.json""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/vocab.json""" ), }, """merges_file""": { """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/merges.txt""", """allenai/longformer-large-4096""": ( """https://huggingface.co/allenai/longformer-large-4096/resolve/main/merges.txt""" ), """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/merges.txt""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/merges.txt""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/merges.txt""" ), }, } lowercase__ = { """allenai/longformer-base-4096""": 4096, """allenai/longformer-large-4096""": 4096, """allenai/longformer-large-4096-finetuned-triviaqa""": 4096, """allenai/longformer-base-4096-extra.pos.embd.only""": 4096, """allenai/longformer-large-4096-extra.pos.embd.only""": 4096, } @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def __lowerCamelCase ( ) -> int: """simple docstring""" lowerCAmelCase_ : Union[str, Any] = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) lowerCAmelCase_ : List[Any] = bs[:] lowerCAmelCase_ : Union[str, Any] = 0 for b in range(2**8 ): if b not in bs: bs.append(__UpperCamelCase ) cs.append(2**8 + n ) n += 1 lowerCAmelCase_ : Union[str, Any] = [chr(__UpperCamelCase ) for n in cs] return dict(zip(__UpperCamelCase , __UpperCamelCase ) ) def __lowerCamelCase ( __UpperCamelCase ) -> int: """simple docstring""" lowerCAmelCase_ : Dict = set() lowerCAmelCase_ : Optional[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase_ : List[Any] = char return pairs class __lowerCamelCase ( A__ ): '''simple docstring''' a_ : Optional[int] = VOCAB_FILES_NAMES a_ : int = PRETRAINED_VOCAB_FILES_MAP a_ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ : int = ["""input_ids""", """attention_mask"""] def __init__( self : int , a_ : int , a_ : Optional[Any] , a_ : Tuple="replace" , a_ : Optional[int]="<s>" , a_ : List[str]="</s>" , a_ : Optional[int]="</s>" , a_ : Any="<s>" , a_ : Union[str, Any]="<unk>" , a_ : Any="<pad>" , a_ : Optional[int]="<mask>" , a_ : List[str]=False , **a_ : str , ): lowerCAmelCase_ : Tuple = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else bos_token lowerCAmelCase_ : List[str] = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else eos_token lowerCAmelCase_ : Union[str, Any] = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else sep_token lowerCAmelCase_ : Dict = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else cls_token lowerCAmelCase_ : Optional[Any] = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else unk_token lowerCAmelCase_ : List[str] = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : Any = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else mask_token super().__init__( errors=a_ , bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , cls_token=a_ , pad_token=a_ , mask_token=a_ , add_prefix_space=a_ , **a_ , ) with open(a_ , encoding="utf-8" ) as vocab_handle: lowerCAmelCase_ : Optional[Any] = json.load(a_ ) lowerCAmelCase_ : Tuple = {v: k for k, v in self.encoder.items()} lowerCAmelCase_ : str = errors # how to handle errors in decoding lowerCAmelCase_ : Tuple = bytes_to_unicode() lowerCAmelCase_ : Dict = {v: k for k, v in self.byte_encoder.items()} with open(a_ , encoding="utf-8" ) as merges_handle: lowerCAmelCase_ : List[str] = merges_handle.read().split("\n" )[1:-1] lowerCAmelCase_ : Any = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase_ : List[Any] = dict(zip(a_ , range(len(a_ ) ) ) ) lowerCAmelCase_ : int = {} lowerCAmelCase_ : int = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase_ : Any = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property def lowerCamelCase ( self : str ): return len(self.encoder ) def lowerCamelCase ( self : Union[str, Any] ): return dict(self.encoder , **self.added_tokens_encoder ) def lowerCamelCase ( self : Tuple , a_ : List[Any] ): if token in self.cache: return self.cache[token] lowerCAmelCase_ : Optional[int] = tuple(a_ ) lowerCAmelCase_ : Optional[int] = get_pairs(a_ ) if not pairs: return token while True: lowerCAmelCase_ : Optional[int] = min(a_ , key=lambda a_ : self.bpe_ranks.get(a_ , float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = bigram lowerCAmelCase_ : Optional[int] = [] lowerCAmelCase_ : Union[str, Any] = 0 while i < len(a_ ): try: lowerCAmelCase_ : Tuple = word.index(a_ , a_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase_ : List[str] = j if word[i] == first and i < len(a_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase_ : str = tuple(a_ ) lowerCAmelCase_ : Optional[int] = new_word if len(a_ ) == 1: break else: lowerCAmelCase_ : str = get_pairs(a_ ) lowerCAmelCase_ : Optional[int] = " ".join(a_ ) lowerCAmelCase_ : Optional[int] = word return word def lowerCamelCase ( self : Union[str, Any] , a_ : Optional[int] ): lowerCAmelCase_ : int = [] for token in re.findall(self.pat , a_ ): lowerCAmelCase_ : Any = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a_ ).split(" " ) ) return bpe_tokens def lowerCamelCase ( self : List[str] , a_ : Optional[Any] ): return self.encoder.get(a_ , self.encoder.get(self.unk_token ) ) def lowerCamelCase ( self : List[str] , a_ : Optional[Any] ): return self.decoder.get(a_ ) def lowerCamelCase ( self : Optional[Any] , a_ : List[str] ): lowerCAmelCase_ : List[Any] = "".join(a_ ) lowerCAmelCase_ : Tuple = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def lowerCamelCase ( self : Optional[int] , a_ : str , a_ : Optional[str] = None ): if not os.path.isdir(a_ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase_ : Any = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : Dict = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(a_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a_ , ensure_ascii=a_ ) + "\n" ) lowerCAmelCase_ : int = 0 with open(a_ , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a_ : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowerCAmelCase_ : List[Any] = token_index writer.write(" ".join(a_ ) + "\n" ) index += 1 return vocab_file, merge_file def lowerCamelCase ( self : Optional[int] , a_ : List[int] , a_ : Optional[List[int]] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase_ : Any = [self.cls_token_id] lowerCAmelCase_ : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCamelCase ( self : Dict , a_ : List[int] , a_ : Optional[List[int]] = None , a_ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a_ , token_ids_a=a_ , already_has_special_tokens=a_ ) if token_ids_a is None: return [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1, 1] + ([0] * len(a_ )) + [1] def lowerCamelCase ( self : Tuple , a_ : List[int] , a_ : Optional[List[int]] = None ): lowerCAmelCase_ : Union[str, Any] = [self.sep_token_id] lowerCAmelCase_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCamelCase ( self : Optional[int] , a_ : Union[str, Any] , a_ : Optional[int]=False , **a_ : List[Any] ): lowerCAmelCase_ : List[str] = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a_ ) > 0 and not text[0].isspace()): lowerCAmelCase_ : List[Any] = " " + text return (text, kwargs)
241
"""simple docstring""" lowercase__ = 0 # The first color of the flag. lowercase__ = 1 # The second color of the flag. lowercase__ = 2 # The third color of the flag. lowercase__ = (red, white, blue) def __lowerCamelCase ( __UpperCamelCase ) -> list: """simple docstring""" if not sequence: return [] if len(__UpperCamelCase ) == 1: return list(__UpperCamelCase ) lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : List[Any] = len(__UpperCamelCase ) - 1 lowerCAmelCase_ : Union[str, Any] = 0 while mid <= high: if sequence[mid] == colors[0]: lowerCAmelCase_ , lowerCAmelCase_ : Any = sequence[mid], sequence[low] low += 1 mid += 1 elif sequence[mid] == colors[1]: mid += 1 elif sequence[mid] == colors[2]: lowerCAmelCase_ , lowerCAmelCase_ : str = sequence[high], sequence[mid] high -= 1 else: lowerCAmelCase_ : str = f'''The elements inside the sequence must contains only {colors} values''' raise ValueError(__UpperCamelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod() lowercase__ = input("""Enter numbers separated by commas:\n""").strip() lowercase__ = [int(item.strip()) for item in user_input.split(""",""")] print(F"""{dutch_national_flag_sort(unsorted)}""")
241
1
'''simple docstring''' import copy import random from transformers import CLIPTokenizer class UpperCamelCase ( a_ ): """simple docstring""" def __init__( self : Union[str, Any] , *UpperCAmelCase_ : Dict , **UpperCAmelCase_ : Any): """simple docstring""" super().__init__(*UpperCAmelCase_ , **UpperCAmelCase_) a : str = {} def SCREAMING_SNAKE_CASE_ ( self : int , UpperCAmelCase_ : Tuple , *UpperCAmelCase_ : Optional[int] , **UpperCAmelCase_ : int): """simple docstring""" a : Dict = super().add_tokens(UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_) if num_added_tokens == 0: raise ValueError( f"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ' `placeholder_token` that is not already in the tokenizer.') def SCREAMING_SNAKE_CASE_ ( self : Dict , UpperCAmelCase_ : Optional[int] , *UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str]=1 , **UpperCAmelCase_ : Optional[int]): """simple docstring""" a : Any = [] if num_vec_per_token == 1: self.try_adding_tokens(UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_) output.append(UpperCAmelCase_) else: a : int = [] for i in range(UpperCAmelCase_): a : Union[str, Any] = placeholder_token + f"""_{i}""" self.try_adding_tokens(UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_) output.append(UpperCAmelCase_) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( f"""The tokenizer already has placeholder token {token} that can get confused with""" f""" {placeholder_token}keep placeholder tokens independent""") a : Any = output def SCREAMING_SNAKE_CASE_ ( self : str , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Optional[int]=False , UpperCAmelCase_ : str=1.0): """simple docstring""" if isinstance(UpperCAmelCase_ , UpperCAmelCase_): a : Any = [] for i in range(len(UpperCAmelCase_)): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=UpperCAmelCase_)) return output for placeholder_token in self.token_map: if placeholder_token in text: a : List[Any] = self.token_map[placeholder_token] a : int = tokens[: 1 + int(len(UpperCAmelCase_) * prop_tokens_to_load)] if vector_shuffle: a : List[Any] = copy.copy(UpperCAmelCase_) random.shuffle(UpperCAmelCase_) a : List[str] = text.replace(UpperCAmelCase_ , ' '.join(UpperCAmelCase_)) return text def __call__( self : Optional[int] , UpperCAmelCase_ : Any , *UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Union[str, Any]=False , UpperCAmelCase_ : Optional[int]=1.0 , **UpperCAmelCase_ : str): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( UpperCAmelCase_ , vector_shuffle=UpperCAmelCase_ , prop_tokens_to_load=UpperCAmelCase_) , *UpperCAmelCase_ , **UpperCAmelCase_ , ) def SCREAMING_SNAKE_CASE_ ( self : Dict , UpperCAmelCase_ : Optional[int] , *UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Any=False , UpperCAmelCase_ : Optional[Any]=1.0 , **UpperCAmelCase_ : Dict): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( UpperCAmelCase_ , vector_shuffle=UpperCAmelCase_ , prop_tokens_to_load=UpperCAmelCase_) , *UpperCAmelCase_ , **UpperCAmelCase_ , )
345
'''simple docstring''' from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class UpperCamelCase ( a_ ): """simple docstring""" A : Optional[int] = ["vqvae"] def __init__( self : List[str] , UpperCAmelCase_ : AutoencoderKL , UpperCAmelCase_ : UNetaDConditionModel , UpperCAmelCase_ : Mel , UpperCAmelCase_ : Union[DDIMScheduler, DDPMScheduler] , ): """simple docstring""" super().__init__() self.register_modules(unet=UpperCAmelCase_ , scheduler=UpperCAmelCase_ , mel=UpperCAmelCase_ , vqvae=UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : List[str]): """simple docstring""" return 5_0 if isinstance(self.scheduler , UpperCAmelCase_) else 1_0_0_0 @torch.no_grad() def __call__( self : Dict , UpperCAmelCase_ : int = 1 , UpperCAmelCase_ : str = None , UpperCAmelCase_ : np.ndarray = None , UpperCAmelCase_ : int = 0 , UpperCAmelCase_ : int = 0 , UpperCAmelCase_ : int = None , UpperCAmelCase_ : torch.Generator = None , UpperCAmelCase_ : float = 0 , UpperCAmelCase_ : float = 0 , UpperCAmelCase_ : torch.Generator = None , UpperCAmelCase_ : float = 0 , UpperCAmelCase_ : torch.Tensor = None , UpperCAmelCase_ : torch.Tensor = None , UpperCAmelCase_ : Optional[Any]=True , ): """simple docstring""" a : Optional[Any] = steps or self.get_default_steps() self.scheduler.set_timesteps(UpperCAmelCase_) a : Optional[Any] = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size) == int: a : Dict = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: a : Dict = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=UpperCAmelCase_ , device=self.device , ) a : Tuple = noise a : Optional[int] = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(UpperCAmelCase_ , UpperCAmelCase_) a : List[Any] = self.mel.audio_slice_to_image(UpperCAmelCase_) a : str = np.frombuffer(input_image.tobytes() , dtype='uint8').reshape( (input_image.height, input_image.width)) a : List[str] = (input_image / 2_5_5) * 2 - 1 a : Any = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float).to(self.device) if self.vqvae is not None: a : List[Any] = self.vqvae.encode(torch.unsqueeze(UpperCAmelCase_ , 0)).latent_dist.sample( generator=UpperCAmelCase_)[0] a : str = self.vqvae.config.scaling_factor * input_images if start_step > 0: a : Union[str, Any] = self.scheduler.add_noise(UpperCAmelCase_ , UpperCAmelCase_ , self.scheduler.timesteps[start_step - 1]) a : Dict = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) a : List[Any] = int(mask_start_secs * pixels_per_second) a : Optional[Any] = int(mask_end_secs * pixels_per_second) a : Optional[int] = self.scheduler.add_noise(UpperCAmelCase_ , UpperCAmelCase_ , torch.tensor(self.scheduler.timesteps[start_step:])) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:])): if isinstance(self.unet , UpperCAmelCase_): a : Dict = self.unet(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_)['sample'] else: a : str = self.unet(UpperCAmelCase_ , UpperCAmelCase_)['sample'] if isinstance(self.scheduler , UpperCAmelCase_): a : List[Any] = self.scheduler.step( model_output=UpperCAmelCase_ , timestep=UpperCAmelCase_ , sample=UpperCAmelCase_ , eta=UpperCAmelCase_ , generator=UpperCAmelCase_ , )['prev_sample'] else: a : Any = self.scheduler.step( model_output=UpperCAmelCase_ , timestep=UpperCAmelCase_ , sample=UpperCAmelCase_ , generator=UpperCAmelCase_ , )['prev_sample'] if mask is not None: if mask_start > 0: a : str = mask[:, step, :, :mask_start] if mask_end > 0: a : Dict = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance a : List[str] = 1 / self.vqvae.config.scaling_factor * images a : str = self.vqvae.decode(UpperCAmelCase_)['sample'] a : Tuple = (images / 2 + 0.5).clamp(0 , 1) a : Any = images.cpu().permute(0 , 2 , 3 , 1).numpy() a : List[str] = (images * 2_5_5).round().astype('uint8') a : Tuple = list( (Image.fromarray(_[:, :, 0]) for _ in images) if images.shape[3] == 1 else (Image.fromarray(UpperCAmelCase_ , mode='RGB').convert('L') for _ in images)) a : List[str] = [self.mel.image_to_audio(UpperCAmelCase_) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(UpperCAmelCase_)[:, np.newaxis, :]) , **ImagePipelineOutput(UpperCAmelCase_)) @torch.no_grad() def SCREAMING_SNAKE_CASE_ ( self : List[Any] , UpperCAmelCase_ : List[Image.Image] , UpperCAmelCase_ : int = 5_0): """simple docstring""" assert isinstance(self.scheduler , UpperCAmelCase_) self.scheduler.set_timesteps(UpperCAmelCase_) a : Dict = np.array( [np.frombuffer(image.tobytes() , dtype='uint8').reshape((1, image.height, image.width)) for image in images]) a : Tuple = (sample / 2_5_5) * 2 - 1 a : int = torch.Tensor(UpperCAmelCase_).to(self.device) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,))): a : Optional[Any] = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps a : Optional[Any] = self.scheduler.alphas_cumprod[t] a : List[Any] = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) a : List[str] = 1 - alpha_prod_t a : Optional[Any] = self.unet(UpperCAmelCase_ , UpperCAmelCase_)['sample'] a : Union[str, Any] = (1 - alpha_prod_t_prev) ** 0.5 * model_output a : Dict = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) a : Union[str, Any] = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : torch.Tensor , UpperCAmelCase_ : torch.Tensor , UpperCAmelCase_ : float): """simple docstring""" a : List[Any] = acos(torch.dot(torch.flatten(UpperCAmelCase_) , torch.flatten(UpperCAmelCase_)) / torch.norm(UpperCAmelCase_) / torch.norm(UpperCAmelCase_)) return sin((1 - alpha) * theta) * xa / sin(UpperCAmelCase_) + sin(alpha * theta) * xa / sin(UpperCAmelCase_)
345
1
"""simple docstring""" import heapq as hq import math from collections.abc import Iterator class lowerCamelCase__ : def __init__( self , SCREAMING_SNAKE_CASE ): """simple docstring""" snake_case : List[Any] = str(id_ ) snake_case : Optional[Any] = None snake_case : List[Any] = None snake_case : str = [] snake_case : Optional[Any] = {} # {vertex:distance} def __lt__( self , SCREAMING_SNAKE_CASE ): """simple docstring""" return self.key < other.key def __repr__( self ): """simple docstring""" return self.id def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE ): """simple docstring""" self.neighbors.append(A__ ) def lowerCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" snake_case : str = weight def UpperCamelCase__ ( lowercase__ : List[str] , lowercase__ : str , lowercase__ : int , lowercase__ : Optional[Any] ): graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , lowerCAmelCase__ ) graph[b - 1].add_edge(graph[a - 1] , lowerCAmelCase__ ) def UpperCamelCase__ ( lowercase__ : List[Any] , lowercase__ : str ): snake_case : Optional[int] = [] for u in graph: snake_case : Tuple = math.inf snake_case : Any = None snake_case : Union[str, Any] = 0 snake_case : str = graph[:] while q: snake_case : str = min(lowerCAmelCase__ ) q.remove(lowerCAmelCase__ ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): snake_case : List[str] = u snake_case : Tuple = u.edges[v.id] for i in range(1 , len(lowerCAmelCase__ ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def UpperCamelCase__ ( lowercase__ : Tuple , lowercase__ : Dict ): for u in graph: snake_case : Any = math.inf snake_case : Union[str, Any] = None snake_case : str = 0 snake_case : List[Any] = list(lowerCAmelCase__ ) hq.heapify(lowerCAmelCase__ ) while h: snake_case : Tuple = hq.heappop(lowerCAmelCase__ ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): snake_case : Optional[int] = u snake_case : List[Any] = u.edges[v.id] hq.heapify(lowerCAmelCase__ ) for i in range(1 , len(lowerCAmelCase__ ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def UpperCamelCase__ ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
148
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() lowercase__ :str = 2 class lowercase : def __init__( self ,*, # begin keyword-only arguments A__="<s>" ,A__="<pad>" ,A__="</s>" ,A__="<unk>" ,A__=None ,): lowercase , lowercase , lowercase , lowercase = bos, unk, pad, eos lowercase = [] lowercase = [] lowercase = {} lowercase = self.add_symbol(A__) lowercase = self.add_symbol(A__) lowercase = self.add_symbol(A__) lowercase = self.add_symbol(A__) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(A__) lowercase = len(self.symbols) def __eq__( self ,A__): return self.indices == other.indices def __getitem__( self ,A__): if idx < len(self.symbols): return self.symbols[idx] return self.unk_word def __len__( self): return len(self.symbols) def __contains__( self ,A__): return sym in self.indices @classmethod def A__ ( cls ,A__): lowercase = cls() d.add_from_file(A__) return d def A__ ( self ,A__ ,A__=1 ,A__=False): if word in self.indices and not overwrite: lowercase = self.indices[word] lowercase = self.count[idx] + n return idx else: lowercase = len(self.symbols) lowercase = idx self.symbols.append(A__) self.count.append(A__) return idx def A__ ( self ,A__): return 0 def A__ ( self ,A__): if isinstance(A__ ,A__): try: with open(A__ ,'''r''' ,encoding='''utf-8''') as fd: self.add_from_file(A__) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('''Incorrect encoding detected in {}, please rebuild the dataset'''.format(A__)) return lowercase = f.readlines() lowercase = self._load_meta(A__) for line in lines[indices_start_line:]: try: lowercase , lowercase = line.rstrip().rsplit(''' ''' ,1) if field == "#fairseq:overwrite": lowercase = True lowercase , lowercase = line.rsplit(''' ''' ,1) else: lowercase = False lowercase = int(A__) lowercase = line if word in self and not overwrite: raise RuntimeError( '''Duplicate word found when loading Dictionary: \'{}\'. ''' '''Duplicate words can overwrite earlier ones by adding the ''' '''#fairseq:overwrite flag at the end of the corresponding row ''' '''in the dictionary file. If using the Camembert model, please ''' '''download an updated copy of the model file.'''.format(A__)) self.add_symbol(A__ ,n=A__ ,overwrite=A__) except ValueError: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt> [flags]\'''') def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} lowercase = dict((re.sub(R'''@@$''' , '''''' , lowerCAmelCase__ ), v) if k.endswith('''@@''' ) else (re.sub(R'''$''' , '''</w>''' , lowerCAmelCase__ ), v) for k, v in d.items() ) lowercase = '''<s> <pad> </s> <unk>'''.split() # restore the special tokens for k in keep_keys: del da[f'{k}</w>'] lowercase = d[k] # restore return da def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # prep if not os.path.exists(lowerCAmelCase__ ): raise ValueError(f'path {biogpt_checkpoint_path} does not exist!' ) os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) print(f'Writing results to {pytorch_dump_folder_path}' ) # handle various types of models lowercase = os.path.join(lowerCAmelCase__ , '''checkpoint.pt''' ) if not os.path.isfile(lowerCAmelCase__ ): raise ValueError(f'path to the file {checkpoint_file} does not exist!' ) lowercase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowercase = chkpt['''cfg''']['''model'''] # dicts lowercase = os.path.join(lowerCAmelCase__ , '''dict.txt''' ) if not os.path.isfile(lowerCAmelCase__ ): raise ValueError(f'path to the file {dict_file} does not exist!' ) lowercase = Dictionary.load(lowerCAmelCase__ ) lowercase = rewrite_dict_keys(src_dict.indices ) lowercase = len(lowerCAmelCase__ ) lowercase = os.path.join(lowerCAmelCase__ , VOCAB_FILES_NAMES['''vocab_file'''] ) print(f'Generating {src_vocab_file} of {src_vocab_size} records' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ , indent=lowerCAmelCase__ ) ) # merges_file (bpecodes) lowercase = os.path.join(lowerCAmelCase__ , '''bpecodes''' ) if not os.path.isfile(lowerCAmelCase__ ): raise ValueError(f'path to the file {bpecodes_file} does not exist!' ) lowercase = os.path.join(lowerCAmelCase__ , VOCAB_FILES_NAMES['''merges_file'''] ) shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) # model config lowercase = os.path.join(lowerCAmelCase__ , '''config.json''' ) lowercase = { '''activation_dropout''': args['''activation_dropout'''], '''architectures''': ['''BioGptForCausalLM'''], '''attention_probs_dropout_prob''': args['''attention_dropout'''], '''bos_token_id''': 0, '''eos_token_id''': 2, '''hidden_act''': args['''activation_fn'''], '''hidden_dropout_prob''': args['''dropout'''], '''hidden_size''': args['''decoder_embed_dim'''], '''initializer_range''': 0.02, '''intermediate_size''': args['''decoder_ffn_embed_dim'''], '''layer_norm_eps''': 1E-12, '''layerdrop''': args['''decoder_layerdrop'''], '''max_position_embeddings''': args['''max_target_positions'''], '''model_type''': '''biogpt''', '''num_attention_heads''': args['''decoder_attention_heads'''], '''num_hidden_layers''': args['''decoder_layers'''], '''pad_token_id''': 1, '''scale_embedding''': not args['''no_scale_embedding'''], '''tie_word_embeddings''': args['''share_decoder_input_output_embed'''], '''vocab_size''': src_vocab_size, } # good hparam defaults to start with print(f'Generating {biogpt_model_config_file}' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ , indent=lowerCAmelCase__ ) ) # tokenizer config lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = { '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', '''model_max_length''': 1024, '''pad_token''': '''<pad>''', '''special_tokens_map_file''': None, '''tokenizer_class''': '''BioGptTokenizer''', '''unk_token''': '''<unk>''', } print(f'Generating {biogpt_tokenizer_config_file}' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ , indent=lowerCAmelCase__ ) ) # model lowercase = chkpt['''model'''] # remove unneeded keys lowercase = [ '''decoder.version''', ] for k in ignore_keys: model_state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('''output_projection.weight''' ): lowercase = model_state_dict.pop(lowerCAmelCase__ ) else: lowercase = model_state_dict.pop(lowerCAmelCase__ ) lowercase = BioGptConfig.from_pretrained(lowerCAmelCase__ ) lowercase = BioGptForCausalLM(lowerCAmelCase__ ) # check that it loads ok model_new.load_state_dict(lowerCAmelCase__ ) # save lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) print(f'Generating {pytorch_weights_dump_path}' ) torch.save(lowerCAmelCase__ , lowerCAmelCase__ ) print('''Conversion is done!''' ) if __name__ == "__main__": lowercase__ :Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--biogpt_checkpoint_path", default=None, type=str, required=True, help=( "Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts," " bpecodes, etc." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) lowercase__ :Any = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
101
0
"""simple docstring""" from __future__ import annotations def a__ ( __UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = sorted(numsa + numsa ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = divmod(len(__UpperCamelCase ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() A : Tuple = [float(x) for x in input("Enter the elements of first array: ").split()] A : List[str] = [float(x) for x in input("Enter the elements of second array: ").split()] print(f"The median of two arrays is: {median_of_two_arrays(array_a, array_a)}")
350
from __future__ import annotations import numpy as np def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = np.shape(__UpperCamelCase ) if rows != columns: SCREAMING_SNAKE_CASE_ = ( "'table' has to be of square shaped array but got a " F'''{rows}x{columns} array:\n{table}''' ) raise ValueError(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = np.zeros((rows, columns) ) SCREAMING_SNAKE_CASE_ = np.zeros((rows, columns) ) for i in range(__UpperCamelCase ): for j in range(__UpperCamelCase ): SCREAMING_SNAKE_CASE_ = sum(lower[i][k] * upper[k][j] for k in range(__UpperCamelCase ) ) if upper[j][j] == 0: raise ArithmeticError("No LU decomposition exists" ) SCREAMING_SNAKE_CASE_ = (table[i][j] - total) / upper[j][j] SCREAMING_SNAKE_CASE_ = 1 for j in range(__UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = sum(lower[i][k] * upper[k][j] for k in range(__UpperCamelCase ) ) SCREAMING_SNAKE_CASE_ = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
305
0
'''simple docstring''' import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> List[str]: lowerCamelCase__ : Optional[int] = SwinConfig() lowerCamelCase__ : Any = swin_name.split("""_""" ) lowerCamelCase__ : str = name_split[1] lowerCamelCase__ : int = int(name_split[4] ) lowerCamelCase__ : Tuple = int(name_split[3][-1] ) if model_size == "tiny": lowerCamelCase__ : Any = 96 lowerCamelCase__ : Optional[Any] = (2, 2, 6, 2) lowerCamelCase__ : Optional[Any] = (3, 6, 12, 24) elif model_size == "small": lowerCamelCase__ : List[Any] = 96 lowerCamelCase__ : Dict = (2, 2, 18, 2) lowerCamelCase__ : Optional[int] = (3, 6, 12, 24) elif model_size == "base": lowerCamelCase__ : Tuple = 128 lowerCamelCase__ : Union[str, Any] = (2, 2, 18, 2) lowerCamelCase__ : Optional[int] = (4, 8, 16, 32) else: lowerCamelCase__ : int = 192 lowerCamelCase__ : List[str] = (2, 2, 18, 2) lowerCamelCase__ : Any = (6, 12, 24, 48) if "in22k" in swin_name: lowerCamelCase__ : int = 21841 else: lowerCamelCase__ : List[str] = 1000 lowerCamelCase__ : Tuple = """huggingface/label-files""" lowerCamelCase__ : Dict = """imagenet-1k-id2label.json""" lowerCamelCase__ : Tuple = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Optional[int] = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : List[str] = idalabel lowerCamelCase__ : List[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase__ : str = img_size lowerCamelCase__ : Any = num_classes lowerCamelCase__ : List[Any] = embed_dim lowerCamelCase__ : Union[str, Any] = depths lowerCamelCase__ : Any = num_heads lowerCamelCase__ : Dict = window_size return config def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> List[Any]: if "patch_embed.proj" in name: lowerCamelCase__ : int = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: lowerCamelCase__ : str = name.replace("""patch_embed.norm""" , """embeddings.norm""" ) if "layers" in name: lowerCamelCase__ : str = """encoder.""" + name if "attn.proj" in name: lowerCamelCase__ : List[str] = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: lowerCamelCase__ : Tuple = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: lowerCamelCase__ : List[str] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: lowerCamelCase__ : Optional[Any] = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: lowerCamelCase__ : Optional[int] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: lowerCamelCase__ : List[str] = name.replace("""mlp.fc2""" , """output.dense""" ) if name == "norm.weight": lowerCamelCase__ : Union[str, Any] = """layernorm.weight""" if name == "norm.bias": lowerCamelCase__ : Tuple = """layernorm.bias""" if "head" in name: lowerCamelCase__ : Optional[Any] = name.replace("""head""" , """classifier""" ) else: lowerCamelCase__ : List[str] = """swin.""" + name return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[int]: for key in orig_state_dict.copy().keys(): lowerCamelCase__ : Optional[Any] = orig_state_dict.pop(UpperCamelCase ) if "mask" in key: continue elif "qkv" in key: lowerCamelCase__ : List[Any] = key.split(""".""" ) lowerCamelCase__ : List[str] = int(key_split[1] ) lowerCamelCase__ : List[Any] = int(key_split[3] ) lowerCamelCase__ : Optional[int] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowerCamelCase__ : Any = val[:dim, :] lowerCamelCase__ : Tuple = val[ dim : dim * 2, : ] lowerCamelCase__ : Tuple = val[-dim:, :] else: lowerCamelCase__ : int = val[ :dim ] lowerCamelCase__ : Any = val[ dim : dim * 2 ] lowerCamelCase__ : Any = val[ -dim: ] else: lowerCamelCase__ : Tuple = val return orig_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : Union[str, Any] = timm.create_model(UpperCamelCase , pretrained=UpperCamelCase ) timm_model.eval() lowerCamelCase__ : Union[str, Any] = get_swin_config(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = SwinForImageClassification(UpperCamelCase ) model.eval() lowerCamelCase__ : Optional[int] = convert_state_dict(timm_model.state_dict() , UpperCamelCase ) model.load_state_dict(UpperCamelCase ) lowerCamelCase__ : Dict = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Tuple = AutoImageProcessor.from_pretrained("""microsoft/{}""".format(swin_name.replace("""_""" , """-""" ) ) ) lowerCamelCase__ : Union[str, Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Tuple = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : List[str] = timm_model(inputs["""pixel_values"""] ) lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase ).logits assert torch.allclose(UpperCamelCase , UpperCamelCase , atol=1E-3 ) print(f'''Saving model {swin_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : List[str] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swin_name''', default='''swin_tiny_patch4_window7_224''', type=str, help='''Name of the Swin timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _A : int =parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
41
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = [ ('''bert.bert''', '''visual_bert'''), ('''bert.cls''', '''cls'''), ('''bert.classifier''', '''cls'''), ('''token_type_embeddings_visual''', '''visual_token_type_embeddings'''), ('''position_embeddings_visual''', '''visual_position_embeddings'''), ('''projection''', '''visual_projection'''), ] _UpperCamelCase = [ '''nlvr2_coco_pre_trained.th''', '''nlvr2_fine_tuned.th''', '''nlvr2_pre_trained.th''', '''vcr_coco_pre_train.th''', '''vcr_fine_tune.th''', '''vcr_pre_train.th''', '''vqa_coco_pre_trained.th''', '''vqa_fine_tuned.th''', '''vqa_pre_trained.th''', ] def lowerCAmelCase__( lowercase : str ) -> Optional[Any]: __snake_case : Optional[int] = torch.load(lowercase , map_location="cpu" ) return sd def lowerCAmelCase__( lowercase : List[Any] , lowercase : List[Any] , lowercase : List[Any]=rename_keys_prefix ) -> Dict: __snake_case : Tuple = OrderedDict() __snake_case : str = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue __snake_case : Optional[Any] = key for name_pair in rename_keys_prefix: __snake_case : List[str] = new_key.replace(name_pair[0] , name_pair[1] ) __snake_case : List[str] = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately __snake_case : List[Any] = new_d["cls.predictions.bias"] return new_d @torch.no_grad() def lowerCAmelCase__( lowercase : Optional[Any] , lowercase : Any ) -> List[Any]: assert ( checkpoint_path.split("/" )[-1] in ACCEPTABLE_CHECKPOINTS ), f"""The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.""" # Get Config if "pre" in checkpoint_path: __snake_case : Any = "pretraining" if "vcr" in checkpoint_path: __snake_case : Optional[Any] = {"visual_embedding_dim": 512} elif "vqa_advanced" in checkpoint_path: __snake_case : Tuple = {"visual_embedding_dim": 2048} elif "vqa" in checkpoint_path: __snake_case : Dict = {"visual_embedding_dim": 2048} elif "nlvr" in checkpoint_path: __snake_case : Any = {"visual_embedding_dim": 1024} else: raise NotImplementedError(f"""No implementation found for `{checkpoint_path}`.""" ) else: if "vcr" in checkpoint_path: __snake_case : Dict = {"visual_embedding_dim": 512} __snake_case : Any = "multichoice" elif "vqa_advanced" in checkpoint_path: __snake_case : List[Any] = {"visual_embedding_dim": 2048} __snake_case : Optional[Any] = "vqa_advanced" elif "vqa" in checkpoint_path: __snake_case : Union[str, Any] = {"visual_embedding_dim": 2048, "num_labels": 3129} __snake_case : Union[str, Any] = "vqa" elif "nlvr" in checkpoint_path: __snake_case : Tuple = { "visual_embedding_dim": 1024, "num_labels": 2, } __snake_case : List[Any] = "nlvr" __snake_case : Union[str, Any] = VisualBertConfig(**lowercase ) # Load State Dict __snake_case : Any = load_state_dict(lowercase ) __snake_case : Dict = get_new_dict(lowercase , lowercase ) if model_type == "pretraining": __snake_case : Optional[Any] = VisualBertForPreTraining(lowercase ) elif model_type == "vqa": __snake_case : Tuple = VisualBertForQuestionAnswering(lowercase ) elif model_type == "nlvr": __snake_case : Tuple = VisualBertForVisualReasoning(lowercase ) elif model_type == "multichoice": __snake_case : List[Any] = VisualBertForMultipleChoice(lowercase ) model.load_state_dict(lowercase ) # Save Checkpoints Path(lowercase ).mkdir(exist_ok=lowercase ) model.save_pretrained(lowercase ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('''orig_checkpoint_path''', type=str, help='''A path to .th on local filesystem.''') parser.add_argument('''pytorch_dump_folder_path''', type=str, help='''Path to the output PyTorch model.''') _UpperCamelCase = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
326
0
'''simple docstring''' from decimal import Decimal, getcontext from math import ceil, factorial def snake_case_ ( lowerCAmelCase_ )-> str: '''simple docstring''' if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""Undefined for non-integers""" ) elif precision < 1: raise ValueError("""Undefined for non-natural numbers""" ) _UpperCAmelCase : int = precision _UpperCAmelCase : List[Any] = ceil(precision / 14 ) _UpperCAmelCase : Optional[int] = 426880 * Decimal(10005 ).sqrt() _UpperCAmelCase : List[str] = 1 _UpperCAmelCase : Union[str, Any] = 13591409 _UpperCAmelCase : List[Any] = Decimal(lowerCAmelCase_ ) for k in range(1 , lowerCAmelCase_ ): _UpperCAmelCase : str = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowerCAmelCase_ ) ** 3) linear_term += 545140134 exponential_term *= -262537412640768000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": A_ : Tuple = 5_0 print(f"""The first {n} digits of pi is: {pi(n)}""")
357
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A_ : Union[str, Any] = logging.get_logger(__name__) A_ : Any = { """hustvl/yolos-small""": """https://huggingface.co/hustvl/yolos-small/resolve/main/config.json""", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class lowercase ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = """yolos""" def __init__( self ,a_=768 ,a_=12 ,a_=12 ,a_=3_072 ,a_="gelu" ,a_=0.0 ,a_=0.0 ,a_=0.02 ,a_=1E-1_2 ,a_=[512, 864] ,a_=16 ,a_=3 ,a_=True ,a_=100 ,a_=True ,a_=False ,a_=1 ,a_=5 ,a_=2 ,a_=5 ,a_=2 ,a_=0.1 ,**a_ ,) -> List[str]: super().__init__(**a_ ) _UpperCAmelCase : Optional[Any] = hidden_size _UpperCAmelCase : Optional[Any] = num_hidden_layers _UpperCAmelCase : Tuple = num_attention_heads _UpperCAmelCase : Optional[Any] = intermediate_size _UpperCAmelCase : Union[str, Any] = hidden_act _UpperCAmelCase : List[str] = hidden_dropout_prob _UpperCAmelCase : Optional[int] = attention_probs_dropout_prob _UpperCAmelCase : List[Any] = initializer_range _UpperCAmelCase : Union[str, Any] = layer_norm_eps _UpperCAmelCase : int = image_size _UpperCAmelCase : Dict = patch_size _UpperCAmelCase : Tuple = num_channels _UpperCAmelCase : Optional[Any] = qkv_bias _UpperCAmelCase : List[Any] = num_detection_tokens _UpperCAmelCase : Tuple = use_mid_position_embeddings _UpperCAmelCase : int = auxiliary_loss # Hungarian matcher _UpperCAmelCase : Dict = class_cost _UpperCAmelCase : Dict = bbox_cost _UpperCAmelCase : Optional[int] = giou_cost # Loss coefficients _UpperCAmelCase : int = bbox_loss_coefficient _UpperCAmelCase : Optional[Any] = giou_loss_coefficient _UpperCAmelCase : Union[str, Any] = eos_coefficient class lowercase ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = version.parse("""1.11""" ) @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def _snake_case ( self ) -> float: return 1E-4 @property def _snake_case ( self ) -> int: return 12
349
0
'''simple docstring''' import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from transformers import OneFormerImageProcessor from transformers.models.oneformer.image_processing_oneformer import binary_mask_to_rle from transformers.models.oneformer.modeling_oneformer import OneFormerForUniversalSegmentationOutput if is_vision_available(): from PIL import Image def lowerCamelCase (_SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : List[str]="shi-labs/oneformer_demo" ): with open(hf_hub_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) as f: __a : List[Any] = json.load(_SCREAMING_SNAKE_CASE ) __a : str = {} __a : List[str] = [] __a : List[Any] = [] for key, info in class_info.items(): __a : List[Any] = info['name'] class_names.append(info['name'] ) if info["isthing"]: thing_ids.append(int(_SCREAMING_SNAKE_CASE ) ) __a : List[str] = thing_ids __a : str = class_names return metadata class __UpperCamelCase ( unittest.TestCase ): def __init__( self , __a , __a=7 , __a=3 , __a=30 , __a=400 , __a=None , __a=True , __a=True , __a=[0.5, 0.5, 0.5] , __a=[0.5, 0.5, 0.5] , __a=10 , __a=False , __a=255 , __a="shi-labs/oneformer_demo" , __a="ade20k_panoptic.json" , __a=10 , ): '''simple docstring''' __a : Union[str, Any] = parent __a : Tuple = batch_size __a : List[Any] = num_channels __a : Tuple = min_resolution __a : Union[str, Any] = max_resolution __a : List[str] = do_resize __a : Union[str, Any] = {'shortest_edge': 32, 'longest_edge': 1333} if size is None else size __a : Optional[Any] = do_normalize __a : Dict = image_mean __a : Union[str, Any] = image_std __a : Union[str, Any] = class_info_file __a : int = prepare_metadata(__a , __a ) __a : List[Any] = num_text __a : List[Any] = repo_path # for the post_process_functions __a : Optional[Any] = 2 __a : Dict = 10 __a : Tuple = 10 __a : int = 3 __a : int = 4 __a : Union[str, Any] = num_labels __a : Any = do_reduce_labels __a : Union[str, Any] = ignore_index def __UpperCAmelCase ( self ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "num_labels": self.num_labels, "do_reduce_labels": self.do_reduce_labels, "ignore_index": self.ignore_index, "class_info_file": self.class_info_file, "metadata": self.metadata, "num_text": self.num_text, } def __UpperCAmelCase ( self , __a , __a=False ): '''simple docstring''' if not batched: __a : int = image_inputs[0] if isinstance(__a , Image.Image ): __a , __a : List[Any] = image.size else: __a , __a : Optional[Any] = image.shape[1], image.shape[2] if w < h: __a : Optional[Any] = int(self.size['shortest_edge'] * h / w ) __a : str = self.size['shortest_edge'] elif w > h: __a : Any = self.size['shortest_edge'] __a : Dict = int(self.size['shortest_edge'] * w / h ) else: __a : Union[str, Any] = self.size['shortest_edge'] __a : List[Any] = self.size['shortest_edge'] else: __a : Optional[int] = [] for image in image_inputs: __a , __a : Any = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __a : Union[str, Any] = max(__a , key=lambda __a : item[0] )[0] __a : str = max(__a , key=lambda __a : item[1] )[1] return expected_height, expected_width def __UpperCAmelCase ( self ): '''simple docstring''' return OneFormerForUniversalSegmentationOutput( # +1 for null class class_queries_logits=torch.randn((self.batch_size, self.num_queries, self.num_classes + 1) ) , masks_queries_logits=torch.randn((self.batch_size, self.num_queries, self.height, self.width) ) , ) @require_torch @require_vision class __UpperCamelCase ( lowerCAmelCase_ , unittest.TestCase ): A_ = OneFormerImageProcessor if (is_vision_available() and is_torch_available()) else None # only for test_image_processing_common.test_image_proc_to_json_string A_ = image_processing_class def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[int] = OneFormerImageProcessorTester(self ) @property def __UpperCAmelCase ( self ): '''simple docstring''' return self.image_processing_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__a , 'image_mean' ) ) self.assertTrue(hasattr(__a , 'image_std' ) ) self.assertTrue(hasattr(__a , 'do_normalize' ) ) self.assertTrue(hasattr(__a , 'do_resize' ) ) self.assertTrue(hasattr(__a , 'size' ) ) self.assertTrue(hasattr(__a , 'ignore_index' ) ) self.assertTrue(hasattr(__a , 'class_info_file' ) ) self.assertTrue(hasattr(__a , 'num_text' ) ) self.assertTrue(hasattr(__a , 'repo_path' ) ) self.assertTrue(hasattr(__a , 'metadata' ) ) self.assertTrue(hasattr(__a , 'do_reduce_labels' ) ) def __UpperCAmelCase ( self ): '''simple docstring''' pass def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a : List[Any] = prepare_image_inputs(self.image_processing_tester , equal_resolution=__a ) for image in image_inputs: self.assertIsInstance(__a , Image.Image ) # Test not batched input __a : List[Any] = image_processor(image_inputs[0] , ['semantic'] , return_tensors='pt' ).pixel_values __a , __a : List[str] = self.image_processing_tester.get_expected_values(__a ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __a , __a : Optional[Any] = self.image_processing_tester.get_expected_values(__a , batched=__a ) __a : List[str] = image_processor( __a , ['semantic'] * len(__a ) , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a : int = prepare_image_inputs(self.image_processing_tester , equal_resolution=__a , numpify=__a ) for image in image_inputs: self.assertIsInstance(__a , np.ndarray ) # Test not batched input __a : str = image_processor(image_inputs[0] , ['semantic'] , return_tensors='pt' ).pixel_values __a , __a : Tuple = self.image_processing_tester.get_expected_values(__a ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __a , __a : Union[str, Any] = self.image_processing_tester.get_expected_values(__a , batched=__a ) __a : Any = image_processor( __a , ['semantic'] * len(__a ) , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a : Union[str, Any] = prepare_image_inputs(self.image_processing_tester , equal_resolution=__a , torchify=__a ) for image in image_inputs: self.assertIsInstance(__a , torch.Tensor ) # Test not batched input __a : List[Any] = image_processor(image_inputs[0] , ['semantic'] , return_tensors='pt' ).pixel_values __a , __a : int = self.image_processing_tester.get_expected_values(__a ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __a , __a : List[Any] = self.image_processing_tester.get_expected_values(__a , batched=__a ) __a : str = image_processor( __a , ['semantic'] * len(__a ) , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def __UpperCAmelCase ( self , __a=False , __a=False , __a="np" ): '''simple docstring''' __a : str = self.image_processing_class(**self.image_processor_dict ) # prepare image and target __a : Dict = self.image_processing_tester.num_labels __a : List[str] = None __a : str = None __a : Optional[int] = prepare_image_inputs(self.image_processing_tester , equal_resolution=__a ) if with_segmentation_maps: __a : List[str] = num_labels if is_instance_map: __a : Optional[int] = list(range(__a ) ) * 2 __a : int = dict(enumerate(__a ) ) __a : List[str] = [ np.random.randint(0 , high * 2 , (img.size[1], img.size[0]) ).astype(np.uinta ) for img in image_inputs ] if segmentation_type == "pil": __a : str = [Image.fromarray(__a ) for annotation in annotations] __a : Optional[Any] = image_processor( __a , ['semantic'] * len(__a ) , __a , return_tensors='pt' , instance_id_to_semantic_id=__a , pad_and_return_pixel_mask=__a , ) return inputs def __UpperCAmelCase ( self ): '''simple docstring''' pass def __UpperCAmelCase ( self ): '''simple docstring''' def common(__a=False , __a=None ): __a : int = self.comm_get_image_processor_inputs( with_segmentation_maps=__a , is_instance_map=__a , segmentation_type=__a ) __a : Optional[Any] = inputs['mask_labels'] __a : Optional[Any] = inputs['class_labels'] __a : Optional[Any] = inputs['pixel_values'] __a : Optional[Any] = inputs['text_inputs'] # check the batch_size for mask_label, class_label, text_input in zip(__a , __a , __a ): self.assertEqual(mask_label.shape[0] , class_label.shape[0] ) # this ensure padding has happened self.assertEqual(mask_label.shape[1:] , pixel_values.shape[2:] ) self.assertEqual(len(__a ) , self.image_processing_tester.num_text ) common() common(is_instance_map=__a ) common(is_instance_map=__a , segmentation_type='pil' ) common(is_instance_map=__a , segmentation_type='pil' ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = np.zeros((20, 50) ) __a : List[Any] = 1 __a : List[str] = 1 __a : Optional[Any] = 1 __a : List[str] = binary_mask_to_rle(__a ) self.assertEqual(len(__a ) , 4 ) self.assertEqual(rle[0] , 21 ) self.assertEqual(rle[1] , 45 ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Tuple = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='ade20k_panoptic.json' , num_text=self.image_processing_tester.num_text , repo_path='shi-labs/oneformer_demo' , ) __a : Union[str, Any] = self.image_processing_tester.get_fake_oneformer_outputs() __a : List[Any] = fature_extractor.post_process_semantic_segmentation(__a ) self.assertEqual(len(__a ) , self.image_processing_tester.batch_size ) self.assertEqual( segmentation[0].shape , ( self.image_processing_tester.height, self.image_processing_tester.width, ) , ) __a : str = [(1, 4) for i in range(self.image_processing_tester.batch_size )] __a : Optional[int] = fature_extractor.post_process_semantic_segmentation(__a , target_sizes=__a ) self.assertEqual(segmentation[0].shape , target_sizes[0] ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='ade20k_panoptic.json' , num_text=self.image_processing_tester.num_text , repo_path='shi-labs/oneformer_demo' , ) __a : List[str] = self.image_processing_tester.get_fake_oneformer_outputs() __a : Union[str, Any] = image_processor.post_process_instance_segmentation(__a , threshold=0 ) self.assertTrue(len(__a ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue('segmentation' in el ) self.assertTrue('segments_info' in el ) self.assertEqual(type(el['segments_info'] ) , __a ) self.assertEqual( el['segmentation'].shape , (self.image_processing_tester.height, self.image_processing_tester.width) ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='ade20k_panoptic.json' , num_text=self.image_processing_tester.num_text , repo_path='shi-labs/oneformer_demo' , ) __a : List[str] = self.image_processing_tester.get_fake_oneformer_outputs() __a : Optional[int] = image_processor.post_process_panoptic_segmentation(__a , threshold=0 ) self.assertTrue(len(__a ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue('segmentation' in el ) self.assertTrue('segments_info' in el ) self.assertEqual(type(el['segments_info'] ) , __a ) self.assertEqual( el['segmentation'].shape , (self.image_processing_tester.height, self.image_processing_tester.width) )
27
import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class __magic_name__ ( lowerCamelCase__ ): """simple docstring""" __UpperCamelCase = (DDPMScheduler,) def SCREAMING_SNAKE_CASE ( self :Union[str, Any] , **snake_case :str ): '''simple docstring''' A_ : Dict = { "num_train_timesteps": 1_000, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", "variance_type": "fixed_small", "clip_sample": True, } config.update(**snake_case ) return config def SCREAMING_SNAKE_CASE ( self :int ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=snake_case ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=snake_case , beta_end=snake_case ) def SCREAMING_SNAKE_CASE ( self :int ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=snake_case ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case ) def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' self.check_over_configs(thresholding=snake_case ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=snake_case , prediction_type=snake_case , sample_max_value=snake_case , ) def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=snake_case ) def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=snake_case ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' A_ : Tuple = self.scheduler_classes[0] A_ : List[str] = self.get_scheduler_config() A_ : List[str] = scheduler_class(**snake_case ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.00979 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1e-5 def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : int = self.scheduler_classes[0] A_ : List[str] = self.get_scheduler_config() A_ : int = scheduler_class(**snake_case ) A_ : Tuple = len(snake_case ) A_ : List[str] = self.dummy_model() A_ : Optional[Any] = self.dummy_sample_deter A_ : List[str] = torch.manual_seed(0 ) for t in reversed(range(snake_case ) ): # 1. predict noise residual A_ : Tuple = model(snake_case , snake_case ) # 2. predict previous mean of sample x_t-1 A_ : Dict = scheduler.step(snake_case , snake_case , snake_case , generator=snake_case ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance A_ : Optional[int] = pred_prev_sample A_ : Tuple = torch.sum(torch.abs(snake_case ) ) A_ : str = torch.mean(torch.abs(snake_case ) ) assert abs(result_sum.item() - 258.9606 ) < 1e-2 assert abs(result_mean.item() - 0.3372 ) < 1e-3 def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : Optional[int] = self.scheduler_classes[0] A_ : int = self.get_scheduler_config(prediction_type="v_prediction" ) A_ : List[str] = scheduler_class(**snake_case ) A_ : int = len(snake_case ) A_ : Dict = self.dummy_model() A_ : str = self.dummy_sample_deter A_ : Any = torch.manual_seed(0 ) for t in reversed(range(snake_case ) ): # 1. predict noise residual A_ : Optional[int] = model(snake_case , snake_case ) # 2. predict previous mean of sample x_t-1 A_ : Tuple = scheduler.step(snake_case , snake_case , snake_case , generator=snake_case ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance A_ : List[str] = pred_prev_sample A_ : Optional[Any] = torch.sum(torch.abs(snake_case ) ) A_ : List[str] = torch.mean(torch.abs(snake_case ) ) assert abs(result_sum.item() - 202.0296 ) < 1e-2 assert abs(result_mean.item() - 0.2631 ) < 1e-3 def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : str = self.scheduler_classes[0] A_ : Optional[Any] = self.get_scheduler_config() A_ : Dict = scheduler_class(**snake_case ) A_ : Optional[int] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=snake_case ) A_ : Optional[int] = scheduler.timesteps for i, timestep in enumerate(snake_case ): if i == len(snake_case ) - 1: A_ : str = -1 else: A_ : List[str] = timesteps[i + 1] A_ : Optional[int] = scheduler.previous_timestep(snake_case ) A_ : List[str] = prev_t.item() self.assertEqual(snake_case , snake_case ) def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' A_ : Optional[Any] = self.scheduler_classes[0] A_ : int = self.get_scheduler_config() A_ : Tuple = scheduler_class(**snake_case ) A_ : List[str] = [100, 87, 50, 51, 0] with self.assertRaises(snake_case , msg="`custom_timesteps` must be in descending order." ): scheduler.set_timesteps(timesteps=snake_case ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' A_ : Any = self.scheduler_classes[0] A_ : Union[str, Any] = self.get_scheduler_config() A_ : Optional[int] = scheduler_class(**snake_case ) A_ : Union[str, Any] = [100, 87, 50, 1, 0] A_ : Optional[int] = len(snake_case ) with self.assertRaises(snake_case , msg="Can only pass one of `num_inference_steps` or `custom_timesteps`." ): scheduler.set_timesteps(num_inference_steps=snake_case , timesteps=snake_case ) def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' A_ : Union[str, Any] = self.scheduler_classes[0] A_ : Optional[Any] = self.get_scheduler_config() A_ : Optional[int] = scheduler_class(**snake_case ) A_ : Optional[int] = [scheduler.config.num_train_timesteps] with self.assertRaises( snake_case , msg="`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}" , ): scheduler.set_timesteps(timesteps=snake_case )
300
0
class lowerCamelCase__ : def __init__( self : Dict , _a : Union[str, Any] ): a__: Optional[Any] =val a__: Optional[Any] =None a__: int =None def _lowerCamelCase ( self : Dict , _a : Dict ): if self.val: if val < self.val: if self.left is None: a__: Optional[int] =Node(_a ) else: self.left.insert(_a ) elif val > self.val: if self.right is None: a__: List[Any] =Node(_a ) else: self.right.insert(_a ) else: a__: Any =val def __lowerCamelCase ( __magic_name__ : Tuple , __magic_name__ : int ): # Recursive traversal if root: inorder(root.left , __magic_name__ ) res.append(root.val ) inorder(root.right , __magic_name__ ) def __lowerCamelCase ( __magic_name__ : Any ): # Build BST if len(__magic_name__ ) == 0: return arr a__: Optional[int] =Node(arr[0] ) for i in range(1 , len(__magic_name__ ) ): root.insert(arr[i] ) # Traverse BST in order. a__: Any =[] inorder(__magic_name__ , __magic_name__ ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
42
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __UpperCAmelCase = 5_00_00 __UpperCAmelCase = 50_00 __UpperCAmelCase , __UpperCAmelCase = os.path.split(__file__) __UpperCAmelCase = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def __lowerCamelCase ( __magic_name__ : datasets.Dataset , __magic_name__ : int ): for i in range(__magic_name__ ): a__: int =dataset[i] @get_duration def __lowerCamelCase ( __magic_name__ : datasets.Dataset , __magic_name__ : Any , __magic_name__ : Union[str, Any] ): for i in range(0 , len(__magic_name__ ) , __magic_name__ ): a__: List[str] =dataset[i : i + batch_size] @get_duration def __lowerCamelCase ( __magic_name__ : datasets.Dataset , __magic_name__ : Union[str, Any] , __magic_name__ : Optional[Any] ): with dataset.formatted_as(type=__magic_name__ ): for i in range(__magic_name__ ): a__: Optional[Any] =dataset[i] @get_duration def __lowerCamelCase ( __magic_name__ : datasets.Dataset , __magic_name__ : List[Any] , __magic_name__ : int , __magic_name__ : Optional[Any] ): with dataset.formatted_as(type=__magic_name__ ): for i in range(0 , __magic_name__ , __magic_name__ ): a__: List[Any] =dataset[i : i + batch_size] def __lowerCamelCase ( ): a__: Union[str, Any] ={"num examples": SPEED_TEST_N_EXAMPLES} a__: int =[ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] a__: Optional[Any] =[ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) a__: str =datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) a__: List[str] =generate_example_dataset( os.path.join(__magic_name__ , "dataset.arrow" ) , __magic_name__ , num_examples=__magic_name__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(__magic_name__ ) ) a__: str =func(__magic_name__ , **__magic_name__ ) print("shuffling dataset" ) a__: List[str] =dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(__magic_name__ ) ) a__: Optional[int] =func( __magic_name__ , **__magic_name__ ) with open(__magic_name__ , "wb" ) as f: f.write(json.dumps(__magic_name__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
42
1
import datasets a : Tuple = "\\n@InProceedings{conneau2018xnli,\n author = \"Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin\",\n title = \"XNLI: Evaluating Cross-lingual Sentence Representations\",\n booktitle = \"Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing\",\n year = \"2018\",\n publisher = \"Association for Computational Linguistics\",\n location = \"Brussels, Belgium\",\n}\n" a : Optional[Any] = "\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n" a : Dict = "\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n 'accuracy': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric(\"xnli\")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n" def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : List[Any] ): return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): """simple docstring""" def UpperCAmelCase ( self : Any ) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int64""" if self.config_name != """sts-b""" else """float32""" ), """references""": datasets.Value("""int64""" if self.config_name != """sts-b""" else """float32""" ), } ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" , ) def UpperCAmelCase ( self : Optional[Any] , __lowercase : List[Any] , __lowercase : Union[str, Any] ) -> Dict: return {"accuracy": simple_accuracy(__a , __a )}
114
"""simple docstring""" from __future__ import annotations def lowerCamelCase__ ( __snake_case, __snake_case ) -> float: """simple docstring""" _UpperCamelCase = sorted(numsa + numsa ) _UpperCamelCase , _UpperCamelCase = divmod(len(__snake_case ), 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() _a = [float(x) for x in input("""Enter the elements of first array: """).split()] _a = [float(x) for x in input("""Enter the elements of second array: """).split()] print(F"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
194
0
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 A_ :List[Any] = sys.version_info >= (3, 10) def A ( a_=None ,a_=None ) -> str: return field(default_factory=lambda: default ,metadata=a_ ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : int UpperCamelCase__ : float UpperCamelCase__ : str UpperCamelCase__ : bool @dataclass class __A : """simple docstring""" UpperCamelCase__ : int =4_2 UpperCamelCase__ : str =field(default="""toto""" , metadata={"""help""": """help message"""} ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : bool =False UpperCamelCase__ : bool =True UpperCamelCase__ : Optional[bool] =None class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[int] ="""titi""" UpperCamelCase__ : int ="""toto""" class __A ( a ): """simple docstring""" UpperCamelCase__ : Union[str, Any] ="""titi""" UpperCamelCase__ : Tuple ="""toto""" UpperCamelCase__ : Union[str, Any] =4_2 @dataclass class __A : """simple docstring""" UpperCamelCase__ : BasicEnum ="toto" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =BasicEnum(self.foo ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : MixedTypeEnum ="toto" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =MixedTypeEnum(self.foo ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : Optional[int] =None UpperCamelCase__ : Optional[float] =field(default=a , metadata={"""help""": """help message"""} ) UpperCamelCase__ : Optional[str] =None UpperCamelCase__ : Optional[List[str]] =list_field(default=[] ) UpperCamelCase__ : Optional[List[int]] =list_field(default=[] ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : List[int] =list_field(default=[] ) UpperCamelCase__ : List[int] =list_field(default=[1, 2, 3] ) UpperCamelCase__ : List[str] =list_field(default=["""Hallo""", """Bonjour""", """Hello"""] ) UpperCamelCase__ : List[float] =list_field(default=[0.1, 0.2, 0.3] ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : List[int] =field() UpperCamelCase__ : str =field() UpperCamelCase__ : BasicEnum =field() def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =BasicEnum(self.required_enum ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : int UpperCamelCase__ : "BasicEnum" =field() UpperCamelCase__ : "Optional[bool]" =None UpperCamelCase__ : "str" =field(default="""toto""" , metadata={"""help""": """help message"""} ) UpperCamelCase__ : "List[str]" =list_field(default=["""Hallo""", """Bonjour""", """Hello"""] ) if is_python_no_less_than_3_10: @dataclass class __A : """simple docstring""" UpperCamelCase__ : bool =False UpperCamelCase__ : bool =True UpperCamelCase__ : bool | None =None @dataclass class __A : """simple docstring""" UpperCamelCase__ : int | None =None UpperCamelCase__ : float | None =field(default=a , metadata={"""help""": """help message"""} ) UpperCamelCase__ : str | None =None UpperCamelCase__ : list[str] | None =list_field(default=[] ) UpperCamelCase__ : list[int] | None =list_field(default=[] ) class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" self.assertEqual(len(a._actions ) , len(b._actions ) ) for x, y in zip(a._actions , b._actions ): __UpperCamelCase : Optional[Any] ={k: v for k, v in vars(lowerCamelCase__ ).items() if k != 'container'} __UpperCamelCase : Any ={k: v for k, v in vars(lowerCamelCase__ ).items() if k != 'container'} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get('choices' , lowerCamelCase__ ) and yy.get('choices' , lowerCamelCase__ ): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx['type'](lowerCamelCase__ ) , yy['type'](lowerCamelCase__ ) ) del xx["type"], yy["type"] self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : int =argparse.ArgumentParser() expected.add_argument('--foo' , type=lowerCamelCase__ , required=lowerCamelCase__ ) expected.add_argument('--bar' , type=lowerCamelCase__ , required=lowerCamelCase__ ) expected.add_argument('--baz' , type=lowerCamelCase__ , required=lowerCamelCase__ ) expected.add_argument('--flag' , type=lowerCamelCase__ , default=lowerCamelCase__ , const=lowerCamelCase__ , nargs='?' ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Any =['--foo', '1', '--baz', 'quux', '--bar', '0.5'] ((__UpperCamelCase) , ) : Union[str, Any] =parser.parse_args_into_dataclasses(lowerCamelCase__ , look_for_args_file=lowerCamelCase__ ) self.assertFalse(example.flag ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : int =argparse.ArgumentParser() expected.add_argument('--foo' , default=42 , type=lowerCamelCase__ ) expected.add_argument('--baz' , default='toto' , type=lowerCamelCase__ , help='help message' ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =argparse.ArgumentParser() expected.add_argument('--foo' , type=lowerCamelCase__ , default=lowerCamelCase__ , const=lowerCamelCase__ , nargs='?' ) expected.add_argument('--baz' , type=lowerCamelCase__ , default=lowerCamelCase__ , const=lowerCamelCase__ , nargs='?' ) # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument('--no_baz' , action='store_false' , default=lowerCamelCase__ , dest='baz' ) expected.add_argument('--opt' , type=lowerCamelCase__ , default=lowerCamelCase__ ) __UpperCamelCase : Dict =[WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(lowerCamelCase__ ) for dataclass_type in dataclass_types: __UpperCamelCase : List[Any] =HfArgumentParser(lowerCamelCase__ ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Tuple =parser.parse_args([] ) self.assertEqual(lowerCamelCase__ , Namespace(foo=lowerCamelCase__ , baz=lowerCamelCase__ , opt=lowerCamelCase__ ) ) __UpperCamelCase : Union[str, Any] =parser.parse_args(['--foo', '--no_baz'] ) self.assertEqual(lowerCamelCase__ , Namespace(foo=lowerCamelCase__ , baz=lowerCamelCase__ , opt=lowerCamelCase__ ) ) __UpperCamelCase : Optional[int] =parser.parse_args(['--foo', '--baz'] ) self.assertEqual(lowerCamelCase__ , Namespace(foo=lowerCamelCase__ , baz=lowerCamelCase__ , opt=lowerCamelCase__ ) ) __UpperCamelCase : List[str] =parser.parse_args(['--foo', 'True', '--baz', 'True', '--opt', 'True'] ) self.assertEqual(lowerCamelCase__ , Namespace(foo=lowerCamelCase__ , baz=lowerCamelCase__ , opt=lowerCamelCase__ ) ) __UpperCamelCase : str =parser.parse_args(['--foo', 'False', '--baz', 'False', '--opt', 'False'] ) self.assertEqual(lowerCamelCase__ , Namespace(foo=lowerCamelCase__ , baz=lowerCamelCase__ , opt=lowerCamelCase__ ) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =argparse.ArgumentParser() expected.add_argument( '--foo' , default='toto' , choices=['titi', 'toto', 42] , type=make_choice_type_function(['titi', 'toto', 42] ) , ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =parser.parse_args([] ) self.assertEqual(args.foo , 'toto' ) __UpperCamelCase : Any =parser.parse_args_into_dataclasses([] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.toto ) __UpperCamelCase : Dict =parser.parse_args(['--foo', 'titi'] ) self.assertEqual(args.foo , 'titi' ) __UpperCamelCase : Union[str, Any] =parser.parse_args_into_dataclasses(['--foo', 'titi'] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.titi ) __UpperCamelCase : str =parser.parse_args(['--foo', '42'] ) self.assertEqual(args.foo , 42 ) __UpperCamelCase : Union[str, Any] =parser.parse_args_into_dataclasses(['--foo', '42'] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.fourtytwo ) def __lowercase ( self ): """simple docstring""" @dataclass class __A : """simple docstring""" UpperCamelCase__ : Literal["titi", "toto", 4_2] ="toto" __UpperCamelCase : Optional[Any] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =argparse.ArgumentParser() expected.add_argument( '--foo' , default='toto' , choices=('titi', 'toto', 42) , type=make_choice_type_function(['titi', 'toto', 42] ) , ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =parser.parse_args([] ) self.assertEqual(args.foo , 'toto' ) __UpperCamelCase : List[str] =parser.parse_args(['--foo', 'titi'] ) self.assertEqual(args.foo , 'titi' ) __UpperCamelCase : str =parser.parse_args(['--foo', '42'] ) self.assertEqual(args.foo , 42 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : str =argparse.ArgumentParser() expected.add_argument('--foo_int' , nargs='+' , default=[] , type=lowerCamelCase__ ) expected.add_argument('--bar_int' , nargs='+' , default=[1, 2, 3] , type=lowerCamelCase__ ) expected.add_argument('--foo_str' , nargs='+' , default=['Hallo', 'Bonjour', 'Hello'] , type=lowerCamelCase__ ) expected.add_argument('--foo_float' , nargs='+' , default=[0.1, 0.2, 0.3] , type=lowerCamelCase__ ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =parser.parse_args([] ) self.assertEqual( lowerCamelCase__ , Namespace(foo_int=[] , bar_int=[1, 2, 3] , foo_str=['Hallo', 'Bonjour', 'Hello'] , foo_float=[0.1, 0.2, 0.3] ) , ) __UpperCamelCase : Dict =parser.parse_args('--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7'.split() ) self.assertEqual(lowerCamelCase__ , Namespace(foo_int=[1] , bar_int=[2, 3] , foo_str=['a', 'b', 'c'] , foo_float=[0.1, 0.7] ) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =argparse.ArgumentParser() expected.add_argument('--foo' , default=lowerCamelCase__ , type=lowerCamelCase__ ) expected.add_argument('--bar' , default=lowerCamelCase__ , type=lowerCamelCase__ , help='help message' ) expected.add_argument('--baz' , default=lowerCamelCase__ , type=lowerCamelCase__ ) expected.add_argument('--ces' , nargs='+' , default=[] , type=lowerCamelCase__ ) expected.add_argument('--des' , nargs='+' , default=[] , type=lowerCamelCase__ ) __UpperCamelCase : Dict =[OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(lowerCamelCase__ ) for dataclass_type in dataclass_types: __UpperCamelCase : Optional[int] =HfArgumentParser(lowerCamelCase__ ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =parser.parse_args([] ) self.assertEqual(lowerCamelCase__ , Namespace(foo=lowerCamelCase__ , bar=lowerCamelCase__ , baz=lowerCamelCase__ , ces=[] , des=[] ) ) __UpperCamelCase : Tuple =parser.parse_args('--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3'.split() ) self.assertEqual(lowerCamelCase__ , Namespace(foo=12 , bar=3.14 , baz='42' , ces=['a', 'b', 'c'] , des=[1, 2, 3] ) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : Any =argparse.ArgumentParser() expected.add_argument('--required_list' , nargs='+' , type=lowerCamelCase__ , required=lowerCamelCase__ ) expected.add_argument('--required_str' , type=lowerCamelCase__ , required=lowerCamelCase__ ) expected.add_argument( '--required_enum' , type=make_choice_type_function(['titi', 'toto'] ) , choices=['titi', 'toto'] , required=lowerCamelCase__ , ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =argparse.ArgumentParser() expected.add_argument('--foo' , type=lowerCamelCase__ , required=lowerCamelCase__ ) expected.add_argument( '--required_enum' , type=make_choice_type_function(['titi', 'toto'] ) , choices=['titi', 'toto'] , required=lowerCamelCase__ , ) expected.add_argument('--opt' , type=lowerCamelCase__ , default=lowerCamelCase__ ) expected.add_argument('--baz' , default='toto' , type=lowerCamelCase__ , help='help message' ) expected.add_argument('--foo_str' , nargs='+' , default=['Hallo', 'Bonjour', 'Hello'] , type=lowerCamelCase__ ) self.argparsersEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : Dict ={ 'foo': 12, 'bar': 3.14, 'baz': '42', 'flag': True, } __UpperCamelCase : List[str] =parser.parse_dict(lowerCamelCase__ )[0] __UpperCamelCase : List[str] =BasicExample(**lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : str ={ 'foo': 12, 'bar': 3.14, 'baz': '42', 'flag': True, 'extra': 42, } self.assertRaises(lowerCamelCase__ , parser.parse_dict , lowerCamelCase__ , allow_extra_keys=lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] ={ 'foo': 12, 'bar': 3.14, 'baz': '42', 'flag': True, } with tempfile.TemporaryDirectory() as tmp_dir: __UpperCamelCase : Tuple =os.path.join(lowerCamelCase__ , 'temp_json' ) os.mkdir(lowerCamelCase__ ) with open(temp_local_path + '.json' , 'w+' ) as f: json.dump(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =parser.parse_yaml_file(Path(temp_local_path + '.json' ) )[0] __UpperCamelCase : Optional[Any] =BasicExample(**lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =HfArgumentParser(lowerCamelCase__ ) __UpperCamelCase : str ={ 'foo': 12, 'bar': 3.14, 'baz': '42', 'flag': True, } with tempfile.TemporaryDirectory() as tmp_dir: __UpperCamelCase : List[str] =os.path.join(lowerCamelCase__ , 'temp_yaml' ) os.mkdir(lowerCamelCase__ ) with open(temp_local_path + '.yaml' , 'w+' ) as f: yaml.dump(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =parser.parse_yaml_file(Path(temp_local_path + '.yaml' ) )[0] __UpperCamelCase : Any =BasicExample(**lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =HfArgumentParser(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ )
245
from math import pow, sqrt def A ( *a_ ) -> bool: __UpperCamelCase : Union[str, Any] =len(a_ ) > 0 and all(value > 0.0 for value in values ) return result def A ( a_ ,a_ ) -> float | ValueError: return ( round(sqrt(molar_mass_a / molar_mass_a ) ,6 ) if validate(a_ ,a_ ) else ValueError('Input Error: Molar mass values must greater than 0.' ) ) def A ( a_ ,a_ ,a_ ) -> float | ValueError: return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) ,6 ) if validate(a_ ,a_ ,a_ ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def A ( a_ ,a_ ,a_ ) -> float | ValueError: return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) ,6 ) if validate(a_ ,a_ ,a_ ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def A ( a_ ,a_ ,a_ ) -> float | ValueError: return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a ,2 ) ,6 ) if validate(a_ ,a_ ,a_ ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def A ( a_ ,a_ ,a_ ) -> float | ValueError: return ( round(pow(effusion_rate_a / effusion_rate_a ,2 ) / molar_mass ,6 ) if validate(a_ ,a_ ,a_ ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) )
245
1
from ..utils import DummyObject, requires_backends class __snake_case ( metaclass=UpperCamelCase_ ): _a = ['''note_seq'''] def __init__( self : List[str] , *A_ : Optional[int] , **A_ : Union[str, Any]): requires_backends(self , ['''note_seq''']) @classmethod def UpperCAmelCase__ ( cls : str , *A_ : List[str] , **A_ : int): requires_backends(cls , ['''note_seq''']) @classmethod def UpperCAmelCase__ ( cls : str , *A_ : Dict , **A_ : Optional[int]): requires_backends(cls , ['''note_seq'''])
103
from pathlib import Path import fire def UpperCamelCase( __UpperCamelCase : str ,__UpperCamelCase : str ,__UpperCamelCase : int ): lowerCAmelCase_ : List[str] = Path(__UpperCamelCase ) lowerCAmelCase_ : Union[str, Any] = Path(__UpperCamelCase ) dest_dir.mkdir(exist_ok=__UpperCamelCase ) for path in src_dir.iterdir(): lowerCAmelCase_ : Optional[Any] = [x.rstrip() for x in list(path.open().readlines() )][:n] lowerCAmelCase_ : List[str] = dest_dir.joinpath(path.name ) print(__UpperCamelCase ) dest_path.open('''w''' ).write('''\n'''.join(__UpperCamelCase ) ) if __name__ == "__main__": fire.Fire(minify)
103
1
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn.grep_linear''': '''encoder.layers.*.attention.gru_rel_pos_linear''', '''self_attn.relative_attention_bias''': '''encoder.layers.*.attention.rel_attn_embed''', '''self_attn.grep_a''': '''encoder.layers.*.attention.gru_rel_pos_const''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''ctc_proj''', '''mask_emb''': '''masked_spec_embed''', } lowerCamelCase__ = [ '''ctc_proj''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def lowerCAmelCase__ ( a__ , a__ , a__ , a__ , a__ ) ->Union[str, Any]: '''simple docstring''' for attribute in key.split("." ): _UpperCamelCase = getattr(a__ , a__ ) if weight_type is not None: _UpperCamelCase = getattr(a__ , a__ ).shape else: _UpperCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": _UpperCamelCase = value elif weight_type == "weight_g": _UpperCamelCase = value elif weight_type == "weight_v": _UpperCamelCase = value elif weight_type == "bias": _UpperCamelCase = value else: _UpperCamelCase = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def lowerCAmelCase__ ( a__ , a__ ) ->Any: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = fairseq_model.state_dict() _UpperCamelCase = hf_model.feature_extractor for name, value in fairseq_dict.items(): _UpperCamelCase = False if "conv_layers" in name: load_conv_layer( a__ , a__ , a__ , a__ , hf_model.config.feat_extract_norm == "group" , ) _UpperCamelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: _UpperCamelCase = True if "*" in mapped_key: _UpperCamelCase = name.split(a__ )[0].split("." )[-2] _UpperCamelCase = mapped_key.replace("*" , a__ ) if "weight_g" in name: _UpperCamelCase = "weight_g" elif "weight_v" in name: _UpperCamelCase = "weight_v" elif "bias" in name and "relative_attention_bias" not in name: _UpperCamelCase = "bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj _UpperCamelCase = "weight" else: _UpperCamelCase = None set_recursively(a__ , a__ , a__ , a__ , a__ ) continue if not is_used: unused_weights.append(a__ ) logger.warning(f'Unused weights: {unused_weights}' ) def lowerCAmelCase__ ( a__ , a__ , a__ , a__ , a__ ) ->Dict: '''simple docstring''' _UpperCamelCase = full_name.split("conv_layers." )[-1] _UpperCamelCase = name.split("." ) _UpperCamelCase = int(items[0] ) _UpperCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _UpperCamelCase = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _UpperCamelCase = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _UpperCamelCase = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _UpperCamelCase = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(a__ ) @torch.no_grad() def lowerCAmelCase__ ( a__ , a__ , a__=None ) ->int: '''simple docstring''' _UpperCamelCase = torch.load(a__ ) _UpperCamelCase = WavLMConfigOrig(checkpoint["cfg"] ) _UpperCamelCase = WavLMOrig(a__ ) model.load_state_dict(checkpoint["model"] ) model.eval() if config_path is not None: _UpperCamelCase = WavLMConfig.from_pretrained(a__ ) else: _UpperCamelCase = WavLMConfig() _UpperCamelCase = WavLMModel(a__ ) recursively_load_weights(a__ , a__ ) hf_wavlm.save_pretrained(a__ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') lowerCamelCase__ = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
359
# This code is adapted from OpenAI's release # https://github.com/openai/human-eval/blob/master/human_eval/execution.py import contextlib import faulthandler import io import multiprocessing import os import platform import signal import tempfile def lowerCAmelCase__ ( a__ , a__ , a__ , a__ ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase = multiprocessing.Manager() _UpperCamelCase = manager.list() _UpperCamelCase = multiprocessing.Process(target=a__ , args=(check_program, result, timeout) ) p.start() p.join(timeout=timeout + 1 ) if p.is_alive(): p.kill() if not result: result.append("timed out" ) return { "task_id": task_id, "passed": result[0] == "passed", "result": result[0], "completion_id": completion_id, } def lowerCAmelCase__ ( a__ , a__ , a__ ) ->int: '''simple docstring''' with create_tempdir(): # These system calls are needed when cleaning up tempdir. import os import shutil _UpperCamelCase = shutil.rmtree _UpperCamelCase = os.rmdir _UpperCamelCase = os.chdir # Disable functionalities that can make destructive changes to the test. reliability_guard() # Run program. try: _UpperCamelCase = {} with swallow_io(): with time_limit(a__ ): exec(a__ , a__ ) result.append("passed" ) except TimeoutException: result.append("timed out" ) except BaseException as e: result.append(f'failed: {e}' ) # Needed for cleaning up. _UpperCamelCase = rmtree _UpperCamelCase = rmdir _UpperCamelCase = chdir @contextlib.contextmanager def lowerCAmelCase__ ( a__ ) ->List[Any]: '''simple docstring''' def signal_handler(a__ , a__ ): raise TimeoutException("Timed out!" ) signal.setitimer(signal.ITIMER_REAL , a__ ) signal.signal(signal.SIGALRM , a__ ) try: yield finally: signal.setitimer(signal.ITIMER_REAL , 0 ) @contextlib.contextmanager def lowerCAmelCase__ ( ) ->Tuple: '''simple docstring''' _UpperCamelCase = WriteOnlyStringIO() with contextlib.redirect_stdout(a__ ): with contextlib.redirect_stderr(a__ ): with redirect_stdin(a__ ): yield @contextlib.contextmanager def lowerCAmelCase__ ( ) ->Optional[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as dirname: with chdir(a__ ): yield dirname class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' pass class _UpperCAmelCase ( io.StringIO ): '''simple docstring''' def __UpperCAmelCase ( self : Optional[int] , *lowercase_ : List[Any] , **lowercase_ : Dict) -> Optional[int]: """simple docstring""" raise OSError def __UpperCAmelCase ( self : str , *lowercase_ : Any , **lowercase_ : Optional[Any]) -> str: """simple docstring""" raise OSError def __UpperCAmelCase ( self : Union[str, Any] , *lowercase_ : Optional[Any] , **lowercase_ : Optional[Any]) -> str: """simple docstring""" raise OSError def __UpperCAmelCase ( self : Optional[Any] , *lowercase_ : str , **lowercase_ : List[Any]) -> Union[str, Any]: """simple docstring""" return False class _UpperCAmelCase ( contextlib._RedirectStream ): # type: ignore '''simple docstring''' __A = '''stdin''' @contextlib.contextmanager def lowerCAmelCase__ ( a__ ) ->Union[str, Any]: '''simple docstring''' if root == ".": yield return _UpperCamelCase = os.getcwd() os.chdir(a__ ) try: yield except BaseException as exc: raise exc finally: os.chdir(a__ ) def lowerCAmelCase__ ( a__=None ) ->Tuple: '''simple docstring''' if maximum_memory_bytes is not None: import resource resource.setrlimit(resource.RLIMIT_AS , (maximum_memory_bytes, maximum_memory_bytes) ) resource.setrlimit(resource.RLIMIT_DATA , (maximum_memory_bytes, maximum_memory_bytes) ) if not platform.uname().system == "Darwin": resource.setrlimit(resource.RLIMIT_STACK , (maximum_memory_bytes, maximum_memory_bytes) ) faulthandler.disable() import builtins _UpperCamelCase = None _UpperCamelCase = None import os _UpperCamelCase = "1" _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None import shutil _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None import subprocess _UpperCamelCase = None # type: ignore _UpperCamelCase = None import sys _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None
63
0
class lowercase : def __init__( self , A_ ) -> Dict: """simple docstring""" # we need a list not a string, so do something to change the type UpperCamelCase = arr.split(',' ) def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" UpperCamelCase = [int(self.array[0] )] * len(self.array ) UpperCamelCase = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): UpperCamelCase = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) UpperCamelCase = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": _UpperCAmelCase : Any = input("please input some numbers:") _UpperCAmelCase : Dict = SubArray(whole_array) _UpperCAmelCase : int = array.solve_sub_array() print(("the results is:", re))
222
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) _UpperCAmelCase : List[str] = { "EleutherAI/gpt-neox-20b": "https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json", # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class lowercase ( _SCREAMING_SNAKE_CASE ): __lowercase : List[str] = "gpt_neox" def __init__( self , A_=50_432 , A_=6_144 , A_=44 , A_=64 , A_=24_576 , A_="gelu" , A_=0.25 , A_=10_000 , A_=0.0 , A_=0.0 , A_=0.1 , A_=2_048 , A_=0.02 , A_=1e-5 , A_=True , A_=0 , A_=2 , A_=False , A_=True , A_=None , **A_ , ) -> Tuple: """simple docstring""" super().__init__(bos_token_id=A_ , eos_token_id=A_ , **A_ ) UpperCamelCase = vocab_size UpperCamelCase = max_position_embeddings UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = intermediate_size UpperCamelCase = hidden_act UpperCamelCase = rotary_pct UpperCamelCase = rotary_emb_base UpperCamelCase = attention_dropout UpperCamelCase = hidden_dropout UpperCamelCase = classifier_dropout UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = use_cache UpperCamelCase = tie_word_embeddings UpperCamelCase = use_parallel_residual UpperCamelCase = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( 'The hidden size is not divisble by the number of attention heads! Make sure to update them!' ) def __UpperCamelCase ( self ) -> Tuple: """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , A_ ) or len(self.rope_scaling ) != 2: raise ValueError( '`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ' F'''got {self.rope_scaling}''' ) UpperCamelCase = self.rope_scaling.get('type' , A_ ) UpperCamelCase = self.rope_scaling.get('factor' , A_ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(A_ , A_ ) or rope_scaling_factor <= 1.0: raise ValueError(F'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
222
1
from __future__ import annotations SCREAMING_SNAKE_CASE_ = 1.6_021e-19 # units = C def UpperCamelCase__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> tuple[str, float]: '''simple docstring''' if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError("""You cannot supply more or less than 2 values""" ) elif conductivity < 0: raise ValueError("""Conductivity cannot be negative""" ) elif electron_conc < 0: raise ValueError("""Electron concentration cannot be negative""" ) elif mobility < 0: raise ValueError("""mobility cannot be negative""" ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
363
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class UpperCamelCase__ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' __snake_case : str = AltDiffusionPipeline __snake_case : int = TEXT_TO_IMAGE_PARAMS __snake_case : Dict = TEXT_TO_IMAGE_BATCH_PARAMS __snake_case : Tuple = TEXT_TO_IMAGE_IMAGE_PARAMS __snake_case : Optional[int] = TEXT_TO_IMAGE_IMAGE_PARAMS def SCREAMING_SNAKE_CASE__ ( self : int ) -> Optional[Any]: '''simple docstring''' torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") ,up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") ,cross_attention_dim=32 ,) SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.00085 ,beta_end=0.012 ,beta_schedule="""scaled_linear""" ,clip_sample=lowerCamelCase__ ,set_alpha_to_one=lowerCamelCase__ ,) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] ,up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] ,latent_channels=4 ,) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,projection_dim=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=5002 ,) SCREAMING_SNAKE_CASE = CLIPTextModel(lowerCamelCase__ ) SCREAMING_SNAKE_CASE = XLMRobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-xlm-roberta""" ) SCREAMING_SNAKE_CASE = 77 SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def SCREAMING_SNAKE_CASE__ ( self : int ,lowerCamelCase__ : Dict ,lowerCamelCase__ : int=0 ) -> Any: '''simple docstring''' if str(lowerCamelCase__ ).startswith("""mps""" ): SCREAMING_SNAKE_CASE = torch.manual_seed(lowerCamelCase__ ) else: SCREAMING_SNAKE_CASE = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> List[str]: '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE = self.get_dummy_components() torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = RobertaSeriesConfig( hidden_size=32 ,project_dim=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,vocab_size=5002 ,) # TODO: remove after fixing the non-deterministic text encoder SCREAMING_SNAKE_CASE = RobertaSeriesModelWithTransformation(lowerCamelCase__ ) SCREAMING_SNAKE_CASE = text_encoder SCREAMING_SNAKE_CASE = AltDiffusionPipeline(**lowerCamelCase__ ) SCREAMING_SNAKE_CASE = alt_pipe.to(lowerCamelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) SCREAMING_SNAKE_CASE = self.get_dummy_inputs(lowerCamelCase__ ) SCREAMING_SNAKE_CASE = """A photo of an astronaut""" SCREAMING_SNAKE_CASE = alt_pipe(**lowerCamelCase__ ) SCREAMING_SNAKE_CASE = output.images SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) SCREAMING_SNAKE_CASE = np.array( [0.5748162, 0.60447145, 0.48821217, 0.50100636, 0.5431185, 0.45763683, 0.49657696, 0.48132733, 0.47573093] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE__ ( self : Any ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = RobertaSeriesConfig( hidden_size=32 ,project_dim=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,vocab_size=5002 ,) # TODO: remove after fixing the non-deterministic text encoder SCREAMING_SNAKE_CASE = RobertaSeriesModelWithTransformation(lowerCamelCase__ ) SCREAMING_SNAKE_CASE = text_encoder SCREAMING_SNAKE_CASE = AltDiffusionPipeline(**lowerCamelCase__ ) SCREAMING_SNAKE_CASE = alt_pipe.to(lowerCamelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) SCREAMING_SNAKE_CASE = self.get_dummy_inputs(lowerCamelCase__ ) SCREAMING_SNAKE_CASE = alt_pipe(**lowerCamelCase__ ) SCREAMING_SNAKE_CASE = output.images SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) SCREAMING_SNAKE_CASE = np.array( [0.51605093, 0.5707241, 0.47365507, 0.50578886, 0.5633877, 0.4642503, 0.5182081, 0.48763484, 0.49084237] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : List[str] ) -> List[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE = AltDiffusionPipeline.from_pretrained("""BAAI/AltDiffusion""" ,safety_checker=lowerCamelCase__ ) SCREAMING_SNAKE_CASE = alt_pipe.to(lowerCamelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) SCREAMING_SNAKE_CASE = """A painting of a squirrel eating a burger""" SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = alt_pipe([prompt] ,generator=lowerCamelCase__ ,guidance_scale=6.0 ,num_inference_steps=20 ,output_type="""np""" ) SCREAMING_SNAKE_CASE = output.images SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE = np.array([0.1010, 0.0800, 0.0794, 0.0885, 0.0843, 0.0762, 0.0769, 0.0729, 0.0586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = DDIMScheduler.from_pretrained("""BAAI/AltDiffusion""" ,subfolder="""scheduler""" ) SCREAMING_SNAKE_CASE = AltDiffusionPipeline.from_pretrained("""BAAI/AltDiffusion""" ,scheduler=lowerCamelCase__ ,safety_checker=lowerCamelCase__ ) SCREAMING_SNAKE_CASE = alt_pipe.to(lowerCamelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) SCREAMING_SNAKE_CASE = """A painting of a squirrel eating a burger""" SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) SCREAMING_SNAKE_CASE = alt_pipe([prompt] ,generator=lowerCamelCase__ ,num_inference_steps=2 ,output_type="""numpy""" ) SCREAMING_SNAKE_CASE = output.images SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE = np.array([0.4019, 0.4052, 0.3810, 0.4119, 0.3916, 0.3982, 0.4651, 0.4195, 0.5323] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
193
0
from __future__ import annotations from typing import Any def UpperCAmelCase_ ( __lowerCAmelCase ) -> None: create_state_space_tree(__lowerCAmelCase , [] , 0 ) def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> None: if index == len(__lowerCAmelCase ): print(__lowerCAmelCase ) return create_state_space_tree(__lowerCAmelCase , __lowerCAmelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCAmelCase , __lowerCAmelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __lowerCAmelCase : Tuple = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
156
"""simple docstring""" import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) _A = { """sample_size""": 32, """in_channels""": 3, """out_channels""": 3, """layers_per_block""": 2, """num_class_embeds""": 10_00, """block_out_channels""": [32, 64], """attention_head_dim""": 8, """down_block_types""": [ """ResnetDownsampleBlock2D""", """AttnDownBlock2D""", ], """up_block_types""": [ """AttnUpBlock2D""", """ResnetUpsampleBlock2D""", ], """resnet_time_scale_shift""": """scale_shift""", """upsample_type""": """resnet""", """downsample_type""": """resnet""", } _A = { """sample_size""": 64, """in_channels""": 3, """out_channels""": 3, """layers_per_block""": 3, """num_class_embeds""": 10_00, """block_out_channels""": [1_92, 1_92 * 2, 1_92 * 3, 1_92 * 4], """attention_head_dim""": 64, """down_block_types""": [ """ResnetDownsampleBlock2D""", """AttnDownBlock2D""", """AttnDownBlock2D""", """AttnDownBlock2D""", ], """up_block_types""": [ """AttnUpBlock2D""", """AttnUpBlock2D""", """AttnUpBlock2D""", """ResnetUpsampleBlock2D""", ], """resnet_time_scale_shift""": """scale_shift""", """upsample_type""": """resnet""", """downsample_type""": """resnet""", } _A = { """sample_size""": 2_56, """in_channels""": 3, """out_channels""": 3, """layers_per_block""": 2, """num_class_embeds""": None, """block_out_channels""": [2_56, 2_56, 2_56 * 2, 2_56 * 2, 2_56 * 4, 2_56 * 4], """attention_head_dim""": 64, """down_block_types""": [ """ResnetDownsampleBlock2D""", """ResnetDownsampleBlock2D""", """ResnetDownsampleBlock2D""", """AttnDownBlock2D""", """AttnDownBlock2D""", """AttnDownBlock2D""", ], """up_block_types""": [ """AttnUpBlock2D""", """AttnUpBlock2D""", """AttnUpBlock2D""", """ResnetUpsampleBlock2D""", """ResnetUpsampleBlock2D""", """ResnetUpsampleBlock2D""", ], """resnet_time_scale_shift""": """default""", """upsample_type""": """resnet""", """downsample_type""": """resnet""", } _A = { """num_train_timesteps""": 40, """sigma_min""": 0.002, """sigma_max""": 80.0, } _A = { """num_train_timesteps""": 2_01, """sigma_min""": 0.002, """sigma_max""": 80.0, } _A = { """num_train_timesteps""": 1_51, """sigma_min""": 0.002, """sigma_max""": 80.0, } def a__ ( lowerCAmelCase ) -> Tuple: if isinstance(lowerCAmelCase , lowerCAmelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("""boolean value expected""" ) def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False ) -> List[str]: UpperCAmelCase__ : int = checkpoint[F"""{old_prefix}.in_layers.0.weight"""] UpperCAmelCase__ : Optional[int] = checkpoint[F"""{old_prefix}.in_layers.0.bias"""] UpperCAmelCase__ : Optional[Any] = checkpoint[F"""{old_prefix}.in_layers.2.weight"""] UpperCAmelCase__ : str = checkpoint[F"""{old_prefix}.in_layers.2.bias"""] UpperCAmelCase__ : Any = checkpoint[F"""{old_prefix}.emb_layers.1.weight"""] UpperCAmelCase__ : Optional[int] = checkpoint[F"""{old_prefix}.emb_layers.1.bias"""] UpperCAmelCase__ : str = checkpoint[F"""{old_prefix}.out_layers.0.weight"""] UpperCAmelCase__ : List[Any] = checkpoint[F"""{old_prefix}.out_layers.0.bias"""] UpperCAmelCase__ : Dict = checkpoint[F"""{old_prefix}.out_layers.3.weight"""] UpperCAmelCase__ : Union[str, Any] = checkpoint[F"""{old_prefix}.out_layers.3.bias"""] if has_skip: UpperCAmelCase__ : int = checkpoint[F"""{old_prefix}.skip_connection.weight"""] UpperCAmelCase__ : Dict = checkpoint[F"""{old_prefix}.skip_connection.bias"""] return new_checkpoint def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=None ) -> Optional[int]: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = checkpoint[F"""{old_prefix}.qkv.weight"""].chunk(3 , dim=0 ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Any = checkpoint[F"""{old_prefix}.qkv.bias"""].chunk(3 , dim=0 ) UpperCAmelCase__ : List[Any] = checkpoint[F"""{old_prefix}.norm.weight"""] UpperCAmelCase__ : str = checkpoint[F"""{old_prefix}.norm.bias"""] UpperCAmelCase__ : Union[str, Any] = weight_q.squeeze(-1 ).squeeze(-1 ) UpperCAmelCase__ : Optional[Any] = bias_q.squeeze(-1 ).squeeze(-1 ) UpperCAmelCase__ : Any = weight_k.squeeze(-1 ).squeeze(-1 ) UpperCAmelCase__ : int = bias_k.squeeze(-1 ).squeeze(-1 ) UpperCAmelCase__ : Dict = weight_v.squeeze(-1 ).squeeze(-1 ) UpperCAmelCase__ : int = bias_v.squeeze(-1 ).squeeze(-1 ) UpperCAmelCase__ : Any = ( checkpoint[F"""{old_prefix}.proj_out.weight"""].squeeze(-1 ).squeeze(-1 ) ) UpperCAmelCase__ : str = checkpoint[F"""{old_prefix}.proj_out.bias"""].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def a__ ( lowerCAmelCase , lowerCAmelCase ) -> str: UpperCAmelCase__ : Optional[Any] = torch.load(lowerCAmelCase , map_location="""cpu""" ) UpperCAmelCase__ : List[Any] = {} UpperCAmelCase__ : List[Any] = checkpoint["""time_embed.0.weight"""] UpperCAmelCase__ : str = checkpoint["""time_embed.0.bias"""] UpperCAmelCase__ : List[str] = checkpoint["""time_embed.2.weight"""] UpperCAmelCase__ : Dict = checkpoint["""time_embed.2.bias"""] if unet_config["num_class_embeds"] is not None: UpperCAmelCase__ : Dict = checkpoint["""label_emb.weight"""] UpperCAmelCase__ : str = checkpoint["""input_blocks.0.0.weight"""] UpperCAmelCase__ : List[str] = checkpoint["""input_blocks.0.0.bias"""] UpperCAmelCase__ : List[str] = unet_config["""down_block_types"""] UpperCAmelCase__ : Tuple = unet_config["""layers_per_block"""] UpperCAmelCase__ : int = unet_config["""attention_head_dim"""] UpperCAmelCase__ : Union[str, Any] = unet_config["""block_out_channels"""] UpperCAmelCase__ : Union[str, Any] = 1 UpperCAmelCase__ : Union[str, Any] = channels_list[0] for i, layer_type in enumerate(lowerCAmelCase ): UpperCAmelCase__ : Union[str, Any] = channels_list[i] UpperCAmelCase__ : int = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(lowerCAmelCase ): UpperCAmelCase__ : Tuple = F"""down_blocks.{i}.resnets.{j}""" UpperCAmelCase__ : List[Any] = F"""input_blocks.{current_layer}.0""" UpperCAmelCase__ : Dict = True if j == 0 and downsample_block_has_skip else False UpperCAmelCase__ : Optional[Any] = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , has_skip=lowerCAmelCase ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(lowerCAmelCase ): UpperCAmelCase__ : Any = F"""down_blocks.{i}.resnets.{j}""" UpperCAmelCase__ : Optional[Any] = F"""input_blocks.{current_layer}.0""" UpperCAmelCase__ : int = True if j == 0 and downsample_block_has_skip else False UpperCAmelCase__ : Union[str, Any] = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , has_skip=lowerCAmelCase ) UpperCAmelCase__ : Dict = F"""down_blocks.{i}.attentions.{j}""" UpperCAmelCase__ : int = F"""input_blocks.{current_layer}.1""" UpperCAmelCase__ : Union[str, Any] = convert_attention( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) current_layer += 1 if i != len(lowerCAmelCase ) - 1: UpperCAmelCase__ : Any = F"""down_blocks.{i}.downsamplers.0""" UpperCAmelCase__ : List[str] = F"""input_blocks.{current_layer}.0""" UpperCAmelCase__ : Tuple = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) current_layer += 1 UpperCAmelCase__ : Tuple = current_channels # hardcoded the mid-block for now UpperCAmelCase__ : List[Any] = """mid_block.resnets.0""" UpperCAmelCase__ : str = """middle_block.0""" UpperCAmelCase__ : List[str] = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase__ : List[str] = """mid_block.attentions.0""" UpperCAmelCase__ : Any = """middle_block.1""" UpperCAmelCase__ : Optional[int] = convert_attention(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase__ : List[Any] = """mid_block.resnets.1""" UpperCAmelCase__ : Tuple = """middle_block.2""" UpperCAmelCase__ : Union[str, Any] = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase__ : Any = 0 UpperCAmelCase__ : Dict = unet_config["""up_block_types"""] for i, layer_type in enumerate(lowerCAmelCase ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): UpperCAmelCase__ : Tuple = F"""up_blocks.{i}.resnets.{j}""" UpperCAmelCase__ : Optional[Any] = F"""output_blocks.{current_layer}.0""" UpperCAmelCase__ : Dict = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , has_skip=lowerCAmelCase ) current_layer += 1 if i != len(lowerCAmelCase ) - 1: UpperCAmelCase__ : List[str] = F"""up_blocks.{i}.upsamplers.0""" UpperCAmelCase__ : Any = F"""output_blocks.{current_layer-1}.1""" UpperCAmelCase__ : Union[str, Any] = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): UpperCAmelCase__ : List[str] = F"""up_blocks.{i}.resnets.{j}""" UpperCAmelCase__ : Dict = F"""output_blocks.{current_layer}.0""" UpperCAmelCase__ : Any = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , has_skip=lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = F"""up_blocks.{i}.attentions.{j}""" UpperCAmelCase__ : List[str] = F"""output_blocks.{current_layer}.1""" UpperCAmelCase__ : Dict = convert_attention( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) current_layer += 1 if i != len(lowerCAmelCase ) - 1: UpperCAmelCase__ : int = F"""up_blocks.{i}.upsamplers.0""" UpperCAmelCase__ : int = F"""output_blocks.{current_layer-1}.2""" UpperCAmelCase__ : Union[str, Any] = convert_resnet(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = checkpoint["""out.0.weight"""] UpperCAmelCase__ : List[Any] = checkpoint["""out.0.bias"""] UpperCAmelCase__ : Tuple = checkpoint["""out.2.weight"""] UpperCAmelCase__ : Optional[Any] = checkpoint["""out.2.bias"""] return new_checkpoint if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument("""--unet_path""", default=None, type=str, required=True, help="""Path to the unet.pt to convert.""") parser.add_argument( """--dump_path""", default=None, type=str, required=True, help="""Path to output the converted UNet model.""" ) parser.add_argument("""--class_cond""", default=True, type=str, help="""Whether the model is class-conditional.""") _A = parser.parse_args() _A = strabool(args.class_cond) _A = os.path.basename(args.unet_path) print(f'''Checkpoint: {ckpt_name}''') # Get U-Net config if "imagenet64" in ckpt_name: _A = IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): _A = LSUN_256_UNET_CONFIG elif "test" in ckpt_name: _A = TEST_UNET_CONFIG else: raise ValueError(f'''Checkpoint type {ckpt_name} is not currently supported.''') if not args.class_cond: _A = None _A = con_pt_to_diffuser(args.unet_path, unet_config) _A = UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: _A = CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: _A = CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): _A = CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(f'''Checkpoint type {ckpt_name} is not currently supported.''') _A = CMStochasticIterativeScheduler(**scheduler_config) _A = ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
171
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase =logging.get_logger(__name__) __UpperCAmelCase ={ "google/switch-base-8": "https://huggingface.co/google/switch-base-8/blob/main/config.json", } class a__ ( UpperCAmelCase__ ): lowerCamelCase : Optional[int] ="switch_transformers" lowerCamelCase : Optional[int] =["past_key_values"] lowerCamelCase : Any ={"hidden_size": "d_model", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers"} def __init__( self : List[Any] , a : int=3_21_28 , a : List[Any]=7_68 , a : Any=64 , a : Optional[Any]=20_48 , a : Optional[int]=64 , a : Optional[int]=12 , a : Optional[Any]=3 , a : int=12 , a : Optional[Any]=3 , a : Tuple=12 , a : Union[str, Any]=8 , a : Dict=False , a : Optional[int]=0.01 , a : str="float32" , a : List[Any]=False , a : Union[str, Any]=32 , a : Optional[int]=1_28 , a : Dict=0.1 , a : List[str]=1e-6 , a : Optional[int]=0.0_01 , a : Tuple=0.0_01 , a : int=1.0 , a : Optional[int]="relu" , a : List[Any]=True , a : Dict=False , a : Any=True , a : Dict=0 , a : int=1 , **a : Optional[int] , ): """simple docstring""" __lowerCamelCase = vocab_size __lowerCamelCase = d_model __lowerCamelCase = d_kv __lowerCamelCase = d_ff __lowerCamelCase = num_sparse_encoder_layers __lowerCamelCase = num_layers __lowerCamelCase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry __lowerCamelCase = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: __lowerCamelCase = self.num_layers // self.num_sparse_encoder_layers else: __lowerCamelCase = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: __lowerCamelCase = self.num_decoder_layers // self.num_sparse_decoder_layers else: __lowerCamelCase = self.num_decoder_layers # HACK: this will create 0 sparse layers __lowerCamelCase = num_heads __lowerCamelCase = num_experts __lowerCamelCase = expert_capacity __lowerCamelCase = router_bias __lowerCamelCase = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(f"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) __lowerCamelCase = router_dtype __lowerCamelCase = router_ignore_padding_tokens __lowerCamelCase = relative_attention_num_buckets __lowerCamelCase = relative_attention_max_distance __lowerCamelCase = dropout_rate __lowerCamelCase = layer_norm_epsilon __lowerCamelCase = initializer_factor __lowerCamelCase = feed_forward_proj __lowerCamelCase = use_cache __lowerCamelCase = add_router_probs __lowerCamelCase = router_z_loss_coef __lowerCamelCase = router_aux_loss_coef __lowerCamelCase = self.feed_forward_proj.split('''-''' ) __lowerCamelCase = act_info[-1] __lowerCamelCase = act_info[0] == '''gated''' if len(a ) > 1 and act_info[0] != "gated" or len(a ) > 2: raise ValueError( f"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": __lowerCamelCase = '''gelu_new''' super().__init__( pad_token_id=a , eos_token_id=a , is_encoder_decoder=a , **a , )
237
'''simple docstring''' import collections import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCAmelCase =logging.get_logger(__name__) __UpperCAmelCase ="▁" __UpperCAmelCase ={"vocab_file": "prophetnet.tokenizer"} __UpperCAmelCase ={ "vocab_file": { "microsoft/xprophetnet-large-wiki100-cased": ( "https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/prophetnet.tokenizer" ), } } __UpperCAmelCase ={ "microsoft/xprophetnet-large-wiki100-cased": {"do_lower_case": False}, } __UpperCAmelCase ={ "microsoft/xprophetnet-large-wiki100-cased": 5_1_2, } def __lowerCAmelCase ( UpperCamelCase__ ) -> List[str]: __lowerCamelCase = collections.OrderedDict() with open(UpperCamelCase__ , '''r''' , encoding='''utf-8''' ) as reader: __lowerCamelCase = reader.readlines() for index, token in enumerate(UpperCamelCase__ ): __lowerCamelCase = token.rstrip('''\n''' ) __lowerCamelCase = index return vocab class a__ ( UpperCAmelCase__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : Any =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Optional[int] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Union[str, Any] =["input_ids", "attention_mask"] def __init__( self : int , a : List[str] , a : Optional[int]="[SEP]" , a : int="[SEP]" , a : str="[SEP]" , a : List[Any]="[UNK]" , a : List[Any]="[PAD]" , a : str="[CLS]" , a : List[str]="[MASK]" , a : Optional[Dict[str, Any]] = None , **a : str , ): """simple docstring""" __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=a , eos_token=a , sep_token=a , unk_token=a , pad_token=a , cls_token=a , mask_token=a , sp_model_kwargs=self.sp_model_kwargs , **a , ) try: import sentencepiece as spm except ImportError: logger.warning( '''You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece''' ''' pip install sentencepiece''' ) raise __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(a ) ) __lowerCamelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # put special tokens and [unused] tokens into the vocab __lowerCamelCase = {'''[PAD]''': 0, '''[CLS]''': 1, '''[SEP]''': 2, '''[UNK]''': 3, '''[MASK]''': 4} for i in range(10 ): __lowerCamelCase = f"""[unused{i}]""" __lowerCamelCase = 5 + i # The first "real" token "," has position 15 in the embedding vocab and position 3 in the spm vocab __lowerCamelCase = 12 __lowerCamelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} for k in self.fairseq_tokens_to_ids.keys(): self.unique_no_split_tokens.append(a ) def __getstate__( self : List[str] ): """simple docstring""" __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self : int , a : List[Any] ): """simple docstring""" __lowerCamelCase = d try: import sentencepiece as spm except ImportError: logger.warning( '''You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece''' ''' pip install sentencepiece''' ) raise # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE__ ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return ([0] * len(a )) + [1] return ([0] * len(a )) + [1] + ([0] * len(a )) + [1] def SCREAMING_SNAKE_CASE__ ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): """simple docstring""" __lowerCamelCase = [self.sep_token_id] if token_ids_a is None: return len(token_ids_a + sep ) * [0] return len(token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" return len(self.sp_model ) + self.fairseq_offset def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" __lowerCamelCase = {self.convert_ids_to_tokens(a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def SCREAMING_SNAKE_CASE__ ( self : Tuple , a : str ): """simple docstring""" return self.sp_model.encode(a , out_type=a ) def SCREAMING_SNAKE_CASE__ ( self : Dict , a : int ): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __lowerCamelCase = self.sp_model.PieceToId(a ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : Union[str, Any] ): """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , a : Tuple ): """simple docstring""" __lowerCamelCase = ''''''.join(a ).replace(a , ''' ''' ).strip() return out_string def SCREAMING_SNAKE_CASE__ ( self : int , a : str , a : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , a ) elif not os.path.isfile(self.vocab_file ): with open(a , '''wb''' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(a ) return (out_vocab_file,) def SCREAMING_SNAKE_CASE__ ( self : Any , a : List[int] , a : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return token_ids_a + [self.sep_token_id] __lowerCamelCase = [self.sep_token_id] return token_ids_a + sep + token_ids_a + sep
237
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if not all(x.isalpha() for x in string ): raise ValueError('String must only contain alphabetic characters.' ) lowerCamelCase__ : Tuple = sorted(string.lower() ) return len(_UpperCAmelCase ) == len(set(_UpperCAmelCase ) ) if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = input("""Enter a string """).strip() _UpperCAmelCase : Optional[int] = is_isogram(input_str) print(F"""{input_str} is {"an" if isogram else "not an"} isogram.""")
50
'''simple docstring''' def lowercase ( __magic_name__ ): '''simple docstring''' if number < 0: raise ValueError("number must not be negative" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
311
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __a = { "configuration_bert": ["BERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BertConfig", "BertOnnxConfig"], "tokenization_bert": ["BasicTokenizer", "BertTokenizer", "WordpieceTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["BertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "BertForMaskedLM", "BertForMultipleChoice", "BertForNextSentencePrediction", "BertForPreTraining", "BertForQuestionAnswering", "BertForSequenceClassification", "BertForTokenClassification", "BertLayer", "BertLMHeadModel", "BertModel", "BertPreTrainedModel", "load_tf_weights_in_bert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFBertEmbeddings", "TFBertForMaskedLM", "TFBertForMultipleChoice", "TFBertForNextSentencePrediction", "TFBertForPreTraining", "TFBertForQuestionAnswering", "TFBertForSequenceClassification", "TFBertForTokenClassification", "TFBertLMHeadModel", "TFBertMainLayer", "TFBertModel", "TFBertPreTrainedModel", ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["TFBertTokenizer"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "FlaxBertForCausalLM", "FlaxBertForMaskedLM", "FlaxBertForMultipleChoice", "FlaxBertForNextSentencePrediction", "FlaxBertForPreTraining", "FlaxBertForQuestionAnswering", "FlaxBertForSequenceClassification", "FlaxBertForTokenClassification", "FlaxBertModel", "FlaxBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
371
'''simple docstring''' from math import factorial def __UpperCAmelCase ( a_: int = 100 ): return sum(map(a_, str(factorial(a_ ) ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
17
0
'''simple docstring''' # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def UpperCAmelCase_ ( __lowercase : int , __lowercase : Optional[int] , __lowercase : Union[str, Any] ) -> str: '''simple docstring''' _UpperCAmelCase = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] _UpperCAmelCase = { "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } _UpperCAmelCase = f'{src_lang}-{tgt_lang}' _UpperCAmelCase = f'\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "facebook/wmt19-{src_lang}-{tgt_lang}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n' os.makedirs(__lowercase , exist_ok=__lowercase ) _UpperCAmelCase = os.path.join(__lowercase , "README.md" ) print(f'Generating {path}' ) with open(__lowercase , "w" , encoding="utf-8" ) as f: f.write(__lowercase ) # make sure we are under the root of the project __SCREAMING_SNAKE_CASE :Optional[Any] = Path(__file__).resolve().parent.parent.parent __SCREAMING_SNAKE_CASE :Any = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :Optional[Any] = model_name.split('''-''') __SCREAMING_SNAKE_CASE :List[str] = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
22
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) UpperCAmelCase = '''\ Text data. Second line of data.''' UpperCAmelCase = '''file''' @pytest.fixture(scope='session' ) def __UpperCamelCase ( lowercase__ : List[Any] ): '''simple docstring''' __lowercase =tmp_path_factory.mktemp('data' ) / (FILE_PATH + '.zstd') __lowercase =bytes(lowercase__, 'utf-8' ) with zstd.open(lowercase__, 'wb' ) as f: f.write(lowercase__ ) return path @pytest.fixture def __UpperCamelCase ( lowercase__ : str ): '''simple docstring''' with open(os.path.join(tmpfs.local_root_dir, lowercase__ ), 'w' ) as f: f.write(lowercase__ ) return FILE_PATH @pytest.mark.parametrize('compression_format', ['gzip', 'xz', 'zstd'] ) def __UpperCamelCase ( lowercase__ : Any, lowercase__ : List[str], lowercase__ : Optional[int], lowercase__ : str, lowercase__ : int, lowercase__ : Dict ): '''simple docstring''' __lowercase ={'gzip': gz_file, 'xz': xz_file, 'zstd': zstd_path} __lowercase =input_paths[compression_format] __lowercase =tmp_path / 'cache' __lowercase =DownloadConfig(cache_dir=lowercase__, extract_compressed_file=lowercase__ ) __lowercase =cached_path(lowercase__, download_config=lowercase__ ) with open(lowercase__ ) as f: __lowercase =f.read() with open(lowercase__ ) as f: __lowercase =f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('default_extracted', [True, False] ) @pytest.mark.parametrize('default_cache_dir', [True, False] ) def __UpperCamelCase ( lowercase__ : Union[str, Any], lowercase__ : Tuple, lowercase__ : int, lowercase__ : int, lowercase__ : Optional[int] ): '''simple docstring''' __lowercase ='custom_cache' __lowercase ='custom_extracted_dir' __lowercase =tmp_path / 'custom_extracted_path' if default_extracted: __lowercase =('downloads' if default_cache_dir else custom_cache_dir, 'extracted') else: monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_DIR', lowercase__ ) monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH', str(lowercase__ ) ) __lowercase =custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __lowercase =xz_file __lowercase =( DownloadConfig(extract_compressed_file=lowercase__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir, extract_compressed_file=lowercase__ ) ) __lowercase =cached_path(lowercase__, download_config=lowercase__ ) assert Path(lowercase__ ).parent.parts[-2:] == expected def __UpperCamelCase ( lowercase__ : List[Any] ): '''simple docstring''' __lowercase =str(Path(lowercase__ ).resolve() ) assert cached_path(lowercase__ ) == text_file # relative path __lowercase =str(Path(lowercase__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(lowercase__ ) == text_file def __UpperCamelCase ( lowercase__ : Optional[Any] ): '''simple docstring''' __lowercase =str(tmp_path.resolve() / '__missing_file__.txt' ) with pytest.raises(lowercase__ ): cached_path(lowercase__ ) # relative path __lowercase ='./__missing_file__.txt' with pytest.raises(lowercase__ ): cached_path(lowercase__ ) def __UpperCamelCase ( lowercase__ : List[str] ): '''simple docstring''' __lowercase =get_from_cache(F'''tmp://{tmpfs_file}''' ) with open(lowercase__ ) as f: __lowercase =f.read() assert output_file_content == FILE_CONTENT @patch('datasets.config.HF_DATASETS_OFFLINE', lowercase__ ) def __UpperCamelCase ( ): '''simple docstring''' with pytest.raises(lowercase__ ): cached_path('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE', lowercase__ ) def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' __lowercase =tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): http_get('https://huggingface.co', temp_file=lowercase__ ) with pytest.raises(lowercase__ ): http_head('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE', lowercase__ ) def __UpperCamelCase ( lowercase__ : Optional[int] ): '''simple docstring''' __lowercase =tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): ftp_get('ftp://huggingface.co', temp_file=lowercase__ ) with pytest.raises(lowercase__ ): ftp_head('ftp://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE', lowercase__ ) def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' __lowercase =tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): fsspec_get('s3://huggingface.co', temp_file=lowercase__ ) with pytest.raises(lowercase__ ): fsspec_head('s3://huggingface.co' )
141
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING __magic_name__: int = logging.get_logger(__name__) __magic_name__: List[Any] = { '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class snake_case__ ( lowerCamelCase__ ): lowercase__ : Optional[int] = '''table-transformer''' lowercase__ : Dict = ['''past_key_values'''] lowercase__ : Optional[Any] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=3 , lowerCAmelCase__=1_00 , lowerCAmelCase__=6 , lowerCAmelCase__=20_48 , lowerCAmelCase__=8 , lowerCAmelCase__=6 , lowerCAmelCase__=20_48 , lowerCAmelCase__=8 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=True , lowerCAmelCase__="relu" , lowerCAmelCase__=2_56 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0_2 , lowerCAmelCase__=1.0 , lowerCAmelCase__=False , lowerCAmelCase__="sine" , lowerCAmelCase__="resnet50" , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=1 , lowerCAmelCase__=5 , lowerCAmelCase__=2 , lowerCAmelCase__=1 , lowerCAmelCase__=1 , lowerCAmelCase__=5 , lowerCAmelCase__=2 , lowerCAmelCase__=0.1 , **lowerCAmelCase__ , ) -> Union[str, Any]: if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) __magic_name__ : List[Any] = CONFIG_MAPPING["resnet"](out_features=["""stage4"""] ) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): __magic_name__ : int = backbone_config.get("""model_type""" ) __magic_name__ : Optional[Any] = CONFIG_MAPPING[backbone_model_type] __magic_name__ : Union[str, Any] = config_class.from_dict(lowerCAmelCase__ ) # set timm attributes to None __magic_name__ : Dict = None, None, None __magic_name__ : Tuple = use_timm_backbone __magic_name__ : Tuple = backbone_config __magic_name__ : Union[str, Any] = num_channels __magic_name__ : str = num_queries __magic_name__ : Dict = d_model __magic_name__ : List[Any] = encoder_ffn_dim __magic_name__ : str = encoder_layers __magic_name__ : str = encoder_attention_heads __magic_name__ : List[str] = decoder_ffn_dim __magic_name__ : Any = decoder_layers __magic_name__ : str = decoder_attention_heads __magic_name__ : Union[str, Any] = dropout __magic_name__ : List[Any] = attention_dropout __magic_name__ : str = activation_dropout __magic_name__ : List[str] = activation_function __magic_name__ : List[str] = init_std __magic_name__ : Tuple = init_xavier_std __magic_name__ : Tuple = encoder_layerdrop __magic_name__ : Dict = decoder_layerdrop __magic_name__ : List[str] = encoder_layers __magic_name__ : Any = auxiliary_loss __magic_name__ : List[Any] = position_embedding_type __magic_name__ : Union[str, Any] = backbone __magic_name__ : List[str] = use_pretrained_backbone __magic_name__ : List[Any] = dilation # Hungarian matcher __magic_name__ : Any = class_cost __magic_name__ : Union[str, Any] = bbox_cost __magic_name__ : str = giou_cost # Loss coefficients __magic_name__ : List[Any] = mask_loss_coefficient __magic_name__ : str = dice_loss_coefficient __magic_name__ : List[Any] = bbox_loss_coefficient __magic_name__ : Optional[int] = giou_loss_coefficient __magic_name__ : str = eos_coefficient super().__init__(is_encoder_decoder=lowerCAmelCase__ , **lowerCAmelCase__ ) @property def __magic_name__ ( self ) -> Union[str, Any]: return self.encoder_attention_heads @property def __magic_name__ ( self ) -> List[Any]: return self.d_model class snake_case__ ( lowerCamelCase__ ): lowercase__ : int = version.parse('''1.11''' ) @property def __magic_name__ ( self ) -> Optional[int]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def __magic_name__ ( self ) -> Optional[int]: return 1e-5 @property def __magic_name__ ( self ) -> Dict: return 12
354
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def UpperCamelCase ( _A, _A ): """simple docstring""" return math.sqrt(sum(pow(a - b, 2 ) for a, b in zip(_A, _A ) ) ) def UpperCamelCase ( _A, _A ): """simple docstring""" if dataset.ndim != value_array.ndim: __magic_name__ : str = ( """Wrong input data's dimensions... """ f'dataset : {dataset.ndim}, value_array : {value_array.ndim}' ) raise ValueError(_A ) try: if dataset.shape[1] != value_array.shape[1]: __magic_name__ : Optional[Any] = ( """Wrong input data's shape... """ f'dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}' ) raise ValueError(_A ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("""Wrong shape""" ) if dataset.dtype != value_array.dtype: __magic_name__ : List[Any] = ( """Input data have different datatype... """ f'dataset : {dataset.dtype}, value_array : {value_array.dtype}' ) raise TypeError(_A ) __magic_name__ : Dict = [] for value in value_array: __magic_name__ : Tuple = euclidean(_A, dataset[0] ) __magic_name__ : Any = dataset[0].tolist() for dataset_value in dataset[1:]: __magic_name__ : Any = euclidean(_A, _A ) if dist > temp_dist: __magic_name__ : Dict = temp_dist __magic_name__ : Any = dataset_value.tolist() answer.append([vector, dist] ) return answer def UpperCamelCase ( _A, _A ): """simple docstring""" return np.dot(_A, _A ) / (norm(_A ) * norm(_A )) if __name__ == "__main__": import doctest doctest.testmod()
138
0
import copy import random from transformers import CLIPTokenizer class _snake_case ( __snake_case ): '''simple docstring''' def __init__( self: str ,*lowerCamelCase_: int ,**lowerCamelCase_: Optional[Any] ) -> str: super().__init__(*lowerCamelCase_ ,**lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = {} def A__ ( self: str ,lowerCamelCase_: str ,*lowerCamelCase_: List[Any] ,**lowerCamelCase_: Dict ) -> Union[str, Any]: UpperCAmelCase_ : Union[str, Any] = super().add_tokens(lowerCamelCase_ ,*lowerCamelCase_ ,**lowerCamelCase_ ) if num_added_tokens == 0: raise ValueError( F'''The tokenizer already contains the token {placeholder_token}. Please pass a different''' """ `placeholder_token` that is not already in the tokenizer.""" ) def A__ ( self: List[Any] ,lowerCamelCase_: Dict ,*lowerCamelCase_: List[str] ,lowerCamelCase_: List[Any]=1 ,**lowerCamelCase_: List[str] ) -> List[str]: UpperCAmelCase_ : str = [] if num_vec_per_token == 1: self.try_adding_tokens(lowerCamelCase_ ,*lowerCamelCase_ ,**lowerCamelCase_ ) output.append(lowerCamelCase_ ) else: UpperCAmelCase_ : Optional[int] = [] for i in range(lowerCamelCase_ ): UpperCAmelCase_ : int = placeholder_token + F'''_{i}''' self.try_adding_tokens(lowerCamelCase_ ,*lowerCamelCase_ ,**lowerCamelCase_ ) output.append(lowerCamelCase_ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F'''The tokenizer already has placeholder token {token} that can get confused with''' F''' {placeholder_token}keep placeholder tokens independent''' ) UpperCAmelCase_ : Optional[int] = output def A__ ( self: int ,lowerCamelCase_: List[str] ,lowerCamelCase_: int=False ,lowerCamelCase_: Optional[Any]=1.0 ) -> Tuple: if isinstance(lowerCamelCase_ ,lowerCamelCase_ ): UpperCAmelCase_ : List[Any] = [] for i in range(len(lowerCamelCase_ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] ,vector_shuffle=lowerCamelCase_ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: UpperCAmelCase_ : Tuple = self.token_map[placeholder_token] UpperCAmelCase_ : List[str] = tokens[: 1 + int(len(lowerCamelCase_ ) * prop_tokens_to_load )] if vector_shuffle: UpperCAmelCase_ : Union[str, Any] = copy.copy(lowerCamelCase_ ) random.shuffle(lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = text.replace(lowerCamelCase_ ,""" """.join(lowerCamelCase_ ) ) return text def __call__( self: Any ,lowerCamelCase_: str ,*lowerCamelCase_: Dict ,lowerCamelCase_: Optional[Any]=False ,lowerCamelCase_: Union[str, Any]=1.0 ,**lowerCamelCase_: List[Any] ) -> Optional[Any]: return super().__call__( self.replace_placeholder_tokens_in_text( lowerCamelCase_ ,vector_shuffle=lowerCamelCase_ ,prop_tokens_to_load=lowerCamelCase_ ) ,*lowerCamelCase_ ,**lowerCamelCase_ ,) def A__ ( self: Tuple ,lowerCamelCase_: Any ,*lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: Optional[Any]=False ,lowerCamelCase_: str=1.0 ,**lowerCamelCase_: Union[str, Any] ) -> str: return super().encode( self.replace_placeholder_tokens_in_text( lowerCamelCase_ ,vector_shuffle=lowerCamelCase_ ,prop_tokens_to_load=lowerCamelCase_ ) ,*lowerCamelCase_ ,**lowerCamelCase_ ,)
345
import unittest from transformers import MobileBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertModel, ) class _snake_case : '''simple docstring''' def __init__( self: Optional[int] ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: Tuple=13 ,lowerCamelCase_: int=7 ,lowerCamelCase_: Union[str, Any]=True ,lowerCamelCase_: Dict=True ,lowerCamelCase_: str=True ,lowerCamelCase_: Tuple=True ,lowerCamelCase_: int=99 ,lowerCamelCase_: List[str]=64 ,lowerCamelCase_: Tuple=32 ,lowerCamelCase_: List[str]=5 ,lowerCamelCase_: str=4 ,lowerCamelCase_: str=37 ,lowerCamelCase_: Union[str, Any]="gelu" ,lowerCamelCase_: Union[str, Any]=0.1 ,lowerCamelCase_: str=0.1 ,lowerCamelCase_: List[str]=512 ,lowerCamelCase_: Dict=16 ,lowerCamelCase_: List[str]=2 ,lowerCamelCase_: List[str]=0.0_2 ,lowerCamelCase_: Optional[Any]=3 ,lowerCamelCase_: Union[str, Any]=4 ,lowerCamelCase_: str=None ,) -> List[str]: UpperCAmelCase_ : Any = parent UpperCAmelCase_ : List[Any] = batch_size UpperCAmelCase_ : Union[str, Any] = seq_length UpperCAmelCase_ : Optional[int] = is_training UpperCAmelCase_ : Dict = use_input_mask UpperCAmelCase_ : Any = use_token_type_ids UpperCAmelCase_ : Tuple = use_labels UpperCAmelCase_ : List[Any] = vocab_size UpperCAmelCase_ : str = hidden_size UpperCAmelCase_ : List[str] = embedding_size UpperCAmelCase_ : List[Any] = num_hidden_layers UpperCAmelCase_ : List[Any] = num_attention_heads UpperCAmelCase_ : List[Any] = intermediate_size UpperCAmelCase_ : Tuple = hidden_act UpperCAmelCase_ : str = hidden_dropout_prob UpperCAmelCase_ : List[str] = attention_probs_dropout_prob UpperCAmelCase_ : Any = max_position_embeddings UpperCAmelCase_ : List[str] = type_vocab_size UpperCAmelCase_ : Any = type_sequence_label_size UpperCAmelCase_ : Optional[Any] = initializer_range UpperCAmelCase_ : Optional[int] = num_labels UpperCAmelCase_ : Optional[int] = num_choices UpperCAmelCase_ : List[str] = scope def A__ ( self: Any ) -> Optional[int]: UpperCAmelCase_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase_ : List[str] = None if self.use_input_mask: UpperCAmelCase_ : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ : Dict = None if self.use_token_type_ids: UpperCAmelCase_ : str = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) UpperCAmelCase_ : int = None UpperCAmelCase_ : Union[str, Any] = None UpperCAmelCase_ : Union[str, Any] = None if self.use_labels: UpperCAmelCase_ : List[str] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase_ : int = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase_ : Tuple = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A__ ( self: Any ) -> Dict: return MobileBertConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,embedding_size=self.embedding_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=lowerCamelCase_ ,initializer_range=self.initializer_range ,) def A__ ( self: List[Any] ,lowerCamelCase_: str ,lowerCamelCase_: Optional[int] ,lowerCamelCase_: Any ,lowerCamelCase_: List[Any] ,lowerCamelCase_: List[str] ,lowerCamelCase_: str ,lowerCamelCase_: str ) -> int: UpperCAmelCase_ : Any = MobileBertModel(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : List[Any] = model(lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = model(lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ) UpperCAmelCase_ : Tuple = model(lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def A__ ( self: Optional[Any] ,lowerCamelCase_: List[str] ,lowerCamelCase_: List[str] ,lowerCamelCase_: Tuple ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Dict ) -> int: UpperCAmelCase_ : Union[str, Any] = MobileBertForMaskedLM(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Optional[Any] = model(lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ,labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def A__ ( self: str ,lowerCamelCase_: Any ,lowerCamelCase_: Dict ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: List[str] ,lowerCamelCase_: str ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: int ) -> int: UpperCAmelCase_ : List[Any] = MobileBertForNextSentencePrediction(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Union[str, Any] = model( lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ,labels=lowerCamelCase_ ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, 2) ) def A__ ( self: Tuple ,lowerCamelCase_: Tuple ,lowerCamelCase_: Dict ,lowerCamelCase_: List[str] ,lowerCamelCase_: Tuple ,lowerCamelCase_: Tuple ,lowerCamelCase_: Dict ,lowerCamelCase_: Any ) -> Optional[Any]: UpperCAmelCase_ : Tuple = MobileBertForPreTraining(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Optional[int] = model( lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ,labels=lowerCamelCase_ ,next_sentence_label=lowerCamelCase_ ,) self.parent.assertEqual(result.prediction_logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape ,(self.batch_size, 2) ) def A__ ( self: Any ,lowerCamelCase_: Optional[int] ,lowerCamelCase_: Any ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: List[str] ,lowerCamelCase_: Any ,lowerCamelCase_: int ,lowerCamelCase_: List[Any] ) -> List[str]: UpperCAmelCase_ : Optional[Any] = MobileBertForQuestionAnswering(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : int = model( lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ,start_positions=lowerCamelCase_ ,end_positions=lowerCamelCase_ ,) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def A__ ( self: List[str] ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Tuple ,lowerCamelCase_: Any ,lowerCamelCase_: Tuple ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: Any ) -> str: UpperCAmelCase_ : Optional[Any] = self.num_labels UpperCAmelCase_ : Union[str, Any] = MobileBertForSequenceClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Optional[int] = model(lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ,labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def A__ ( self: Union[str, Any] ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: str ,lowerCamelCase_: Dict ,lowerCamelCase_: Any ,lowerCamelCase_: List[str] ) -> Any: UpperCAmelCase_ : str = self.num_labels UpperCAmelCase_ : Optional[int] = MobileBertForTokenClassification(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : List[Any] = model(lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ,labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def A__ ( self: Tuple ,lowerCamelCase_: str ,lowerCamelCase_: int ,lowerCamelCase_: Tuple ,lowerCamelCase_: List[Any] ,lowerCamelCase_: str ,lowerCamelCase_: Optional[int] ,lowerCamelCase_: List[Any] ) -> Union[str, Any]: UpperCAmelCase_ : Union[str, Any] = self.num_choices UpperCAmelCase_ : Tuple = MobileBertForMultipleChoice(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Dict = input_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() UpperCAmelCase_ : Union[str, Any] = token_type_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() UpperCAmelCase_ : str = input_mask.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() UpperCAmelCase_ : Optional[int] = model( lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,token_type_ids=lowerCamelCase_ ,labels=lowerCamelCase_ ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def A__ ( self: List[str] ) -> str: UpperCAmelCase_ : str = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) : Union[str, Any] = config_and_inputs UpperCAmelCase_ : Dict = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _snake_case ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' A__ : Dict = ( ( MobileBertModel, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, ) if is_torch_available() else () ) A__ : List[str] = ( { "feature-extraction": MobileBertModel, "fill-mask": MobileBertForMaskedLM, "question-answering": MobileBertForQuestionAnswering, "text-classification": MobileBertForSequenceClassification, "token-classification": MobileBertForTokenClassification, "zero-shot": MobileBertForSequenceClassification, } if is_torch_available() else {} ) A__ : List[str] = True def A__ ( self: Dict ,lowerCamelCase_: Tuple ,lowerCamelCase_: Tuple ,lowerCamelCase_: int=False ) -> Union[str, Any]: UpperCAmelCase_ : List[Any] = super()._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ,return_labels=lowerCamelCase_ ) if return_labels: if model_class in get_values(lowerCamelCase_ ): UpperCAmelCase_ : Any = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) ,dtype=torch.long ,device=lowerCamelCase_ ) UpperCAmelCase_ : List[str] = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=lowerCamelCase_ ) return inputs_dict def A__ ( self: List[str] ) -> Any: UpperCAmelCase_ : List[str] = MobileBertModelTester(self ) UpperCAmelCase_ : Union[str, Any] = ConfigTester(self ,config_class=lowerCamelCase_ ,hidden_size=37 ) def A__ ( self: Optional[Any] ) -> List[Any]: self.config_tester.run_common_tests() def A__ ( self: List[str] ) -> Optional[Any]: UpperCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*lowerCamelCase_ ) def A__ ( self: Optional[int] ) -> Optional[int]: UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*lowerCamelCase_ ) def A__ ( self: Optional[Any] ) -> Tuple: UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*lowerCamelCase_ ) def A__ ( self: List[Any] ) -> List[str]: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*lowerCamelCase_ ) def A__ ( self: Optional[Any] ) -> Dict: UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*lowerCamelCase_ ) def A__ ( self: Optional[int] ) -> Optional[int]: UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*lowerCamelCase_ ) def A__ ( self: Union[str, Any] ) -> Optional[int]: UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*lowerCamelCase_ ) def A__ ( self: Any ) -> Optional[int]: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*lowerCamelCase_ ) def lowerCamelCase_ ( _a : Union[str, Any] ): '''simple docstring''' return torch.tensor( _a , dtype=torch.long , device=_a , ) UpperCamelCase_ = 1E-3 @require_torch @require_sentencepiece @require_tokenizers class _snake_case ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self: List[Any] ) -> str: UpperCAmelCase_ : Any = MobileBertModel.from_pretrained("""google/mobilebert-uncased""" ).to(lowerCamelCase_ ) UpperCAmelCase_ : str = _long_tensor([[101, 7110, 1005, 1056, 2023, 11333, 17413, 1029, 102]] ) with torch.no_grad(): UpperCAmelCase_ : Union[str, Any] = model(lowerCamelCase_ )[0] UpperCAmelCase_ : Union[str, Any] = torch.Size((1, 9, 512) ) self.assertEqual(output.shape ,lowerCamelCase_ ) UpperCAmelCase_ : Tuple = torch.tensor( [ [ [-2.473_6526e07, 8.269_1656e04, 1.652_1838e05], [-5.754_1704e-01, 3.905_6022e00, 4.401_1507e00], [2.604_7359e00, 1.567_7652e00, -1.732_4188e-01], ] ] ,device=lowerCamelCase_ ,) # MobileBERT results range from 10e0 to 10e8. Even a 0.0000001% difference with a value of 10e8 results in a # ~1 difference, it's therefore not a good idea to measure using addition. # Here, we instead divide the expected result with the result in order to obtain ~1. We then check that the # result is held between bounds: 1 - TOLERANCE < expected_result / result < 1 + TOLERANCE UpperCAmelCase_ : Dict = torch.all((expected_slice / output[..., :3, :3]) >= 1 - TOLERANCE ) UpperCAmelCase_ : Dict = torch.all((expected_slice / output[..., :3, :3]) <= 1 + TOLERANCE ) self.assertTrue(lower_bound and upper_bound )
345
1
'''simple docstring''' import json import os import shutil import tempfile import unittest from transformers import BatchEncoding, CanineTokenizer from transformers.testing_utils import require_tokenizers, require_torch from transformers.tokenization_utils import AddedToken from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin class _lowercase ( _lowercase , unittest.TestCase ): a = CanineTokenizer a = False def lowerCamelCase_ ( self: Optional[Any] ): super().setUp() lowerCamelCase__ : Dict = CanineTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase_ ( self: Dict ): return CanineTokenizer.from_pretrained("""google/canine-s""" ) def lowerCamelCase_ ( self: str , **UpperCamelCase__: str ): lowerCamelCase__ : str = self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) lowerCamelCase__ : Dict = 1_024 return tokenizer @require_torch def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : List[str] = self.canine_tokenizer lowerCamelCase__ : Tuple = ["""Life is like a box of chocolates.""", """You never know what you're gonna get."""] # fmt: off lowerCamelCase__ : Optional[Any] = [57_344, 76, 105, 102, 101, 32, 105, 115, 32, 108, 105, 107, 101, 32, 97, 32, 98, 111, 120, 32, 111, 102, 32, 99, 104, 111, 99, 111, 108, 97, 116, 101, 115, 46, 57_345, 0, 0, 0, 0] # fmt: on lowerCamelCase__ : str = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors="""pt""" ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = list(batch.input_ids.numpy()[0] ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual((2, 39) , batch.input_ids.shape ) self.assertEqual((2, 39) , batch.attention_mask.shape ) @require_torch def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : str = self.canine_tokenizer lowerCamelCase__ : int = ["""Once there was a man.""", """He wrote a test in HuggingFace Tranformers."""] lowerCamelCase__ : Union[str, Any] = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors="""pt""" ) # check if input_ids, attention_mask and token_type_ids are returned self.assertIn("""input_ids""" , UpperCamelCase__ ) self.assertIn("""attention_mask""" , UpperCamelCase__ ) self.assertIn("""token_type_ids""" , UpperCamelCase__ ) @require_torch def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = self.canine_tokenizer lowerCamelCase__ : List[str] = [ """What's the weater?""", """It's about 25 degrees.""", ] lowerCamelCase__ : str = tokenizer( text_target=UpperCamelCase__ , max_length=32 , padding="""max_length""" , truncation=UpperCamelCase__ , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def lowerCamelCase_ ( self: Optional[Any] ): # safety check on max_len default value so we are sure the test works lowerCamelCase__ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test lowerCamelCase__ : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : str = tempfile.mkdtemp() lowerCamelCase__ : Union[str, Any] = """ He is very happy, UNwant\u00E9d,running""" lowerCamelCase__ : Optional[int] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) shutil.rmtree(UpperCamelCase__ ) lowerCamelCase__ : Tuple = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Any = tempfile.mkdtemp() lowerCamelCase__ : str = """ He is very happy, UNwant\u00E9d,running""" lowerCamelCase__ : str = tokenizer.additional_special_tokens # We can add a new special token for Canine as follows: lowerCamelCase__ : int = chr(0xE007 ) additional_special_tokens.append(UpperCamelCase__ ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Tuple = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : int = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertIn(UpperCamelCase__ , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) lowerCamelCase__ : Optional[int] = tokenizer.__class__.from_pretrained(UpperCamelCase__ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = self.get_tokenizers(do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : Tuple = self.get_clean_sequence(UpperCamelCase__ ) # a special token for Canine can be defined as follows: lowerCamelCase__ : Tuple = 0xE005 lowerCamelCase__ : Optional[Any] = chr(UpperCamelCase__ ) tokenizer.add_special_tokens({"""cls_token""": special_token} ) lowerCamelCase__ : List[str] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(len(UpperCamelCase__ ) , 1 ) lowerCamelCase__ : Optional[int] = tokenizer.decode(ids + encoded_special_token , clean_up_tokenization_spaces=UpperCamelCase__ ) lowerCamelCase__ : str = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) lowerCamelCase__ : List[str] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) lowerCamelCase__ : int = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , input_encoded + special_token_id ) lowerCamelCase__ : Dict = tokenizer.decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) self.assertTrue(special_token not in decoded ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : str = self.get_tokenizers(do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : Any = chr(0xE005 ) lowerCamelCase__ : Tuple = chr(0xE006 ) # `add_tokens` method stores special tokens only in `tokenizer.unique_no_split_tokens`. (in tokenization_utils.py) tokenizer.add_tokens([SPECIAL_TOKEN_1] , special_tokens=UpperCamelCase__ ) # `add_special_tokens` method stores special tokens in `tokenizer.additional_special_tokens`, # which also occur in `tokenizer.all_special_tokens`. (in tokenization_utils_base.py) tokenizer.add_special_tokens({"""additional_special_tokens""": [SPECIAL_TOKEN_2]} ) lowerCamelCase__ : int = tokenizer.tokenize(UpperCamelCase__ ) lowerCamelCase__ : int = tokenizer.tokenize(UpperCamelCase__ ) self.assertEqual(len(UpperCamelCase__ ) , 1 ) self.assertEqual(len(UpperCamelCase__ ) , 1 ) self.assertEqual(token_a[0] , UpperCamelCase__ ) self.assertEqual(token_a[0] , UpperCamelCase__ ) @require_tokenizers def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Dict = self.get_tokenizers(do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # a special token for Canine can be defined as follows: lowerCamelCase__ : Tuple = 0xE006 lowerCamelCase__ : Any = chr(UpperCamelCase__ ) lowerCamelCase__ : Dict = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ ) tokenizer.add_special_tokens({"""additional_special_tokens""": [new_token]} ) with tempfile.TemporaryDirectory() as tmp_dir_name: tokenizer.save_pretrained(UpperCamelCase__ ) tokenizer.from_pretrained(UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Union[str, Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Dict = json.load(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Optional[int] = json.load(UpperCamelCase__ ) # a special token for Canine can be defined as follows: lowerCamelCase__ : Union[str, Any] = 0xE006 lowerCamelCase__ : Optional[int] = chr(UpperCamelCase__ ) lowerCamelCase__ : int = [new_token_a] lowerCamelCase__ : Any = [new_token_a] with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCamelCase__ : str = tokenizer_class.from_pretrained(UpperCamelCase__ , extra_ids=0 ) self.assertIn(UpperCamelCase__ , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids([new_token_a] ) ) , ) lowerCamelCase__ : List[Any] = 0xE007 lowerCamelCase__ : Optional[int] = chr(UpperCamelCase__ ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCamelCase__ : Optional[int] = [AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ )] lowerCamelCase__ : str = tokenizer_class.from_pretrained( UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , extra_ids=0 ) self.assertIn(UpperCamelCase__ , tokenizer.additional_special_tokens ) # self.assertIn(new_token_2,tokenizer.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer.convert_ids_to_tokens(tokenizer.convert_tokens_to_ids([new_token_a] ) ) ) @require_tokenizers def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Optional[int] = self.get_tokenizers(do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : Tuple = """hello world""" if self.space_between_special_tokens: lowerCamelCase__ : str = """[CLS] hello world [SEP]""" else: lowerCamelCase__ : str = input lowerCamelCase__ : Optional[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = tokenizer.decode(UpperCamelCase__ , spaces_between_special_tokens=self.space_between_special_tokens ) self.assertIn(UpperCamelCase__ , [output, output.lower()] ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : List[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : List[Any] = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] lowerCamelCase__ : List[Any] = """a""" lowerCamelCase__ : Tuple = ord(UpperCamelCase__ ) for attr in attributes_list: setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [] ) lowerCamelCase__ : int = 0xE006 lowerCamelCase__ : Optional[int] = chr(UpperCamelCase__ ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [additional_special_token_id] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [additional_special_token] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [additional_special_token_id] ) def lowerCamelCase_ ( self: Union[str, Any] ): pass def lowerCamelCase_ ( self: Tuple ): pass def lowerCamelCase_ ( self: str ): pass def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: int ): pass def lowerCamelCase_ ( self: int ): pass def lowerCamelCase_ ( self: List[Any] ): pass def lowerCamelCase_ ( self: Optional[Any] ): pass
355
'''simple docstring''' import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": _A : Optional[int] =pd.read_csv('''sample_data.csv''', header=None) _A : Any =df.shape[:1][0] # If you're using some other dataset input the target column _A : List[str] =df.iloc[:, 1:2] _A : int =actual_data.values.reshape(len_data, 1) _A : Union[str, Any] =MinMaxScaler().fit_transform(actual_data) _A : Optional[int] =10 _A : Union[str, Any] =5 _A : Union[str, Any] =20 _A : str =len_data - periods * look_back _A : List[Any] =actual_data[:division] _A : Optional[Any] =actual_data[division - look_back :] _A , _A : Tuple =[], [] _A , _A : List[str] =[], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) _A : List[Any] =np.array(train_x) _A : str =np.array(test_x) _A : List[Any] =np.array([list(i.ravel()) for i in train_y]) _A : Any =np.array([list(i.ravel()) for i in test_y]) _A : Optional[Any] =Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss='''mean_squared_error''', optimizer='''adam''') _A : Dict =model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) _A : List[str] =model.predict(x_test)
129
0
'''simple docstring''' from ....utils import logging a__ : str = logging.get_logger(__name__) class lowercase_ ( UpperCAmelCase__ ): def __init__( self , a , a=None , a=20_48 ): UpperCamelCase__ = config.__dict__ UpperCamelCase__ = modal_hidden_size if num_labels: UpperCamelCase__ = num_labels
80
def UpperCamelCase ( __magic_name__ : str ) -> list: """simple docstring""" if n_term == "": return [] lowercase__ = [] for temp in range(int(__magic_name__ ) ): series.append(f'''1/{temp + 1}''' if series else """1""" ) return series if __name__ == "__main__": A : Tuple = input('Enter the last number (nth term) of the Harmonic Series') print('Formula of Harmonic Series => 1+1/2+1/3 ..... 1/n') print(harmonic_series(nth_term))
305
0
def lowerCAmelCase_ (lowerCAmelCase__: int ): """simple docstring""" if n == 1 or not isinstance(__a , __a ): return 0 elif n == 2: return 1 else: UpperCAmelCase_: Any = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def lowerCAmelCase_ (lowerCAmelCase__: int ): """simple docstring""" UpperCAmelCase_: Any = 0 UpperCAmelCase_: Dict = 2 while digits < n: index += 1 UpperCAmelCase_: Dict = len(str(fibonacci(__a ) ) ) return index def lowerCAmelCase_ (lowerCAmelCase__: int = 1_0_0_0 ): """simple docstring""" return fibonacci_digits_index(__a ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
365
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. a : Tuple = {'LayoutLMv2Config', 'LayoutLMv3Config'} @is_pipeline_test class _a ( unittest.TestCase ): A = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING A = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: A = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: A = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Dict: UpperCAmelCase_: Dict = ZeroShotClassificationPipeline( model=SCREAMING_SNAKE_CASE_, tokenizer=SCREAMING_SNAKE_CASE_, candidate_labels=["""polics""", """health"""] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Dict: UpperCAmelCase_: Dict = classifier("""Who are you voting for in 2020?""", candidate_labels="""politics""" ) self.assertEqual(SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ )]} ) # No kwarg UpperCAmelCase_: Optional[int] = classifier("""Who are you voting for in 2020?""", ["""politics"""] ) self.assertEqual(SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ )]} ) UpperCAmelCase_: Optional[int] = classifier("""Who are you voting for in 2020?""", candidate_labels=["""politics"""] ) self.assertEqual(SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ )]} ) UpperCAmelCase_: List[Any] = classifier("""Who are you voting for in 2020?""", candidate_labels="""politics, public health""" ) self.assertEqual( SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["""scores"""] ) ), 1.0 ) UpperCAmelCase_: Tuple = classifier("""Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health"""] ) self.assertEqual( SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["""scores"""] ) ), 1.0 ) UpperCAmelCase_: str = classifier( """Who are you voting for in 2020?""", candidate_labels="""politics""", hypothesis_template="""This text is about {}""" ) self.assertEqual(SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ )]} ) # https://github.com/huggingface/transformers/issues/13846 UpperCAmelCase_: Union[str, Any] = classifier(["""I am happy"""], ["""positive""", """negative"""] ) self.assertEqual( SCREAMING_SNAKE_CASE_, [ {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )]} for i in range(1 ) ], ) UpperCAmelCase_: Dict = classifier(["""I am happy""", """I am sad"""], ["""positive""", """negative"""] ) self.assertEqual( SCREAMING_SNAKE_CASE_, [ {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )]} for i in range(2 ) ], ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier("""""", candidate_labels="""politics""" ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier(SCREAMING_SNAKE_CASE_, candidate_labels="""politics""" ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier("""Who are you voting for in 2020?""", candidate_labels="""""" ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier("""Who are you voting for in 2020?""", candidate_labels=SCREAMING_SNAKE_CASE_ ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier( """Who are you voting for in 2020?""", candidate_labels="""politics""", hypothesis_template="""Not formatting template""", ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier( """Who are you voting for in 2020?""", candidate_labels="""politics""", hypothesis_template=SCREAMING_SNAKE_CASE_, ) self.run_entailment_id(SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: UpperCAmelCase_: int = zero_shot_classifier.model.config UpperCAmelCase_: Optional[int] = config.labelaid UpperCAmelCase_: str = zero_shot_classifier.entailment_id UpperCAmelCase_: Union[str, Any] = {"""LABEL_0""": 0, """LABEL_1""": 1, """LABEL_2""": 2} self.assertEqual(zero_shot_classifier.entailment_id, -1 ) UpperCAmelCase_: int = {"""entailment""": 0, """neutral""": 1, """contradiction""": 2} self.assertEqual(zero_shot_classifier.entailment_id, 0 ) UpperCAmelCase_: Dict = {"""ENTAIL""": 0, """NON-ENTAIL""": 1} self.assertEqual(zero_shot_classifier.entailment_id, 0 ) UpperCAmelCase_: Tuple = {"""ENTAIL""": 2, """NEUTRAL""": 1, """CONTR""": 0} self.assertEqual(zero_shot_classifier.entailment_id, 2 ) UpperCAmelCase_: Any = original_labelaid self.assertEqual(SCREAMING_SNAKE_CASE_, zero_shot_classifier.entailment_id ) @require_torch def __snake_case (self ) -> str: UpperCAmelCase_: Any = pipeline( """zero-shot-classification""", model="""sshleifer/tiny-distilbert-base-cased-distilled-squad""", framework="""pt""", ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( """Who are you voting for in 2020?""" * 100, candidate_labels=["""politics""", """public health""", """science"""] ) @require_torch def __snake_case (self ) -> Union[str, Any]: UpperCAmelCase_: str = pipeline( """zero-shot-classification""", model="""sshleifer/tiny-distilbert-base-cased-distilled-squad""", framework="""pt""", ) UpperCAmelCase_: Tuple = zero_shot_classifier( """Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health""", """science"""] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": """Who are you voting for in 2020?""", """labels""": ["""science""", """public health""", """politics"""], """scores""": [0.3_3_3, 0.3_3_3, 0.3_3_3], }, ) @require_tf def __snake_case (self ) -> int: UpperCAmelCase_: List[Any] = pipeline( """zero-shot-classification""", model="""sshleifer/tiny-distilbert-base-cased-distilled-squad""", framework="""tf""", ) UpperCAmelCase_: Optional[Any] = zero_shot_classifier( """Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health""", """science"""] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": """Who are you voting for in 2020?""", """labels""": ["""science""", """public health""", """politics"""], """scores""": [0.3_3_3, 0.3_3_3, 0.3_3_3], }, ) @slow @require_torch def __snake_case (self ) -> Optional[int]: UpperCAmelCase_: List[Any] = pipeline("""zero-shot-classification""", model="""roberta-large-mnli""", framework="""pt""" ) UpperCAmelCase_: Optional[int] = zero_shot_classifier( """Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health""", """science"""] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": """Who are you voting for in 2020?""", """labels""": ["""politics""", """public health""", """science"""], """scores""": [0.9_7_6, 0.0_1_5, 0.0_0_9], }, ) UpperCAmelCase_: Optional[Any] = zero_shot_classifier( """The dominant sequence transduction models are based on complex recurrent or convolutional neural networks""" """ in an encoder-decoder configuration. The best performing models also connect the encoder and decoder""" """ through an attention mechanism. We propose a new simple network architecture, the Transformer, based""" """ solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two""" """ machine translation tasks show these models to be superior in quality while being more parallelizable""" """ and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014""" """ English-to-German translation task, improving over the existing best results, including ensembles by""" """ over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new""" """ single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small""" """ fraction of the training costs of the best models from the literature. We show that the Transformer""" """ generalizes well to other tasks by applying it successfully to English constituency parsing both with""" """ large and limited training data.""", candidate_labels=["""machine learning""", """statistics""", """translation""", """vision"""], multi_label=SCREAMING_SNAKE_CASE_, ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": ( """The dominant sequence transduction models are based on complex recurrent or convolutional neural""" """ networks in an encoder-decoder configuration. The best performing models also connect the""" """ encoder and decoder through an attention mechanism. We propose a new simple network""" """ architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence""" """ and convolutions entirely. Experiments on two machine translation tasks show these models to be""" """ superior in quality while being more parallelizable and requiring significantly less time to""" """ train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,""" """ improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014""" """ English-to-French translation task, our model establishes a new single-model state-of-the-art""" """ BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training""" """ costs of the best models from the literature. We show that the Transformer generalizes well to""" """ other tasks by applying it successfully to English constituency parsing both with large and""" """ limited training data.""" ), """labels""": ["""translation""", """machine learning""", """vision""", """statistics"""], """scores""": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], }, ) @slow @require_tf def __snake_case (self ) -> Optional[int]: UpperCAmelCase_: List[str] = pipeline("""zero-shot-classification""", model="""roberta-large-mnli""", framework="""tf""" ) UpperCAmelCase_: Optional[Any] = zero_shot_classifier( """Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health""", """science"""] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": """Who are you voting for in 2020?""", """labels""": ["""politics""", """public health""", """science"""], """scores""": [0.9_7_6, 0.0_1_5, 0.0_0_9], }, ) UpperCAmelCase_: Any = zero_shot_classifier( """The dominant sequence transduction models are based on complex recurrent or convolutional neural networks""" """ in an encoder-decoder configuration. The best performing models also connect the encoder and decoder""" """ through an attention mechanism. We propose a new simple network architecture, the Transformer, based""" """ solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two""" """ machine translation tasks show these models to be superior in quality while being more parallelizable""" """ and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014""" """ English-to-German translation task, improving over the existing best results, including ensembles by""" """ over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new""" """ single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small""" """ fraction of the training costs of the best models from the literature. We show that the Transformer""" """ generalizes well to other tasks by applying it successfully to English constituency parsing both with""" """ large and limited training data.""", candidate_labels=["""machine learning""", """statistics""", """translation""", """vision"""], multi_label=SCREAMING_SNAKE_CASE_, ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": ( """The dominant sequence transduction models are based on complex recurrent or convolutional neural""" """ networks in an encoder-decoder configuration. The best performing models also connect the""" """ encoder and decoder through an attention mechanism. We propose a new simple network""" """ architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence""" """ and convolutions entirely. Experiments on two machine translation tasks show these models to be""" """ superior in quality while being more parallelizable and requiring significantly less time to""" """ train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,""" """ improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014""" """ English-to-French translation task, our model establishes a new single-model state-of-the-art""" """ BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training""" """ costs of the best models from the literature. We show that the Transformer generalizes well to""" """ other tasks by applying it successfully to English constituency parsing both with large and""" """ limited training data.""" ), """labels""": ["""translation""", """machine learning""", """vision""", """statistics"""], """scores""": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], }, )
82
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : Tuple = { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/config.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/config.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/config.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/config.json', 'bert-base-multilingual-uncased': 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json', 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/config.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/config.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-base-cased-finetuned-mrpc': 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json', 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json', 'bert-base-german-dbmdz-uncased': 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json', 'cl-tohoku/bert-base-japanese': 'https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json', 'cl-tohoku/bert-base-japanese-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json' ), 'wietsedv/bert-base-dutch-cased': 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json', # See all BERT models at https://huggingface.co/models?filter=bert } class _snake_case ( UpperCAmelCase_ ): SCREAMING_SNAKE_CASE__ = 'bert' def __init__( self , _lowerCamelCase=3_0522 , _lowerCamelCase=768 , _lowerCamelCase=12 , _lowerCamelCase=12 , _lowerCamelCase=3072 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=2 , _lowerCamelCase=0.02 , _lowerCamelCase=1e-12 , _lowerCamelCase=0 , _lowerCamelCase="absolute" , _lowerCamelCase=True , _lowerCamelCase=None , **_lowerCamelCase , ): super().__init__(pad_token_id=_lowerCamelCase , **_lowerCamelCase ) a :Union[str, Any] = vocab_size a :Optional[int] = hidden_size a :Optional[Any] = num_hidden_layers a :Union[str, Any] = num_attention_heads a :List[Any] = hidden_act a :Tuple = intermediate_size a :Any = hidden_dropout_prob a :Tuple = attention_probs_dropout_prob a :List[Any] = max_position_embeddings a :Optional[Any] = type_vocab_size a :Any = initializer_range a :Optional[int] = layer_norm_eps a :Optional[int] = position_embedding_type a :Optional[int] = use_cache a :Tuple = classifier_dropout class _snake_case ( UpperCAmelCase_ ): @property def SCREAMING_SNAKE_CASE__ ( self ): if self.task == "multiple-choice": a :Any = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: a :List[str] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
94
'''simple docstring''' import logging import os from .state import PartialState class UpperCAmelCase__ ( logging.LoggerAdapter): @staticmethod def __lowerCamelCase ( lowercase ) -> Dict: __UpperCamelCase = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def __lowerCamelCase ( self , lowercase , lowercase , *lowercase , **lowercase ) -> List[str]: if PartialState._shared_state == {}: raise RuntimeError( """You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.""" ) __UpperCamelCase = kwargs.pop("""main_process_only""" , lowercase ) __UpperCamelCase = kwargs.pop("""in_order""" , lowercase ) if self.isEnabledFor(lowercase ): if self._should_log(lowercase ): __UpperCamelCase , __UpperCamelCase = self.process(lowercase , lowercase ) self.logger.log(lowercase , lowercase , *lowercase , **lowercase ) elif in_order: __UpperCamelCase = PartialState() for i in range(state.num_processes ): if i == state.process_index: __UpperCamelCase , __UpperCamelCase = self.process(lowercase , lowercase ) self.logger.log(lowercase , lowercase , *lowercase , **lowercase ) state.wait_for_everyone() def _lowercase ( __A ,__A = None ): '''simple docstring''' if log_level is None: __UpperCamelCase = os.environ.get("""ACCELERATE_LOG_LEVEL""" ,__A ) __UpperCamelCase = logging.getLogger(__A ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(__A ,{} )
349
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class snake_case_ ( unittest.TestCase ): def __init__( self : Any , _snake_case : Optional[int] , _snake_case : Any=7 , _snake_case : Dict=3 , _snake_case : str=18 , _snake_case : Any=30 , _snake_case : str=400 , _snake_case : List[str]=True , _snake_case : Optional[Any]=None , _snake_case : str=True , _snake_case : Optional[int]=None , _snake_case : Union[str, Any]=True , )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : str = size if size is not None else {"""shortest_edge""": 20} __lowerCAmelCase : Union[str, Any] = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : int = batch_size __lowerCAmelCase : List[str] = num_channels __lowerCAmelCase : Dict = image_size __lowerCAmelCase : Any = min_resolution __lowerCAmelCase : str = max_resolution __lowerCAmelCase : List[Any] = do_resize __lowerCAmelCase : Dict = size __lowerCAmelCase : List[Any] = do_center_crop __lowerCAmelCase : Union[str, Any] = crop_size __lowerCAmelCase : Optional[Any] = do_flip_channel_order def UpperCAmelCase__ ( self : Dict )->Any: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class snake_case_ ( __lowercase ,unittest.TestCase ): A_ = MobileViTImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self : Union[str, Any] )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Any = MobileViTImageProcessingTester(self ) @property def UpperCAmelCase__ ( self : List[Any] )->Dict: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self : Union[str, Any] )->Any: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_snake_case , """do_resize""" ) ) self.assertTrue(hasattr(_snake_case , """size""" ) ) self.assertTrue(hasattr(_snake_case , """do_center_crop""" ) ) self.assertTrue(hasattr(_snake_case , """center_crop""" ) ) self.assertTrue(hasattr(_snake_case , """do_flip_channel_order""" ) ) def UpperCAmelCase__ ( self : int )->List[Any]: '''simple docstring''' __lowerCAmelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) __lowerCAmelCase : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) def UpperCAmelCase__ ( self : str )->str: '''simple docstring''' pass def UpperCAmelCase__ ( self : Optional[int] )->int: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case ) for image in image_inputs: self.assertIsInstance(_snake_case , Image.Image ) # Test not batched input __lowerCAmelCase : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched __lowerCAmelCase : str = image_processing(_snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase__ ( self : Any )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , numpify=_snake_case ) for image in image_inputs: self.assertIsInstance(_snake_case , np.ndarray ) # Test not batched input __lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched __lowerCAmelCase : List[str] = image_processing(_snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase__ ( self : Union[str, Any] )->Any: '''simple docstring''' __lowerCAmelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , torchify=_snake_case ) for image in image_inputs: self.assertIsInstance(_snake_case , torch.Tensor ) # Test not batched input __lowerCAmelCase : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched __lowerCAmelCase : Optional[Any] = image_processing(_snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
354
import random def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :list , SCREAMING_SNAKE_CASE :Dict ) -> tuple: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : List[str] = [], [], [] for element in data: if element < pivot: less.append(SCREAMING_SNAKE_CASE ) elif element > pivot: greater.append(SCREAMING_SNAKE_CASE ) else: equal.append(SCREAMING_SNAKE_CASE ) return less, equal, greater def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :list , SCREAMING_SNAKE_CASE :int ) -> Dict: # index = len(items) // 2 when trying to find the median # (value of index when items is sorted) # invalid input if index >= len(SCREAMING_SNAKE_CASE ) or index < 0: return None __lowerCAmelCase : Union[str, Any] = items[random.randint(0 , len(SCREAMING_SNAKE_CASE ) - 1 )] __lowerCAmelCase : int = 0 __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Tuple = _partition(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = len(SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = len(SCREAMING_SNAKE_CASE ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # must be in larger else: return quick_select(SCREAMING_SNAKE_CASE , index - (m + count) )
232
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available lowercase : Optional[Any] = { "configuration_audio_spectrogram_transformer": [ "AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "ASTConfig", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = [ "AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "ASTForAudioClassification", "ASTModel", "ASTPreTrainedModel", ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Dict = ["ASTFeatureExtractor"] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys lowercase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Union[str, Any] = { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/config.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/config.json", } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """xlnet""" __lowercase = ["""mems"""] __lowercase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase_=3_20_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=24 , lowerCAmelCase_=16 , lowerCAmelCase_=40_96 , lowerCAmelCase_="gelu" , lowerCAmelCase_=True , lowerCAmelCase_="bi" , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-12 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=-1 , lowerCAmelCase_=False , lowerCAmelCase_="last" , lowerCAmelCase_=True , lowerCAmelCase_="tanh" , lowerCAmelCase_=0.1 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=1 , lowerCAmelCase_=2 , **lowerCAmelCase_ , ): """simple docstring""" _snake_case = vocab_size _snake_case = d_model _snake_case = n_layer _snake_case = n_head if d_model % n_head != 0: raise ValueError(F'\'d_model % n_head\' ({d_model % n_head}) should be equal to 0' ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F'`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})' ) _snake_case = d_model // n_head _snake_case = ff_activation _snake_case = d_inner _snake_case = untie_r _snake_case = attn_type _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = dropout _snake_case = mem_len _snake_case = reuse_len _snake_case = bi_data _snake_case = clamp_len _snake_case = same_length _snake_case = summary_type _snake_case = summary_use_proj _snake_case = summary_activation _snake_case = summary_last_dropout _snake_case = start_n_top _snake_case = end_n_top _snake_case = bos_token_id _snake_case = pad_token_id _snake_case = eos_token_id if "use_cache" in kwargs: warnings.warn( 'The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`' ' instead.' , lowerCAmelCase_ , ) _snake_case = kwargs['use_cache'] _snake_case = use_mems_eval _snake_case = use_mems_train super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) @property def lowerCamelCase ( self ): """simple docstring""" logger.info(F'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" raise NotImplementedError( F'The model {self.model_type} is one of the few models that has no sequence length limit.' )
42
1
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCAmelCase_ : '''simple docstring''' def __init__( self , _UpperCAmelCase , _UpperCAmelCase=99 , _UpperCAmelCase=13 , _UpperCAmelCase=16 , _UpperCAmelCase=7 , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=False , _UpperCAmelCase=True , _UpperCAmelCase=2 , _UpperCAmelCase=32 , _UpperCAmelCase=4 , _UpperCAmelCase=4 , _UpperCAmelCase=30 , _UpperCAmelCase=0 , _UpperCAmelCase=1 , _UpperCAmelCase=2 , _UpperCAmelCase=None , ): snake_case_ = parent snake_case_ = batch_size snake_case_ = decoder_seq_length # For common tests snake_case_ = self.decoder_seq_length snake_case_ = is_training snake_case_ = use_attention_mask snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = d_model snake_case_ = d_model snake_case_ = decoder_layers snake_case_ = decoder_layers snake_case_ = decoder_ffn_dim snake_case_ = decoder_attention_heads snake_case_ = decoder_attention_heads snake_case_ = eos_token_id snake_case_ = bos_token_id snake_case_ = pad_token_id snake_case_ = decoder_start_token_id snake_case_ = use_cache snake_case_ = max_position_embeddings snake_case_ = None snake_case_ = decoder_seq_length snake_case_ = 2 snake_case_ = 1 def UpperCamelCase__ ( self ): snake_case_ = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) snake_case_ = None if self.use_attention_mask: snake_case_ = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) snake_case_ = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): snake_case_ = True snake_case_ = TrOCRDecoder(config=_UpperCAmelCase ).to(_UpperCAmelCase ).eval() snake_case_ = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass snake_case_ = model(_UpperCAmelCase , use_cache=_UpperCAmelCase ) snake_case_ = model(_UpperCAmelCase ) snake_case_ = model(_UpperCAmelCase , use_cache=_UpperCAmelCase ) self.parent.assertTrue(len(_UpperCAmelCase ) == len(_UpperCAmelCase ) ) self.parent.assertTrue(len(_UpperCAmelCase ) == len(_UpperCAmelCase ) + 1 ) snake_case_ = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids snake_case_ = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and snake_case_ = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case_ = model(_UpperCAmelCase )['''last_hidden_state'''] snake_case_ = model(_UpperCAmelCase , past_key_values=_UpperCAmelCase )['''last_hidden_state'''] # select random slice snake_case_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case_ = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() snake_case_ = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-3 ) def UpperCamelCase__ ( self ): snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCAmelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): '''simple docstring''' __snake_case = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () __snake_case = (TrOCRForCausalLM,) if is_torch_available() else () __snake_case = {"text-generation": TrOCRForCausalLM} if is_torch_available() else {} __snake_case = True __snake_case = False def UpperCamelCase__ ( self ): snake_case_ = TrOCRStandaloneDecoderModelTester(self , is_training=_UpperCAmelCase ) snake_case_ = ConfigTester(self , config_class=_UpperCAmelCase ) def UpperCamelCase__ ( self ): pass def UpperCamelCase__ ( self ): pass def UpperCamelCase__ ( self ): pass def UpperCamelCase__ ( self ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*_UpperCAmelCase ) def UpperCamelCase__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def UpperCamelCase__ ( self ): pass
267
import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): snake_case_ = ['''a''', '''b''', '''c'''] # Defaults to last layer if both are None snake_case_ , snake_case_ = get_aligned_output_features_output_indices(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , ['''c'''] ) self.assertEqual(_UpperCAmelCase , [2] ) # Out indices set to match out features snake_case_ , snake_case_ = get_aligned_output_features_output_indices(['''a''', '''c'''] , _UpperCAmelCase , _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , ['''a''', '''c'''] ) self.assertEqual(_UpperCAmelCase , [0, 2] ) # Out features set to match out indices snake_case_ , snake_case_ = get_aligned_output_features_output_indices(_UpperCAmelCase , [0, 2] , _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , ['''a''', '''c'''] ) self.assertEqual(_UpperCAmelCase , [0, 2] ) # Out features selected from negative indices snake_case_ , snake_case_ = get_aligned_output_features_output_indices(_UpperCAmelCase , [-3, -1] , _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase , ['''a''', '''c'''] ) self.assertEqual(_UpperCAmelCase , [-3, -1] ) def UpperCamelCase__ ( self ): # Stage names must be set with self.assertRaises(_UpperCAmelCase ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 1) , _UpperCAmelCase ) # Out features must be a list with self.assertRaises(_UpperCAmelCase ): verify_out_features_out_indices(('''a''', '''b''') , (0, 1) , ['''a''', '''b'''] ) # Out features must be a subset of stage names with self.assertRaises(_UpperCAmelCase ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 1) , ['''a'''] ) # Out indices must be a list or tuple with self.assertRaises(_UpperCAmelCase ): verify_out_features_out_indices(_UpperCAmelCase , 0 , ['''a''', '''b'''] ) # Out indices must be a subset of stage names with self.assertRaises(_UpperCAmelCase ): verify_out_features_out_indices(_UpperCAmelCase , (0, 1) , ['''a'''] ) # Out features and out indices must be the same length with self.assertRaises(_UpperCAmelCase ): verify_out_features_out_indices(['''a''', '''b'''] , (0,) , ['''a''', '''b''', '''c'''] ) # Out features should match out indices with self.assertRaises(_UpperCAmelCase ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 2) , ['''a''', '''b''', '''c'''] ) # Out features and out indices should be in order with self.assertRaises(_UpperCAmelCase ): verify_out_features_out_indices(['''b''', '''a'''] , (0, 1) , ['''a''', '''b'''] ) # Check passes with valid inputs verify_out_features_out_indices(['''a''', '''b''', '''d'''] , (0, 1, -1) , ['''a''', '''b''', '''c''', '''d'''] ) def UpperCamelCase__ ( self ): snake_case_ = BackboneMixin() snake_case_ = ['''a''', '''b''', '''c'''] snake_case_ = ['''a''', '''c'''] snake_case_ = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ['''a''', '''c'''] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly snake_case_ = ['''a''', '''b'''] self.assertEqual(backbone.out_features , ['''a''', '''b'''] ) self.assertEqual(backbone.out_indices , [0, 1] ) snake_case_ = [-3, -1] self.assertEqual(backbone.out_features , ['''a''', '''c'''] ) self.assertEqual(backbone.out_indices , [-3, -1] )
267
1
UpperCAmelCase__ : Tuple = [ """DownloadConfig""", """DownloadManager""", """DownloadMode""", """StreamingDownloadManager""", ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
245
from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING UpperCAmelCase__ : Tuple = logging.get_logger(__name__) @add_end_docstrings(UpperCAmelCase ) class a__ ( UpperCAmelCase ): """simple docstring""" def __init__( self : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : Dict ) ->List[str]: """simple docstring""" super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) requires_backends(self , """vision""" ) self.check_model_type(UpperCAmelCase__ ) def __call__( self : Any , UpperCAmelCase__ : Union[str, List[str], "Image.Image", List["Image.Image"]] , **UpperCAmelCase__ : List[str] ) ->Any: """simple docstring""" return super().__call__(UpperCAmelCase__ , **UpperCAmelCase__ ) def _lowercase ( self : Dict , **UpperCAmelCase__ : Optional[int] ) ->Any: """simple docstring""" return {}, {}, {} def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : Tuple ) ->Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = load_image(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Any = image.size SCREAMING_SNAKE_CASE : Tuple = self.image_processor(images=UpperCAmelCase__ , return_tensors=self.framework ) return model_inputs def _lowercase ( self : int , UpperCAmelCase__ : Any ) ->List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = self.model(**UpperCAmelCase__ ) return model_outputs def _lowercase ( self : Tuple , UpperCAmelCase__ : Union[str, Any] ) ->str: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = model_outputs.predicted_depth SCREAMING_SNAKE_CASE : Any = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) , size=self.image_size[::-1] , mode="""bicubic""" , align_corners=UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Dict = prediction.squeeze().cpu().numpy() SCREAMING_SNAKE_CASE : str = (output * 2_5_5 / np.max(UpperCAmelCase__ )).astype("""uint8""" ) SCREAMING_SNAKE_CASE : Tuple = Image.fromarray(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Tuple = {} SCREAMING_SNAKE_CASE : Dict = predicted_depth SCREAMING_SNAKE_CASE : Optional[int] = depth return output_dict
245
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __SCREAMING_SNAKE_CASE ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ :Any = StableDiffusionSAGPipeline lowerCamelCase_ :Dict = TEXT_TO_IMAGE_PARAMS lowerCamelCase_ :Dict = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase_ :Union[str, Any] = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase_ :Optional[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase_ :Any = False def _UpperCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase_ : Optional[Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) UpperCAmelCase_ : List[Any] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) torch.manual_seed(0 ) UpperCAmelCase_ : Dict = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) UpperCAmelCase_ : Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) UpperCAmelCase_ : Tuple = CLIPTextModel(snake_case_ ) UpperCAmelCase_ : Tuple = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) UpperCAmelCase_ : Tuple = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def _UpperCamelCase ( self , snake_case_ , snake_case_=0 ): '''simple docstring''' if str(snake_case_ ).startswith('mps' ): UpperCAmelCase_ : Optional[Any] = torch.manual_seed(snake_case_ ) else: UpperCAmelCase_ : List[str] = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) UpperCAmelCase_ : List[Any] = { 'prompt': '.', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 1.0, 'sag_scale': 1.0, 'output_type': 'numpy', } return inputs def _UpperCamelCase ( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def _UpperCamelCase ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : List[Any] = StableDiffusionSAGPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' ) UpperCAmelCase_ : Any = sag_pipe.to(snake_case_ ) sag_pipe.set_progress_bar_config(disable=snake_case_ ) UpperCAmelCase_ : Union[str, Any] = '.' UpperCAmelCase_ : Tuple = torch.manual_seed(0 ) UpperCAmelCase_ : Dict = sag_pipe( [prompt] , generator=snake_case_ , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=2_0 , output_type='np' ) UpperCAmelCase_ : Any = output.images UpperCAmelCase_ : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) UpperCAmelCase_ : Tuple = np.array([0.15_68, 0.17_38, 0.16_95, 0.16_93, 0.15_07, 0.17_05, 0.15_47, 0.17_51, 0.19_49] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-2 def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Any = StableDiffusionSAGPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) UpperCAmelCase_ : Optional[int] = sag_pipe.to(snake_case_ ) sag_pipe.set_progress_bar_config(disable=snake_case_ ) UpperCAmelCase_ : List[Any] = '.' UpperCAmelCase_ : List[str] = torch.manual_seed(0 ) UpperCAmelCase_ : str = sag_pipe( [prompt] , generator=snake_case_ , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=2_0 , output_type='np' ) UpperCAmelCase_ : List[str] = output.images UpperCAmelCase_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) UpperCAmelCase_ : Tuple = np.array([0.34_59, 0.28_76, 0.25_37, 0.30_02, 0.26_71, 0.21_60, 0.30_26, 0.22_62, 0.23_71] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-2 def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : str = StableDiffusionSAGPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) UpperCAmelCase_ : str = sag_pipe.to(snake_case_ ) sag_pipe.set_progress_bar_config(disable=snake_case_ ) UpperCAmelCase_ : Union[str, Any] = '.' UpperCAmelCase_ : List[Any] = torch.manual_seed(0 ) UpperCAmelCase_ : List[str] = sag_pipe( [prompt] , width=7_6_8 , height=5_1_2 , generator=snake_case_ , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=2_0 , output_type='np' , ) UpperCAmelCase_ : List[str] = output.images assert image.shape == (1, 5_1_2, 7_6_8, 3)
274
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __SCREAMING_SNAKE_CASE ( lowerCamelCase_ ): '''simple docstring''' lowerCamelCase_ :Tuple = ['''image_processor''', '''tokenizer'''] lowerCamelCase_ :Optional[Any] = '''ViTImageProcessor''' lowerCamelCase_ :int = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): '''simple docstring''' UpperCAmelCase_ : Dict = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case_ , ) UpperCAmelCase_ : int = kwargs.pop('feature_extractor' ) UpperCAmelCase_ : str = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case_ , snake_case_ ) def __call__( self , snake_case_=None , snake_case_=None , snake_case_=None , snake_case_=None , **snake_case_ ): '''simple docstring''' if text is None and visual_prompt is None and images is None: raise ValueError('You have to specify either text, visual prompt or images.' ) if text is not None and visual_prompt is not None: raise ValueError('You have to specify exactly one type of prompt. Either text or visual prompt.' ) if text is not None: UpperCAmelCase_ : Optional[int] = self.tokenizer(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if visual_prompt is not None: UpperCAmelCase_ : Optional[Any] = self.image_processor(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if images is not None: UpperCAmelCase_ : int = self.image_processor(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if visual_prompt is not None and images is not None: UpperCAmelCase_ : Tuple = { 'pixel_values': image_features.pixel_values, 'conditional_pixel_values': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: UpperCAmelCase_ : Optional[int] = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: UpperCAmelCase_ : Dict = { 'conditional_pixel_values': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**snake_case_ ) , tensor_type=snake_case_ ) def _UpperCamelCase ( self , *snake_case_ , **snake_case_ ): '''simple docstring''' return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _UpperCamelCase ( self , *snake_case_ , **snake_case_ ): '''simple docstring''' return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def _UpperCamelCase ( self ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case_ , ) return self.image_processor_class @property def _UpperCamelCase ( self ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case_ , ) return self.image_processor
274
1
"""simple docstring""" from math import sqrt def __A (_SCREAMING_SNAKE_CASE ) ->bool: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and ( number >= 0 ), "'number' must been an int and positive" lowerCAmelCase__ :Tuple = True # 0 and 1 are none primes. if number <= 1: lowerCAmelCase__ :str = False for divisor in range(2 , int(round(sqrt(_SCREAMING_SNAKE_CASE ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowerCAmelCase__ :Tuple = False break # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), "'status' must been from type bool" return status def __A (_SCREAMING_SNAKE_CASE ) ->Dict: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowerCAmelCase__ :Optional[int] = list(range(2 , n + 1 ) ) lowerCAmelCase__ :List[Any] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(_SCREAMING_SNAKE_CASE ) ): for j in range(i + 1 , len(_SCREAMING_SNAKE_CASE ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowerCAmelCase__ :str = 0 # filters actual prime numbers. lowerCAmelCase__ :Union[str, Any] = [x for x in begin_list if x != 0] # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), "'ans' must been from type list" return ans def __A (_SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (n > 2), "'N' must been an int and > 2" lowerCAmelCase__ :Dict = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(_SCREAMING_SNAKE_CASE ): ans.append(_SCREAMING_SNAKE_CASE ) # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), "'ans' must been from type list" return ans def __A (_SCREAMING_SNAKE_CASE ) ->List[Any]: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and number >= 0, "'number' must been an int and >= 0" lowerCAmelCase__ :List[str] = [] # this list will be returns of the function. # potential prime number factors. lowerCAmelCase__ :Optional[Any] = 2 lowerCAmelCase__ :Any = number if number == 0 or number == 1: ans.append(_SCREAMING_SNAKE_CASE ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(_SCREAMING_SNAKE_CASE ): while quotient != 1: if is_prime(_SCREAMING_SNAKE_CASE ) and (quotient % factor == 0): ans.append(_SCREAMING_SNAKE_CASE ) quotient /= factor else: factor += 1 else: ans.append(_SCREAMING_SNAKE_CASE ) # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), "'ans' must been from type list" return ans def __A (_SCREAMING_SNAKE_CASE ) ->List[Any]: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase__ :Any = 0 # prime factorization of 'number' lowerCAmelCase__ :Dict = prime_factorization(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Tuple = max(_SCREAMING_SNAKE_CASE ) # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), "'ans' must been from type int" return ans def __A (_SCREAMING_SNAKE_CASE ) ->List[Any]: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase__ :int = 0 # prime factorization of 'number' lowerCAmelCase__ :Optional[int] = prime_factorization(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[Any] = min(_SCREAMING_SNAKE_CASE ) # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), "'ans' must been from type int" return ans def __A (_SCREAMING_SNAKE_CASE ) ->Dict: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), "'number' must been an int" assert isinstance(number % 2 == 0 , _SCREAMING_SNAKE_CASE ), "compare bust been from type bool" return number % 2 == 0 def __A (_SCREAMING_SNAKE_CASE ) ->Tuple: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), "'number' must been an int" assert isinstance(number % 2 != 0 , _SCREAMING_SNAKE_CASE ), "compare bust been from type bool" return number % 2 != 0 def __A (_SCREAMING_SNAKE_CASE ) ->Optional[int]: """simple docstring""" assert ( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (number > 2) and is_even(_SCREAMING_SNAKE_CASE ) ), "'number' must been an int, even and > 2" lowerCAmelCase__ :List[str] = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowerCAmelCase__ :Optional[int] = get_prime_numbers(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Any = len(_SCREAMING_SNAKE_CASE ) # run variable for while-loops. lowerCAmelCase__ :Tuple = 0 lowerCAmelCase__ :Optional[Any] = None # exit variable. for break up the loops lowerCAmelCase__ :int = True while i < len_pn and loop: lowerCAmelCase__ :Optional[int] = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowerCAmelCase__ :Union[str, Any] = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (len(_SCREAMING_SNAKE_CASE ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" assert ( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase__ :Dict = 0 while numbera != 0: lowerCAmelCase__ :Dict = numbera % numbera lowerCAmelCase__ :Optional[int] = numbera lowerCAmelCase__ :Optional[Any] = rest # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: """simple docstring""" assert ( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase__ :Union[str, Any] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowerCAmelCase__ :List[Any] = prime_factorization(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :str = prime_factorization(_SCREAMING_SNAKE_CASE ) elif numbera == 1 or numbera == 1: lowerCAmelCase__ :List[str] = [] lowerCAmelCase__ :Optional[int] = [] lowerCAmelCase__ :List[Any] = max(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[Any] = 0 lowerCAmelCase__ :Tuple = 0 lowerCAmelCase__ :Optional[int] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowerCAmelCase__ :Optional[int] = prime_fac_a.count(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :int = prime_fac_a.count(_SCREAMING_SNAKE_CASE ) for _ in range(max(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ): ans *= n else: lowerCAmelCase__ :List[str] = prime_fac_a.count(_SCREAMING_SNAKE_CASE ) for _ in range(_SCREAMING_SNAKE_CASE ): ans *= n done.append(_SCREAMING_SNAKE_CASE ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowerCAmelCase__ :Tuple = prime_fac_a.count(_SCREAMING_SNAKE_CASE ) for _ in range(_SCREAMING_SNAKE_CASE ): ans *= n done.append(_SCREAMING_SNAKE_CASE ) # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def __A (_SCREAMING_SNAKE_CASE ) ->Optional[int]: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (n >= 0), "'number' must been a positive int" lowerCAmelCase__ :Optional[Any] = 0 lowerCAmelCase__ :Tuple = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(_SCREAMING_SNAKE_CASE ): ans += 1 # precondition assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and is_prime( _SCREAMING_SNAKE_CASE ), "'ans' must been a prime number and from type int" return ans def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: """simple docstring""" assert ( is_prime(_SCREAMING_SNAKE_CASE ) and is_prime(_SCREAMING_SNAKE_CASE ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowerCAmelCase__ :Tuple = p_number_a + 1 # jump to the next number lowerCAmelCase__ :Any = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(_SCREAMING_SNAKE_CASE ): number += 1 while number < p_number_a: ans.append(_SCREAMING_SNAKE_CASE ) number += 1 # fetch the next prime number. while not is_prime(_SCREAMING_SNAKE_CASE ): number += 1 # precondition assert ( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and ans[0] != p_number_a and ans[len(_SCREAMING_SNAKE_CASE ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def __A (_SCREAMING_SNAKE_CASE ) ->Dict: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (n >= 1), "'n' must been int and >= 1" lowerCAmelCase__ :int = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(_SCREAMING_SNAKE_CASE ) # precondition assert ans[0] == 1 and ans[len(_SCREAMING_SNAKE_CASE ) - 1] == n, "Error in function getDivisiors(...)" return ans def __A (_SCREAMING_SNAKE_CASE ) ->Optional[Any]: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and ( number > 1 ), "'number' must been an int and >= 1" lowerCAmelCase__ :List[Any] = get_divisors(_SCREAMING_SNAKE_CASE ) # precondition assert ( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (divisors[0] == 1) and (divisors[len(_SCREAMING_SNAKE_CASE ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" assert ( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowerCAmelCase__ :Any = gcd(abs(_SCREAMING_SNAKE_CASE ) , abs(_SCREAMING_SNAKE_CASE ) ) # precondition assert ( isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def __A (_SCREAMING_SNAKE_CASE ) ->Any: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (n >= 0), "'n' must been a int and >= 0" lowerCAmelCase__ :str = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def __A (_SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and (n >= 0), "'n' must been an int and >= 0" lowerCAmelCase__ :int = 0 lowerCAmelCase__ :Any = 1 lowerCAmelCase__ :Tuple = 1 # this will be return for _ in range(n - 1 ): lowerCAmelCase__ :Any = ans ans += fiba lowerCAmelCase__ :int = tmp return ans
293
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ : Dict = logging.get_logger(__name__) lowerCAmelCase_ : int = { 'bigcode/gpt_bigcode-santacoder': 'https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json', } class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a ='gpt_bigcode' __a =['past_key_values'] __a ={ 'hidden_size': 'n_embd', 'max_position_embeddings': 'n_positions', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Optional[Any] , __a : Tuple=5_02_57 , __a : str=10_24 , __a : Dict=7_68 , __a : Tuple=12 , __a : str=12 , __a : Optional[int]=None , __a : Dict="gelu_pytorch_tanh" , __a : Tuple=0.1 , __a : Tuple=0.1 , __a : Union[str, Any]=0.1 , __a : Tuple=1e-5 , __a : str=0.02 , __a : Dict=True , __a : Union[str, Any]=True , __a : Optional[int]=5_02_56 , __a : Optional[int]=5_02_56 , __a : Union[str, Any]=True , __a : Dict=True , __a : Union[str, Any]=True , **__a : List[Any] , ): _a = vocab_size _a = n_positions _a = n_embd _a = n_layer _a = n_head _a = n_inner _a = activation_function _a = resid_pdrop _a = embd_pdrop _a = attn_pdrop _a = layer_norm_epsilon _a = initializer_range _a = scale_attn_weights _a = use_cache _a = attention_softmax_in_fpaa _a = scale_attention_softmax_in_fpaa _a = multi_query _a = bos_token_id _a = eos_token_id super().__init__(bos_token_id=__a , eos_token_id=__a , **__a )
63
0
from __future__ import annotations def lowerCAmelCase_ ( _lowercase : float , _lowercase : float , _lowercase : float , ) -> tuple[str, float]: """simple docstring""" if (stress, tangential_force, area).count(0) != 1: raise ValueError("""You cannot supply more or less than 2 values""") elif stress < 0: raise ValueError("""Stress cannot be negative""") elif tangential_force < 0: raise ValueError("""Tangential Force cannot be negative""") elif area < 0: raise ValueError("""Area cannot be negative""") elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
266
import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _lowercase : List[str] =logging.get_logger(__name__) _lowercase : List[str] ={ "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } _lowercase : Optional[Any] ={ "vocab_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json"}, "merges_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt"}, "tokenizer_config_file": { "facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json" }, } _lowercase : int ={"facebook/blenderbot-3B": 128} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def lowerCAmelCase_ ( ) -> Optional[int]: """simple docstring""" a__ : List[str] = ( list(range(ord("""!""") , ord("""~""") + 1)) + list(range(ord("""¡""") , ord("""¬""") + 1)) + list(range(ord("""®""") , ord("""ÿ""") + 1)) ) a__ : Optional[Any] = bs[:] a__ : List[Any] = 0 for b in range(2**8): if b not in bs: bs.append(_lowercase) cs.append(2**8 + n) n += 1 a__ : Tuple = [chr(_lowercase) for n in cs] return dict(zip(_lowercase , _lowercase)) def lowerCAmelCase_ ( _lowercase : Tuple) -> List[str]: """simple docstring""" a__ : int = set() a__ : Tuple = word[0] for char in word[1:]: pairs.add((prev_char, char)) a__ : Any = char return pairs class snake_case__ (A__ ): """simple docstring""" __lowerCAmelCase :str = VOCAB_FILES_NAMES __lowerCAmelCase :int = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase :int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase :Optional[Any] = ["input_ids", "attention_mask"] def __init__( self , __lowercase , __lowercase , __lowercase="replace" , __lowercase="<s>" , __lowercase="</s>" , __lowercase="</s>" , __lowercase="<s>" , __lowercase="<unk>" , __lowercase="<pad>" , __lowercase="<mask>" , __lowercase=False , **__lowercase , ) -> List[Any]: """simple docstring""" a__ : Any = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else bos_token a__ : Optional[Any] = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else eos_token a__ : Tuple = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else sep_token a__ : Any = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else cls_token a__ : List[Any] = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else unk_token a__ : str = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it a__ : List[str] = AddedToken(__lowercase , lstrip=__lowercase , rstrip=__lowercase ) if isinstance(__lowercase , __lowercase ) else mask_token super().__init__( errors=__lowercase , bos_token=__lowercase , eos_token=__lowercase , unk_token=__lowercase , sep_token=__lowercase , cls_token=__lowercase , pad_token=__lowercase , mask_token=__lowercase , add_prefix_space=__lowercase , **__lowercase , ) with open(__lowercase , encoding="""utf-8""" ) as vocab_handle: a__ : str = json.load(__lowercase ) a__ : Dict = {v: k for k, v in self.encoder.items()} a__ : Any = errors # how to handle errors in decoding a__ : Union[str, Any] = bytes_to_unicode() a__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(__lowercase , encoding="""utf-8""" ) as merges_handle: a__ : List[Any] = merges_handle.read().split("""\n""" )[1:-1] a__ : Union[str, Any] = [tuple(merge.split() ) for merge in bpe_merges] a__ : str = dict(zip(__lowercase , range(len(__lowercase ) ) ) ) a__ : Optional[Any] = {} a__ : Optional[Any] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions a__ : Union[str, Any] = re.compile(r"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" return len(self.encoder ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Tuple: """simple docstring""" if token in self.cache: return self.cache[token] a__ : List[Any] = tuple(__lowercase ) a__ : Optional[int] = get_pairs(__lowercase ) if not pairs: return token while True: a__ : List[Any] = min(__lowercase , key=lambda __lowercase : self.bpe_ranks.get(__lowercase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break a__ , a__ : Dict = bigram a__ : List[Any] = [] a__ : int = 0 while i < len(__lowercase ): try: a__ : str = word.index(__lowercase , __lowercase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) a__ : Optional[int] = j if word[i] == first and i < len(__lowercase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 a__ : List[Any] = tuple(__lowercase ) a__ : Any = new_word if len(__lowercase ) == 1: break else: a__ : List[Any] = get_pairs(__lowercase ) a__ : Optional[Any] = """ """.join(__lowercase ) a__ : Tuple = word return word def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Union[str, Any]: """simple docstring""" a__ : int = [] for token in re.findall(self.pat , __lowercase ): a__ : List[Any] = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(__lowercase ).split(""" """ ) ) return bpe_tokens def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> List[Any]: """simple docstring""" return self.encoder.get(__lowercase , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Tuple: """simple docstring""" return self.decoder.get(__lowercase ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> int: """simple docstring""" a__ : Union[str, Any] = """""".join(__lowercase ) a__ : Tuple = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" , errors=self.errors ) return text def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__lowercase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return a__ : int = os.path.join( __lowercase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) a__ : Optional[Any] = os.path.join( __lowercase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(__lowercase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__lowercase , ensure_ascii=__lowercase ) + """\n""" ) a__ : str = 0 with open(__lowercase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __lowercase : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) a__ : Tuple = token_index writer.write(""" """.join(__lowercase ) + """\n""" ) index += 1 return vocab_file, merge_file def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase = None , __lowercase = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowercase , token_ids_a=__lowercase , already_has_special_tokens=__lowercase ) if token_ids_a is None: return [1] + ([0] * len(__lowercase )) + [1] return [1] + ([0] * len(__lowercase )) + [1, 1] + ([0] * len(__lowercase )) + [1] def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase = None ) -> List[int]: """simple docstring""" a__ : Any = [self.sep_token_id] a__ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase=False , **__lowercase ) -> int: """simple docstring""" a__ : Tuple = kwargs.pop("""add_prefix_space""" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(__lowercase ) > 0 and not text[0].isspace()): a__ : Union[str, Any] = """ """ + text return (text, kwargs) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase = None ) -> List[str]: """simple docstring""" return token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> List[int]: """simple docstring""" a__ : List[Any] = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(""" """ + text ) else: # Generated responses should contain them already. inputs.append(__lowercase ) a__ : Optional[int] = """ """.join(__lowercase ) a__ : Any = self.encode(__lowercase ) if len(__lowercase ) > self.model_max_length: a__ : List[str] = input_ids[-self.model_max_length :] logger.warning(F'''Trimmed input from conversation as it was longer than {self.model_max_length} tokens.''' ) return input_ids
266
1
import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" a_ = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING a_ = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def lowercase ( self : str , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : List[str] ) -> List[str]: __lowerCAmelCase = TextaTextGenerationPipeline(model=__lowerCamelCase , tokenizer=__lowerCamelCase ) return generator, ["Something to write", "Something else"] def lowercase ( self : Tuple , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : int ) -> str: __lowerCAmelCase = generator('Something there' ) self.assertEqual(__lowerCamelCase , [{'generated_text': ANY(__lowerCamelCase )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]['generated_text'].startswith('Something there' ) ) __lowerCAmelCase = generator(['This is great !', 'Something else'] , num_return_sequences=2 , do_sample=__lowerCamelCase ) self.assertEqual( __lowerCamelCase , [ [{'generated_text': ANY(__lowerCamelCase )}, {'generated_text': ANY(__lowerCamelCase )}], [{'generated_text': ANY(__lowerCamelCase )}, {'generated_text': ANY(__lowerCamelCase )}], ] , ) __lowerCAmelCase = generator( ['This is great !', 'Something else'] , num_return_sequences=2 , batch_size=2 , do_sample=__lowerCamelCase ) self.assertEqual( __lowerCamelCase , [ [{'generated_text': ANY(__lowerCamelCase )}, {'generated_text': ANY(__lowerCamelCase )}], [{'generated_text': ANY(__lowerCamelCase )}, {'generated_text': ANY(__lowerCamelCase )}], ] , ) with self.assertRaises(__lowerCamelCase ): generator(4 ) @require_torch def lowercase ( self : Union[str, Any] ) -> Tuple: __lowerCAmelCase = pipeline('text2text-generation' , model='patrickvonplaten/t5-tiny-random' , framework='pt' ) # do_sample=False necessary for reproducibility __lowerCAmelCase = generator('Something there' , do_sample=__lowerCamelCase ) self.assertEqual(__lowerCamelCase , [{'generated_text': ''}] ) __lowerCAmelCase = 3 __lowerCAmelCase = generator( 'Something there' , num_return_sequences=__lowerCamelCase , num_beams=__lowerCamelCase , ) __lowerCAmelCase = [ {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': ''}, ] self.assertEqual(__lowerCamelCase , __lowerCamelCase ) __lowerCAmelCase = generator('This is a test' , do_sample=__lowerCamelCase , num_return_sequences=2 , return_tensors=__lowerCamelCase ) self.assertEqual( __lowerCamelCase , [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ] , ) __lowerCAmelCase = generator.model.config.eos_token_id __lowerCAmelCase = '<pad>' __lowerCAmelCase = generator( ['This is a test', 'This is a second test'] , do_sample=__lowerCamelCase , num_return_sequences=2 , batch_size=2 , return_tensors=__lowerCamelCase , ) self.assertEqual( __lowerCamelCase , [ [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], ] , ) @require_tf def lowercase ( self : Optional[Any] ) -> List[Any]: __lowerCAmelCase = pipeline('text2text-generation' , model='patrickvonplaten/t5-tiny-random' , framework='tf' ) # do_sample=False necessary for reproducibility __lowerCAmelCase = generator('Something there' , do_sample=__lowerCamelCase ) self.assertEqual(__lowerCamelCase , [{'generated_text': ''}] )
284
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer a__: Dict = logging.get_logger(__name__) a__: str = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} a__: Any = { 'vocab_file': { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt' ), 'distilbert-base-german-cased': 'https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt', 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json' ), 'distilbert-base-german-cased': ( 'https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json' ), 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json' ), }, } a__: Any = { 'distilbert-base-uncased': 512, 'distilbert-base-uncased-distilled-squad': 512, 'distilbert-base-cased': 512, 'distilbert-base-cased-distilled-squad': 512, 'distilbert-base-german-cased': 512, 'distilbert-base-multilingual-cased': 512, } a__: Optional[Any] = { 'distilbert-base-uncased': {'do_lower_case': True}, 'distilbert-base-uncased-distilled-squad': {'do_lower_case': True}, 'distilbert-base-cased': {'do_lower_case': False}, 'distilbert-base-cased-distilled-squad': {'do_lower_case': False}, 'distilbert-base-german-cased': {'do_lower_case': False}, 'distilbert-base-multilingual-cased': {'do_lower_case': False}, } class SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = PRETRAINED_INIT_CONFIGURATION __SCREAMING_SNAKE_CASE = ['''input_ids''', '''attention_mask'''] __SCREAMING_SNAKE_CASE = DistilBertTokenizer def __init__( self,__lowerCamelCase=None,__lowerCamelCase=None,__lowerCamelCase=True,__lowerCamelCase="[UNK]",__lowerCamelCase="[SEP]",__lowerCamelCase="[PAD]",__lowerCamelCase="[CLS]",__lowerCamelCase="[MASK]",__lowerCamelCase=True,__lowerCamelCase=None,**__lowerCamelCase,): super().__init__( __lowerCamelCase,tokenizer_file=__lowerCamelCase,do_lower_case=__lowerCamelCase,unk_token=__lowerCamelCase,sep_token=__lowerCamelCase,pad_token=__lowerCamelCase,cls_token=__lowerCamelCase,mask_token=__lowerCamelCase,tokenize_chinese_chars=__lowerCamelCase,strip_accents=__lowerCamelCase,**__lowerCamelCase,) A__ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''',__lowerCamelCase ) != do_lower_case or normalizer_state.get('''strip_accents''',__lowerCamelCase ) != strip_accents or normalizer_state.get('''handle_chinese_chars''',__lowerCamelCase ) != tokenize_chinese_chars ): A__ = getattr(__lowerCamelCase,normalizer_state.pop('''type''' ) ) A__ = do_lower_case A__ = strip_accents A__ = tokenize_chinese_chars A__ = normalizer_class(**__lowerCamelCase ) A__ = do_lower_case def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase=None ): A__ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase = None ): A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase = None ): A__ = self._tokenizer.model.save(__lowerCamelCase,name=__lowerCamelCase ) return tuple(__lowerCamelCase )
193
0
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class a__ ( unittest.TestCase ): def __init__( self , _A , _A=1_3 , _A=7 , _A=True , _A=True , _A=True , _A=True , _A=9_9 , _A=3_2 , _A=5 , _A=4 , _A=3_7 , _A="gelu" , _A=0.1 , _A=0.1 , _A=5_1_2 , _A=1_6 , _A=2 , _A=0.02 , _A=4 , ): """simple docstring""" __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_attention_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = num_choices def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_attention_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = None if self.use_token_type_ids: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_A , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class a__ ( snake_case__ , unittest.TestCase ): _a : List[str] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = FlaxAlbertModelTester(self ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_class_name in self.all_model_classes: __lowerCAmelCase = model_class_name.from_pretrained("albert-base-v2" ) __lowerCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(_A ) @require_flax class a__ ( unittest.TestCase ): @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = FlaxAlbertModel.from_pretrained("albert-base-v2" ) __lowerCAmelCase = np.array([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) __lowerCAmelCase = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) __lowerCAmelCase = model(_A , attention_mask=_A )[0] __lowerCAmelCase = (1, 1_1, 7_6_8) self.assertEqual(output.shape , _A ) __lowerCAmelCase = np.array( [[[-0.65_13, 1.50_35, -0.27_66], [-0.65_15, 1.50_46, -0.27_80], [-0.65_12, 1.50_49, -0.27_84]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , _A , atol=1E-4 ) )
102
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class a__ : _a : str = field( default=snake_case__ , metadata={"""help""": """Model type selected in the list: """ + """, """.join(snake_case__ )} ) _a : str = field( default=snake_case__ , metadata={"""help""": """The input data dir. Should contain the .json files for the SQuAD task."""} ) _a : int = field( default=1_2_8 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) _a : int = field( default=1_2_8 , metadata={"""help""": """When splitting up a long document into chunks, how much stride to take between chunks."""} , ) _a : int = field( default=6_4 , metadata={ """help""": ( """The maximum number of tokens for the question. Questions longer than this will """ """be truncated to this length.""" ) } , ) _a : int = field( default=3_0 , metadata={ """help""": ( """The maximum length of an answer that can be generated. This is needed because the start """ """and end predictions are not conditioned on one another.""" ) } , ) _a : bool = field( default=snake_case__ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) _a : bool = field( default=snake_case__ , metadata={"""help""": """If true, the SQuAD examples contain some that do not have an answer."""} ) _a : float = field( default=0.0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) _a : int = field( default=2_0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) _a : int = field( default=0 , metadata={ """help""": ( """language id of input for language-specific xlm models (see""" """ tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)""" ) } , ) _a : int = field(default=1 , metadata={"""help""": """multiple threads for converting example to features"""} ) class a__ ( snake_case__ ): _a : Any = """train""" _a : Union[str, Any] = """dev""" class a__ ( snake_case__ ): _a : SquadDataTrainingArguments _a : List[SquadFeatures] _a : Split _a : bool def __init__( self , _A , _A , _A = None , _A = Split.train , _A = False , _A = None , _A = "pt" , ): """simple docstring""" __lowerCAmelCase = args __lowerCAmelCase = is_language_sensitive __lowerCAmelCase = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(_A , _A ): try: __lowerCAmelCase = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) __lowerCAmelCase = mode # Load data features from cache or dataset file __lowerCAmelCase = "v2" if args.version_2_with_negative else "v1" __lowerCAmelCase = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"""cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}""" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __lowerCAmelCase = cached_features_file + ".lock" with FileLock(_A ): if os.path.exists(_A ) and not args.overwrite_cache: __lowerCAmelCase = time.time() __lowerCAmelCase = torch.load(_A ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. __lowerCAmelCase = self.old_features["features"] __lowerCAmelCase = self.old_features.get("dataset" , _A ) __lowerCAmelCase = self.old_features.get("examples" , _A ) logger.info( f"""Loading features from cached file {cached_features_file} [took %.3f s]""" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"""Deleting cached file {cached_features_file} will allow dataset and examples to be cached in""" " future run" ) else: if mode == Split.dev: __lowerCAmelCase = self.processor.get_dev_examples(args.data_dir ) else: __lowerCAmelCase = self.processor.get_train_examples(args.data_dir ) __lowerCAmelCase , __lowerCAmelCase = squad_convert_examples_to_features( examples=self.examples , tokenizer=_A , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=_A , ) __lowerCAmelCase = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , _A , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"""Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]""" ) def __len__( self ): """simple docstring""" return len(self.features ) def __getitem__( self , _A ): """simple docstring""" __lowerCAmelCase = self.features[i] __lowerCAmelCase = torch.tensor(feature.input_ids , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.attention_mask , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.token_type_ids , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.cls_index , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.p_mask , dtype=torch.float ) __lowerCAmelCase = torch.tensor(feature.is_impossible , dtype=torch.float ) __lowerCAmelCase = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: __lowerCAmelCase = torch.tensor(feature.start_position , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
102
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase : Union[str, Any] =logging.get_logger(__name__) def UpperCamelCase ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Union[str, Any]=False ): A__ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"blocks.{i}.norm1.weight", F"deit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"blocks.{i}.norm1.bias", F"deit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append((F"blocks.{i}.attn.proj.weight", F"deit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"blocks.{i}.attn.proj.bias", F"deit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"blocks.{i}.norm2.weight", F"deit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"blocks.{i}.norm2.bias", F"deit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc1.weight", F"deit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc1.bias", F"deit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc2.weight", F"deit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc2.bias", F"deit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" A__ = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def UpperCamelCase ( _lowerCamelCase : int , _lowerCamelCase : Optional[int] , _lowerCamelCase : Optional[int]=False ): for i in range(config.num_hidden_layers ): if base_model: A__ = "" else: A__ = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) A__ = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict A__ = in_proj_weight[ : config.hidden_size, : ] A__ = in_proj_bias[: config.hidden_size] A__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ = in_proj_weight[ -config.hidden_size :, : ] A__ = in_proj_bias[-config.hidden_size :] def UpperCamelCase ( _lowerCamelCase : Tuple , _lowerCamelCase : Tuple , _lowerCamelCase : List[str] ): A__ = dct.pop(_lowerCamelCase ) A__ = val def UpperCamelCase ( ): A__ = "http://images.cocodataset.org/val2017/000000039769.jpg" A__ = Image.open(requests.get(_lowerCamelCase , stream=_lowerCamelCase ).raw ) return im @torch.no_grad() def UpperCamelCase ( _lowerCamelCase : Any , _lowerCamelCase : Optional[Any] ): A__ = DeiTConfig() # all deit models have fine-tuned heads A__ = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size A__ = 10_00 A__ = "huggingface/label-files" A__ = "imagenet-1k-id2label.json" A__ = json.load(open(hf_hub_download(_lowerCamelCase , _lowerCamelCase , repo_type="dataset" ) , "r" ) ) A__ = {int(_lowerCamelCase ): v for k, v in idalabel.items()} A__ = idalabel A__ = {v: k for k, v in idalabel.items()} A__ = int(deit_name[-6:-4] ) A__ = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): A__ = 1_92 A__ = 7_68 A__ = 12 A__ = 3 elif deit_name[9:].startswith("small" ): A__ = 3_84 A__ = 15_36 A__ = 12 A__ = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): A__ = 10_24 A__ = 40_96 A__ = 24 A__ = 16 # load original model from timm A__ = timm.create_model(_lowerCamelCase , pretrained=_lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A__ = timm_model.state_dict() A__ = create_rename_keys(_lowerCamelCase , _lowerCamelCase ) for src, dest in rename_keys: rename_key(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) read_in_q_k_v(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # load HuggingFace model A__ = DeiTForImageClassificationWithTeacher(_lowerCamelCase ).eval() model.load_state_dict(_lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor A__ = int( (2_56 / 2_24) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 A__ = DeiTImageProcessor(size=_lowerCamelCase , crop_size=config.image_size ) A__ = image_processor(images=prepare_img() , return_tensors="pt" ) A__ = encoding["pixel_values"] A__ = model(_lowerCamelCase ) A__ = timm_model(_lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_lowerCamelCase , outputs.logits , atol=1e-3 ) Path(_lowerCamelCase ).mkdir(exist_ok=_lowerCamelCase ) print(F"Saving model {deit_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_lowerCamelCase ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(_lowerCamelCase ) if __name__ == "__main__": __lowerCAmelCase : Optional[int] =argparse.ArgumentParser() # Required parameters parser.add_argument( "--deit_name", default="vit_deit_base_distilled_patch16_224", type=str, help="Name of the DeiT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) __lowerCAmelCase : Dict =parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
237
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPImageProcessor, CLIPProcessor @require_vision class UpperCAmelCase ( unittest.TestCase ): def UpperCAmelCase_ ( self :Optional[Any] )-> Tuple: A__ = tempfile.mkdtemp() # fmt: off A__ = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on A__ = dict(zip(lowercase_ , range(len(lowercase_ ) ) ) ) A__ = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] A__ = {"unk_token": "<unk>"} A__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) A__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowercase_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowercase_ ) ) A__ = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], "image_std": [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], } A__ = os.path.join(self.tmpdirname , lowercase_ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(lowercase_ , lowercase_ ) def UpperCAmelCase_ ( self :Any , **lowercase_ :Union[str, Any] )-> Tuple: return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCAmelCase_ ( self :Any , **lowercase_ :Tuple )-> Dict: return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCAmelCase_ ( self :Dict , **lowercase_ :Union[str, Any] )-> Any: return CLIPImageProcessor.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCAmelCase_ ( self :List[str] )-> int: shutil.rmtree(self.tmpdirname ) def UpperCAmelCase_ ( self :Optional[int] )-> Optional[int]: A__ = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta )] A__ = [Image.fromarray(np.moveaxis(lowercase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCAmelCase_ ( self :int )-> List[Any]: A__ = self.get_tokenizer() A__ = self.get_rust_tokenizer() A__ = self.get_image_processor() A__ = CLIPProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_slow.save_pretrained(self.tmpdirname ) A__ = CLIPProcessor.from_pretrained(self.tmpdirname , use_fast=lowercase_ ) A__ = CLIPProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_fast.save_pretrained(self.tmpdirname ) A__ = CLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , lowercase_ ) self.assertIsInstance(processor_fast.tokenizer , lowercase_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , lowercase_ ) self.assertIsInstance(processor_fast.image_processor , lowercase_ ) def UpperCAmelCase_ ( self :Optional[Any] )-> Optional[int]: A__ = CLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) A__ = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) A__ = self.get_image_processor(do_normalize=lowercase_ , padding_value=1.0 ) A__ = CLIPProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowercase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowercase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowercase_ ) def UpperCAmelCase_ ( self :List[Any] )-> Tuple: A__ = self.get_image_processor() A__ = self.get_tokenizer() A__ = CLIPProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) A__ = self.prepare_image_inputs() A__ = image_processor(lowercase_ , return_tensors="np" ) A__ = processor(images=lowercase_ , return_tensors="np" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase_ ( self :Optional[int] )-> Dict: A__ = self.get_image_processor() A__ = self.get_tokenizer() A__ = CLIPProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) A__ = "lower newer" A__ = processor(text=lowercase_ ) A__ = tokenizer(lowercase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCAmelCase_ ( self :str )-> Any: A__ = self.get_image_processor() A__ = self.get_tokenizer() A__ = CLIPProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) A__ = "lower newer" A__ = self.prepare_image_inputs() A__ = processor(text=lowercase_ , images=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCAmelCase_ ( self :Tuple )-> Tuple: A__ = self.get_image_processor() A__ = self.get_tokenizer() A__ = CLIPProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) A__ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] A__ = processor.batch_decode(lowercase_ ) A__ = tokenizer.batch_decode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def UpperCAmelCase_ ( self :List[Any] )-> Dict: A__ = self.get_image_processor() A__ = self.get_tokenizer() A__ = CLIPProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) A__ = "lower newer" A__ = self.prepare_image_inputs() A__ = processor(text=lowercase_ , images=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
237
1
def UpperCamelCase_( snake_case__: str , snake_case__: str ) -> int: if len(snake_case__ ) != len(snake_case__ ): raise ValueError('String lengths must match!' ) UpperCAmelCase__ = 0 for chara, chara in zip(snake_case__ , snake_case__ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
371
import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class lowercase ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = BioGptTokenizer __SCREAMING_SNAKE_CASE = False def UpperCamelCase__ (self ) -> str: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] UpperCAmelCase__ = dict(zip(__a , range(len(__a ) ) ) ) UpperCAmelCase__ = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] UpperCAmelCase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(__a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(__a ) ) def UpperCamelCase__ (self , __a ) -> Any: """simple docstring""" UpperCAmelCase__ = 'lower newer' UpperCAmelCase__ = 'lower newer' return input_text, output_text def UpperCamelCase__ (self ) -> Any: """simple docstring""" UpperCAmelCase__ = BioGptTokenizer(self.vocab_file , self.merges_file ) UpperCAmelCase__ = 'lower' UpperCAmelCase__ = ['low', 'er</w>'] UpperCAmelCase__ = tokenizer.tokenize(__a ) self.assertListEqual(__a , __a ) UpperCAmelCase__ = tokens + ['<unk>'] UpperCAmelCase__ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__a ) , __a ) @slow def UpperCamelCase__ (self ) -> int: """simple docstring""" UpperCAmelCase__ = BioGptTokenizer.from_pretrained('microsoft/biogpt' ) UpperCAmelCase__ = tokenizer.encode('sequence builders' , add_special_tokens=__a ) UpperCAmelCase__ = tokenizer.encode('multi-sequence build' , add_special_tokens=__a ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(__a ) UpperCAmelCase__ = tokenizer.build_inputs_with_special_tokens(__a , __a ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
335
0
"""simple docstring""" import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self: List[Any] , snake_case: List[str] , snake_case: Optional[Any]=13 , snake_case: List[str]=7 , snake_case: Dict=True , snake_case: List[str]=True , snake_case: Optional[int]=True , snake_case: Any=True , snake_case: Optional[Any]=99 , snake_case: Tuple=32 , snake_case: Tuple=5 , snake_case: Dict=4 , snake_case: Optional[Any]=37 , snake_case: Union[str, Any]="gelu" , snake_case: Tuple=0.1 , snake_case: List[Any]=0.1 , snake_case: List[str]=512 , snake_case: Optional[int]=16 , snake_case: int=2 , snake_case: List[Any]=0.0_2 , snake_case: Union[str, Any]=4 , ) -> List[str]: snake_case_ :Dict = parent snake_case_ :Any = batch_size snake_case_ :Any = seq_length snake_case_ :List[str] = is_training snake_case_ :Optional[Any] = use_attention_mask snake_case_ :Dict = use_token_type_ids snake_case_ :Union[str, Any] = use_labels snake_case_ :str = vocab_size snake_case_ :int = hidden_size snake_case_ :List[str] = num_hidden_layers snake_case_ :Dict = num_attention_heads snake_case_ :Any = intermediate_size snake_case_ :Tuple = hidden_act snake_case_ :int = hidden_dropout_prob snake_case_ :Optional[Any] = attention_probs_dropout_prob snake_case_ :Any = max_position_embeddings snake_case_ :Union[str, Any] = type_vocab_size snake_case_ :Optional[int] = type_sequence_label_size snake_case_ :Union[str, Any] = initializer_range snake_case_ :Tuple = num_choices def lowerCAmelCase_ ( self: Tuple ) -> str: snake_case_ :Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ :Union[str, Any] = None if self.use_attention_mask: snake_case_ :str = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ :Any = None if self.use_token_type_ids: snake_case_ :List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ :int = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase_ ( self: Optional[int] ) -> int: snake_case_ :str = self.prepare_config_and_inputs() snake_case_, snake_case_, snake_case_, snake_case_ :Optional[int] = config_and_inputs snake_case_ :Union[str, Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def lowerCAmelCase_ ( self: Optional[Any] ) -> Any: snake_case_ :int = self.prepare_config_and_inputs() snake_case_, snake_case_, snake_case_, snake_case_ :Dict = config_and_inputs snake_case_ :Union[str, Any] = True snake_case_ :Optional[int] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ :Tuple = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class lowerCamelCase ( _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' _A : List[str] = True _A : Dict = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase_ ( self: int ) -> List[str]: snake_case_ :Any = FlaxBertModelTester(self ) @slow def lowerCAmelCase_ ( self: List[str] ) -> Dict: # Only check this for base model, not necessary for all model classes. # This will also help speed-up tests. snake_case_ :Dict = FlaxBertModel.from_pretrained("""bert-base-cased""" ) snake_case_ :Dict = model(np.ones((1, 1) ) ) self.assertIsNotNone(snake_case )
66
"""simple docstring""" import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _a = 2 class _lowerCAmelCase : """simple docstring""" def __init__( self : Dict, *, # begin keyword-only arguments UpperCAmelCase__ : str="<s>", UpperCAmelCase__ : Tuple="<pad>", UpperCAmelCase__ : str="</s>", UpperCAmelCase__ : Optional[Any]="<unk>", UpperCAmelCase__ : List[Any]=None, ): __lowercase ,__lowercase ,__lowercase ,__lowercase = bos, unk, pad, eos __lowercase = [] __lowercase = [] __lowercase = {} __lowercase = self.add_symbol(UpperCAmelCase__ ) __lowercase = self.add_symbol(UpperCAmelCase__ ) __lowercase = self.add_symbol(UpperCAmelCase__ ) __lowercase = self.add_symbol(UpperCAmelCase__ ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(UpperCAmelCase__ ) __lowercase = len(self.symbols ) def __eq__( self : List[str], UpperCAmelCase__ : Dict ): return self.indices == other.indices def __getitem__( self : Optional[int], UpperCAmelCase__ : List[str] ): if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self : str ): return len(self.symbols ) def __contains__( self : Any, UpperCAmelCase__ : Optional[Any] ): return sym in self.indices @classmethod def _lowercase ( cls : List[Any], UpperCAmelCase__ : Optional[Any] ): __lowercase = cls() d.add_from_file(UpperCAmelCase__ ) return d def _lowercase ( self : Dict, UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : List[Any]=1, UpperCAmelCase__ : str=False ): if word in self.indices and not overwrite: __lowercase = self.indices[word] __lowercase = self.count[idx] + n return idx else: __lowercase = len(self.symbols ) __lowercase = idx self.symbols.append(UpperCAmelCase__ ) self.count.append(UpperCAmelCase__ ) return idx def _lowercase ( self : Any, UpperCAmelCase__ : str ): return 0 def _lowercase ( self : Tuple, UpperCAmelCase__ : List[Any] ): if isinstance(UpperCAmelCase__, UpperCAmelCase__ ): try: with open(UpperCAmelCase__, "r", encoding="utf-8" ) as fd: self.add_from_file(UpperCAmelCase__ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception("Incorrect encoding detected in {}, please rebuild the dataset".format(UpperCAmelCase__ ) ) return __lowercase = f.readlines() __lowercase = self._load_meta(UpperCAmelCase__ ) for line in lines[indices_start_line:]: try: __lowercase ,__lowercase = line.rstrip().rsplit(" ", 1 ) if field == "#fairseq:overwrite": __lowercase = True __lowercase ,__lowercase = line.rsplit(" ", 1 ) else: __lowercase = False __lowercase = int(UpperCAmelCase__ ) __lowercase = line if word in self and not overwrite: raise RuntimeError( "Duplicate word found when loading Dictionary: '{}'. " "Duplicate words can overwrite earlier ones by adding the " "#fairseq:overwrite flag at the end of the corresponding row " "in the dictionary file. If using the Camembert model, please " "download an updated copy of the model file.".format(UpperCAmelCase__ ) ) self.add_symbol(UpperCAmelCase__, n=UpperCAmelCase__, overwrite=UpperCAmelCase__ ) except ValueError: raise ValueError("Incorrect dictionary format, expected '<token> <cnt> [flags]'" ) def _A ( UpperCamelCase_ : int) -> str: '''simple docstring''' __lowercase = dict((re.sub(r"@@$", "", UpperCamelCase_), v) if k.endswith("@@") else (re.sub(r"$", "</w>", UpperCamelCase_), v) for k, v in d.items()) __lowercase = "<s> <pad> </s> <unk>".split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] __lowercase = d[k] # restore return da def _A ( UpperCamelCase_ : str, UpperCamelCase_ : str) -> List[Any]: '''simple docstring''' if not os.path.exists(UpperCamelCase_): raise ValueError(F"""path {biogpt_checkpoint_path} does not exist!""") os.makedirs(UpperCamelCase_, exist_ok=UpperCamelCase_) print(F"""Writing results to {pytorch_dump_folder_path}""") # handle various types of models __lowercase = os.path.join(UpperCamelCase_, "checkpoint.pt") if not os.path.isfile(UpperCamelCase_): raise ValueError(F"""path to the file {checkpoint_file} does not exist!""") __lowercase = torch.load(UpperCamelCase_, map_location="cpu") __lowercase = chkpt["cfg"]["model"] # dicts __lowercase = os.path.join(UpperCamelCase_, "dict.txt") if not os.path.isfile(UpperCamelCase_): raise ValueError(F"""path to the file {dict_file} does not exist!""") __lowercase = Dictionary.load(UpperCamelCase_) __lowercase = rewrite_dict_keys(src_dict.indices) __lowercase = len(UpperCamelCase_) __lowercase = os.path.join(UpperCamelCase_, VOCAB_FILES_NAMES["vocab_file"]) print(F"""Generating {src_vocab_file} of {src_vocab_size} records""") with open(UpperCamelCase_, "w", encoding="utf-8") as f: f.write(json.dumps(UpperCamelCase_, ensure_ascii=UpperCamelCase_, indent=UpperCamelCase_)) # merges_file (bpecodes) __lowercase = os.path.join(UpperCamelCase_, "bpecodes") if not os.path.isfile(UpperCamelCase_): raise ValueError(F"""path to the file {bpecodes_file} does not exist!""") __lowercase = os.path.join(UpperCamelCase_, VOCAB_FILES_NAMES["merges_file"]) shutil.copyfile(UpperCamelCase_, UpperCamelCase_) # model config __lowercase = os.path.join(UpperCamelCase_, "config.json") __lowercase = { "activation_dropout": args["activation_dropout"], "architectures": ["BioGptForCausalLM"], "attention_probs_dropout_prob": args["attention_dropout"], "bos_token_id": 0, "eos_token_id": 2, "hidden_act": args["activation_fn"], "hidden_dropout_prob": args["dropout"], "hidden_size": args["decoder_embed_dim"], "initializer_range": 0.02, "intermediate_size": args["decoder_ffn_embed_dim"], "layer_norm_eps": 1E-12, "layerdrop": args["decoder_layerdrop"], "max_position_embeddings": args["max_target_positions"], "model_type": "biogpt", "num_attention_heads": args["decoder_attention_heads"], "num_hidden_layers": args["decoder_layers"], "pad_token_id": 1, "scale_embedding": not args["no_scale_embedding"], "tie_word_embeddings": args["share_decoder_input_output_embed"], "vocab_size": src_vocab_size, } # good hparam defaults to start with print(F"""Generating {biogpt_model_config_file}""") with open(UpperCamelCase_, "w", encoding="utf-8") as f: f.write(json.dumps(UpperCamelCase_, ensure_ascii=UpperCamelCase_, indent=UpperCamelCase_)) # tokenizer config __lowercase = os.path.join(UpperCamelCase_, UpperCamelCase_) __lowercase = { "bos_token": "<s>", "eos_token": "</s>", "model_max_length": 1024, "pad_token": "<pad>", "special_tokens_map_file": None, "tokenizer_class": "BioGptTokenizer", "unk_token": "<unk>", } print(F"""Generating {biogpt_tokenizer_config_file}""") with open(UpperCamelCase_, "w", encoding="utf-8") as f: f.write(json.dumps(UpperCamelCase_, ensure_ascii=UpperCamelCase_, indent=UpperCamelCase_)) # model __lowercase = chkpt["model"] # remove unneeded keys __lowercase = [ "decoder.version", ] for k in ignore_keys: model_state_dict.pop(UpperCamelCase_, UpperCamelCase_) __lowercase = list(model_state_dict.keys()) for layer_name in layer_names: if layer_name.endswith("output_projection.weight"): __lowercase = model_state_dict.pop(UpperCamelCase_) else: __lowercase = model_state_dict.pop(UpperCamelCase_) __lowercase = BioGptConfig.from_pretrained(UpperCamelCase_) __lowercase = BioGptForCausalLM(UpperCamelCase_) # check that it loads ok model_new.load_state_dict(UpperCamelCase_) # save __lowercase = os.path.join(UpperCamelCase_, UpperCamelCase_) print(F"""Generating {pytorch_weights_dump_path}""") torch.save(UpperCamelCase_, UpperCamelCase_) print("Conversion is done!") if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--biogpt_checkpoint_path', default=None, type=str, required=True, help=( 'Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,' ' bpecodes, etc.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) _a = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
17
0
import qiskit def lowerCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase )-> qiskit.result.counts.Counts: '''simple docstring''' UpperCAmelCase : Any =qiskit.Aer.get_backend('''aer_simulator''' ) # Create a Quantum Circuit acting on the q register UpperCAmelCase : Any =qiskit.QuantumCircuit(lowercase_ , lowercase_ ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator UpperCAmelCase : int =qiskit.execute(lowercase_ , lowercase_ , shots=10_00 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(lowercase_ ) if __name__ == "__main__": print(f'Total count for various states are: {single_qubit_measure(1, 1)}')
362
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class __snake_case ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): __lowerCamelCase : Dict = StableUnCLIPPipeline __lowerCamelCase : int = TEXT_TO_IMAGE_PARAMS __lowerCamelCase : int = TEXT_TO_IMAGE_BATCH_PARAMS __lowerCamelCase : Dict = TEXT_TO_IMAGE_IMAGE_PARAMS __lowerCamelCase : Dict = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false __lowerCamelCase : Optional[Any] = False def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : int =32 UpperCAmelCase : Union[str, Any] =embedder_hidden_size # prior components torch.manual_seed(0 ) UpperCAmelCase : Optional[Any] =CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) UpperCAmelCase : int =CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=snake_case__ , projection_dim=snake_case__ , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) UpperCAmelCase : Dict =PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=snake_case__ , num_layers=1 , ) torch.manual_seed(0 ) UpperCAmelCase : Tuple =DDPMScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1000 , clip_sample=snake_case__ , clip_sample_range=5.0 , beta_schedule='''squaredcos_cap_v2''' , ) # regular denoising components torch.manual_seed(0 ) UpperCAmelCase : Optional[int] =StableUnCLIPImageNormalizer(embedding_dim=snake_case__ ) UpperCAmelCase : Any =DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' ) torch.manual_seed(0 ) UpperCAmelCase : List[str] =CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) UpperCAmelCase : List[str] =CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=snake_case__ , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) UpperCAmelCase : Optional[int] =UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=snake_case__ , layers_per_block=1 , upcast_attention=snake_case__ , use_linear_projection=snake_case__ , ) torch.manual_seed(0 ) UpperCAmelCase : List[Any] =DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_0085 , beta_end=0.012 , prediction_type='''v_prediction''' , set_alpha_to_one=snake_case__ , steps_offset=1 , ) torch.manual_seed(0 ) UpperCAmelCase : Dict =AutoencoderKL() UpperCAmelCase : Tuple ={ # prior components '''prior_tokenizer''': prior_tokenizer, '''prior_text_encoder''': prior_text_encoder, '''prior''': prior, '''prior_scheduler''': prior_scheduler, # image noising components '''image_normalizer''': image_normalizer, '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder, '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, } return components def UpperCAmelCase__ ( self , snake_case__ , snake_case__=0 ) -> List[Any]: '''simple docstring''' if str(snake_case__ ).startswith('''mps''' ): UpperCAmelCase : Union[str, Any] =torch.manual_seed(snake_case__ ) else: UpperCAmelCase : Any =torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase : str ={ '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''prior_num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Tuple =torch_device == '''cpu''' self._test_attention_slicing_forward_pass(test_max_difference=snake_case__ ) def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' UpperCAmelCase : List[Any] =torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=snake_case__ ) @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : Union[str, Any] =load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy''' ) UpperCAmelCase : Optional[int] =StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() UpperCAmelCase : int =torch.Generator(device='''cpu''' ).manual_seed(0 ) UpperCAmelCase : int =pipe('''anime turle''' , generator=snake_case__ , output_type='''np''' ) UpperCAmelCase : str =output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(snake_case__ , snake_case__ ) def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCAmelCase : List[str] =StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) UpperCAmelCase : str =pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() UpperCAmelCase : Any =pipe( '''anime turtle''' , prior_num_inference_steps=2 , num_inference_steps=2 , output_type='''np''' , ) UpperCAmelCase : Tuple =torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
78
0
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def __lowerCAmelCase ( UpperCamelCase__ ) -> Tuple: __lowerCamelCase = filter(lambda UpperCamelCase__ : p.requires_grad , model.parameters() ) __lowerCamelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params __UpperCAmelCase =logging.getLogger(__name__) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> Optional[int]: if metric == "rouge2": __lowerCamelCase = '''{val_avg_rouge2:.4f}-{step_count}''' elif metric == "bleu": __lowerCamelCase = '''{val_avg_bleu:.4f}-{step_count}''' elif metric == "em": __lowerCamelCase = '''{val_avg_em:.4f}-{step_count}''' elif metric == "loss": __lowerCamelCase = '''{val_avg_loss:.4f}-{step_count}''' else: raise NotImplementedError( f"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ''' function.''' ) __lowerCamelCase = ModelCheckpoint( dirpath=UpperCamelCase__ , filename=UpperCamelCase__ , monitor=f"""val_{metric}""" , mode='''max''' , save_top_k=1 , every_n_epochs=1 , ) return checkpoint_callback def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> Optional[int]: return EarlyStopping( monitor=f"""val_{metric}""" , mode='''min''' if '''loss''' in metric else '''max''' , patience=UpperCamelCase__ , verbose=UpperCamelCase__ , ) class a__ ( pl.Callback ): def SCREAMING_SNAKE_CASE__ ( self : Any , a : str , a : List[Any] ): """simple docstring""" __lowerCamelCase = {f"""lr_group_{i}""": param['''lr'''] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(a ) @rank_zero_only def SCREAMING_SNAKE_CASE__ ( self : Dict , a : pl.Trainer , a : pl.LightningModule , a : str , a : List[Any]=True ): """simple docstring""" logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCamelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['''log''', '''progress_bar''', '''preds''']} ) # Log results __lowerCamelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCamelCase = od / '''test_results.txt''' __lowerCamelCase = od / '''test_generations.txt''' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCamelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCamelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=a ) generations_file.parent.mkdir(exist_ok=a ) with open(a , '''a+''' ) as writer: for key in sorted(a ): if key in ["log", "progress_bar", "preds"]: continue __lowerCamelCase = metrics[key] if isinstance(a , torch.Tensor ): __lowerCamelCase = val.item() __lowerCamelCase = f"""{key}: {val:.6f}\n""" writer.write(a ) if not save_generations: return if "preds" in metrics: __lowerCamelCase = '''\n'''.join(metrics['''preds'''] ) generations_file.open('''w+''' ).write(a ) @rank_zero_only def SCREAMING_SNAKE_CASE__ ( self : str , a : Dict , a : List[str] ): """simple docstring""" try: __lowerCamelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCamelCase = pl_module.model.num_parameters() __lowerCamelCase = count_trainable_parameters(a ) # mp stands for million parameters trainer.logger.log_metrics({'''n_params''': npars, '''mp''': npars / 1e6, '''grad_mp''': n_trainable_pars / 1e6} ) @rank_zero_only def SCREAMING_SNAKE_CASE__ ( self : int , a : pl.Trainer , a : pl.LightningModule ): """simple docstring""" save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(a , a , '''test''' ) @rank_zero_only def SCREAMING_SNAKE_CASE__ ( self : Tuple , a : pl.Trainer , a : Optional[int] ): """simple docstring""" save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
67
from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Dict = logging.get_logger(__name__) __A : List[Any] = {'''ctrl''': '''https://huggingface.co/ctrl/resolve/main/config.json'''} class __A ( lowerCAmelCase ): lowerCAmelCase_ : str = "ctrl" lowerCAmelCase_ : Optional[Any] = ["past_key_values"] lowerCAmelCase_ : Dict = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Any , UpperCAmelCase_ : int=246534 , UpperCAmelCase_ : Optional[Any]=256 , UpperCAmelCase_ : Any=1280 , UpperCAmelCase_ : int=8192 , UpperCAmelCase_ : int=48 , UpperCAmelCase_ : Optional[Any]=16 , UpperCAmelCase_ : Dict=0.1 , UpperCAmelCase_ : Any=0.1 , UpperCAmelCase_ : List[str]=1E-6 , UpperCAmelCase_ : str=0.02 , UpperCAmelCase_ : Optional[Any]=True , **UpperCAmelCase_ : int , ): lowerCAmelCase : int = vocab_size lowerCAmelCase : int = n_positions lowerCAmelCase : Optional[Any] = n_embd lowerCAmelCase : Optional[Any] = n_layer lowerCAmelCase : List[str] = n_head lowerCAmelCase : Union[str, Any] = dff lowerCAmelCase : Dict = resid_pdrop lowerCAmelCase : List[Any] = embd_pdrop lowerCAmelCase : List[Any] = layer_norm_epsilon lowerCAmelCase : Dict = initializer_range lowerCAmelCase : Union[str, Any] = use_cache super().__init__(**UpperCAmelCase_ )
138
0
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo _UpperCAmelCase = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey\n and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin\n Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto\n Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and\n Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes\n and Jeffrey Dean},\n year={2016},\n eprint={1609.08144},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n' _UpperCAmelCase = '\\nThe BLEU score has some undesirable properties when used for single\nsentences, as it was designed to be a corpus measure. We therefore\nuse a slightly different score for our RL experiments which we call\nthe \'GLEU score\'. For the GLEU score, we record all sub-sequences of\n1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then\ncompute a recall, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the target (ground truth) sequence,\nand a precision, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the generated output sequence. Then\nGLEU score is simply the minimum of recall and precision. This GLEU\nscore\'s range is always between 0 (no matches) and 1 (all match) and\nit is symmetrical when switching output and target. According to\nour experiments, GLEU score correlates quite well with the BLEU\nmetric on a corpus level but does not have its drawbacks for our per\nsentence reward objective.\n' _UpperCAmelCase = '\\nComputes corpus-level Google BLEU (GLEU) score of translated segments against one or more references.\nInstead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching\ntokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values.\n\nArgs:\n predictions (list of str): list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references (list of list of str): list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n min_len (int): The minimum order of n-gram this function should extract. Defaults to 1.\n max_len (int): The maximum order of n-gram this function should extract. Defaults to 4.\n\nReturns:\n \'google_bleu\': google_bleu score\n\nExamples:\n Example 1:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results["google_bleu"], 2))\n 0.44\n\n Example 2:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results["google_bleu"], 2))\n 0.61\n\n Example 3:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2)\n >>> print(round(results["google_bleu"], 2))\n 0.53\n\n Example 4:\n >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\',\n ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\']\n >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\',\n ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\',\n ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\']\n >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\',\n ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\',\n ... \'heed\', \'the\', \'cat\', \'commands\']\n >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\',\n ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\',\n ... \'of\', \'the\', \'cat\']\n\n >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\',\n ... \'interested\', \'in\', \'world\', \'history\']\n >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\',\n ... \'because\', \'he\', \'read\', \'the\', \'book\']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric("google_bleu")\n >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6)\n >>> print(round(results["google_bleu"], 2))\n 0.4\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class snake_case_ ( datasets.Metric ): def UpperCAmelCase__ ( self : Union[str, Any] )->MetricInfo: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" , id="""token""" ) , id="""sequence""" ), """references""": datasets.Sequence( datasets.Sequence(datasets.Value("""string""" , id="""token""" ) , id="""sequence""" ) , id="""references""" ), } ) , ) def UpperCAmelCase__ ( self : Tuple , _snake_case : List[List[List[str]]] , _snake_case : List[List[str]] , _snake_case : int = 1 , _snake_case : int = 4 , )->Dict[str, float]: '''simple docstring''' return { "google_bleu": gleu_score.corpus_gleu( list_of_references=_snake_case , hypotheses=_snake_case , min_len=_snake_case , max_len=_snake_case ) }
370
import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax from transformers.models.vit.modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel class snake_case_ ( unittest.TestCase ): def __init__( self : List[Any] , _snake_case : List[Any] , _snake_case : str=13 , _snake_case : int=30 , _snake_case : str=2 , _snake_case : int=3 , _snake_case : Optional[Any]=True , _snake_case : str=True , _snake_case : Optional[int]=32 , _snake_case : Dict=5 , _snake_case : Optional[int]=4 , _snake_case : List[Any]=37 , _snake_case : Union[str, Any]="gelu" , _snake_case : str=0.1 , _snake_case : str=0.1 , _snake_case : str=10 , _snake_case : Any=0.02 , )->Tuple: '''simple docstring''' __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : Any = batch_size __lowerCAmelCase : int = image_size __lowerCAmelCase : int = patch_size __lowerCAmelCase : List[Any] = num_channels __lowerCAmelCase : str = is_training __lowerCAmelCase : str = use_labels __lowerCAmelCase : List[str] = hidden_size __lowerCAmelCase : Dict = num_hidden_layers __lowerCAmelCase : List[str] = num_attention_heads __lowerCAmelCase : Any = intermediate_size __lowerCAmelCase : List[str] = hidden_act __lowerCAmelCase : List[str] = hidden_dropout_prob __lowerCAmelCase : Any = attention_probs_dropout_prob __lowerCAmelCase : Tuple = type_sequence_label_size __lowerCAmelCase : Union[str, Any] = initializer_range # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : Tuple = (image_size // patch_size) ** 2 __lowerCAmelCase : Optional[Any] = num_patches + 1 def UpperCAmelCase__ ( self : Any )->Any: '''simple docstring''' __lowerCAmelCase : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCAmelCase : Tuple = ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_snake_case , initializer_range=self.initializer_range , ) return config, pixel_values def UpperCAmelCase__ ( self : List[Any] , _snake_case : List[str] , _snake_case : List[Any] )->Any: '''simple docstring''' __lowerCAmelCase : str = FlaxViTModel(config=_snake_case ) __lowerCAmelCase : int = model(_snake_case ) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : Dict = (self.image_size, self.image_size) __lowerCAmelCase : Any = (self.patch_size, self.patch_size) __lowerCAmelCase : Any = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, num_patches + 1, self.hidden_size) ) def UpperCAmelCase__ ( self : Dict , _snake_case : str , _snake_case : List[Any] )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Tuple = self.type_sequence_label_size __lowerCAmelCase : Tuple = FlaxViTForImageClassification(config=_snake_case ) __lowerCAmelCase : List[str] = model(_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCAmelCase : str = 1 __lowerCAmelCase : Any = FlaxViTForImageClassification(_snake_case ) __lowerCAmelCase : int = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCAmelCase : Dict = model(_snake_case ) def UpperCAmelCase__ ( self : str )->Any: '''simple docstring''' __lowerCAmelCase : Any = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) : Union[str, Any] = config_and_inputs __lowerCAmelCase : str = {"""pixel_values""": pixel_values} return config, inputs_dict @require_flax class snake_case_ ( __lowercase ,unittest.TestCase ): A_ = (FlaxViTModel, FlaxViTForImageClassification) if is_flax_available() else () def UpperCAmelCase__ ( self : str )->None: '''simple docstring''' __lowerCAmelCase : List[str] = FlaxViTModelTester(self ) __lowerCAmelCase : Optional[Any] = ConfigTester(self , config_class=_snake_case , has_text_modality=_snake_case , hidden_size=37 ) def UpperCAmelCase__ ( self : Any )->Dict: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : Any )->Any: '''simple docstring''' __lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def UpperCAmelCase__ ( self : Union[str, Any] )->str: '''simple docstring''' __lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_snake_case ) def UpperCAmelCase__ ( self : int )->int: '''simple docstring''' __lowerCAmelCase , __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase : int = model_class(_snake_case ) __lowerCAmelCase : Optional[int] = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase : Tuple = [*signature.parameters.keys()] __lowerCAmelCase : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _snake_case ) def UpperCAmelCase__ ( self : str )->str: '''simple docstring''' __lowerCAmelCase , __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __lowerCAmelCase : Optional[int] = self._prepare_for_class(_snake_case , _snake_case ) __lowerCAmelCase : List[str] = model_class(_snake_case ) @jax.jit def model_jitted(_snake_case : Dict , **_snake_case : Union[str, Any] ): return model(pixel_values=_snake_case , **_snake_case ) with self.subTest("""JIT Enabled""" ): __lowerCAmelCase : List[Any] = model_jitted(**_snake_case ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): __lowerCAmelCase : str = model_jitted(**_snake_case ).to_tuple() self.assertEqual(len(_snake_case ) , len(_snake_case ) ) for jitted_output, output in zip(_snake_case , _snake_case ): self.assertEqual(jitted_output.shape , output.shape ) @slow def UpperCAmelCase__ ( self : Dict )->str: '''simple docstring''' for model_class_name in self.all_model_classes: __lowerCAmelCase : List[str] = model_class_name.from_pretrained("""google/vit-base-patch16-224""" ) __lowerCAmelCase : List[str] = model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(_snake_case )
232
0