code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets __magic_name__ : Tuple = '''\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav K...
280
def lowercase__ ( _UpperCamelCase) -> list: """simple docstring""" if bit_count < 0: raise ValueError('The given input must be positive') # get the generated string sequence UpperCamelCase = gray_code_sequence_string(_UpperCamelCase) ...
280
1
"""simple docstring""" import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class snake_case__ ( snake_case_ ): _snake_case : Tuple = """Speech2TextFeatureExtractor""" _snake_case : Dict = """Speech2TextTokenize...
67
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigW...
67
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) SCREAMING_SNAKE_CASE__ : List[str] = { "configuration_trocr": ["TROCR_PRETRAINED_CONFIG_ARCHIVE_...
85
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def SCREAMING_SNAKE_CASE ( __UpperCAmelCase ) -> str: if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise TypeError("Undefined for non-integers...
159
0
from __future__ import annotations def __lowercase( UpperCAmelCase__ = 4 ): """simple docstring""" lowerCamelCase = abs(UpperCAmelCase__ ) or 4 return [[1 + x + y * row_size for x in range(UpperCAmelCase__ )] for y in range(UpperCAmelCase__ )] def __lowercase( UpperCAmelCase...
707
import gc import unittest from diffusers import FlaxStableDiffusionInpaintPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils im...
484
0
'''simple docstring''' from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup _UpperCamelCase : Union[str, Any] = """https://www.indeed.co.in/jobs?q=mobile+app+development&l=""" def __UpperCAmelCase ( A : str = "mu...
541
"""simple docstring""" import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class low...
438
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """EleutherAI/gpt-neox-20b""": """https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json""", # See all GPTNeoX models ...
69
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' ...
69
1
'''simple docstring''' import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class _lowerCAmelCase : def __init__(self , lowercase ): if isinstance(A_ , A_ ): # Do...
667
'''simple docstring''' def lowercase__( _UpperCamelCase : str )-> str: """simple docstring""" return " ".join( "".join(word[::-1] ) if len(_UpperCamelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.te...
138
0
"""simple docstring""" from __future__ import annotations from typing import Any def __lowerCAmelCase ( __lowerCAmelCase : list[Any] ) -> None: create_state_space_tree(__lowerCAmelCase , [] , 0 ) def __lowerCAmelCase ( __lowerCAmelCase : list[Any] ...
711
"""simple docstring""" import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.pat...
239
0
import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : int , UpperCAmelCase_ ...
443
def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : int = 1_0_0_0 ) -> int: SCREAMING_SNAKE_CASE_ : int =2**power SCREAMING_SNAKE_CASE_ : Tuple =0 while n: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] =r + n...
443
1
"""simple docstring""" import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTest...
480
"""simple docstring""" import json import logging import os import socket import git import numpy as np import torch logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - PID: %(process)d - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO, ) _lowerCAmelCase = ...
480
1
def UpperCamelCase_ ( ) -> List[Any]: a__ : Optional[int] = [] a__ : Dict = 1 while len(__a ) < 1e6: constant.append(str(__a ) ) i += 1 a__ : Dict = "".join(__a ) return ( int(constant[0] ) ...
37
"""simple docstring""" from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def _lowerCamelCase ( UpperCAmelC...
232
0
from importlib import import_module from .logging import get_logger lowerCamelCase__ : Optional[Any] = get_logger(__name__) class _snake_case : def __init__( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None): '''simple docstring''' ...
717
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record lowerCamelCase__ : Dict = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author...
495
0
'''simple docstring''' def _lowercase ( lowerCamelCase__ ) -> bool: """simple docstring""" return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
168
'''simple docstring''' import operator def _lowercase ( lowerCamelCase__ , lowerCamelCase__ = False , lowerCamelCase__ = None ) -> list: """simple docstring""" __UpperCAmelCase : Tuple = operator.lt if reverse else operator.gt...
168
1
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMix...
707
'''simple docstring''' import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence ...
419
0
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _a ( UpperCAmelCase__ , UpperCAmelCase__ ): """simple docstring""...
23
'''simple docstring''' import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optim...
588
0
'''simple docstring''' import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __UpperCAmelCase = "src/transformers" #...
702
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common...
692
0
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from ....
64
import inspect import unittest class UpperCamelCase( unittest.TestCase ): def SCREAMING_SNAKE_CASE_ ( self : Dict ) -> List[Any]: '''simple docstring''' try: import diffusers # noqa: F401 except...
371
0
"""simple docstring""" import torch def UpperCamelCase ( ): if torch.cuda.is_available(): __a = torch.cuda.device_count() else: __a = 0 print(f"""Successfully ran on {num_gpus} GPUs""" ) if __name__ == "__main__": main()
708
"""simple docstring""" __A = 6_55_21 def UpperCamelCase ( _lowerCAmelCase : str ): __a = 1 __a = 0 for plain_chr in plain_text: __a = (a + ord(_lowerCAmelCase )) % MOD_ADLER __a = (b + a) % MOD_ADLER return (b << 16) | a
173
0
'''simple docstring''' def _a ( _lowerCamelCase , _lowerCamelCase ) -> int: """simple docstring""" while a != 0: __snake_case , __snake_case : str = b % a, a return b def _a ( _lo...
26
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) a : List[str] = { """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MA...
613
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase: Optional[int] = logging.get_logger(__name__) UpperCAmelCase: Union[str, A...
714
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCAmelCase: List[Any] = logging.get_logger(__name__) class UpperCamelCase ( snake_case ): """sim...
600
0
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): '''simple docstring...
62
import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def lowerCamelCase__ ( lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = os.path.join(args.tf_model_dir , "parameters...
62
1
"""simple docstring""" import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'MCTCTFeatureExtractor' lowercase__ = 'AutoTokenizer' def __init...
720
"""simple docstring""" import copy import re class _UpperCAmelCase: lowercase__ = 'hp' lowercase__ = {} lowercase__ = None @classmethod def UpperCAmelCase ( cls , __a , __a) -> Dict: '''simple...
78
0
from ...configuration_utils import PretrainedConfig from ...utils import logging a__ = logging.get_logger(__name__) a__ = { """microsoft/trocr-base-handwritten""": ( """https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json""" ), # See...
477
"""simple docstring""" class UpperCAmelCase_ : """simple docstring""" def __init__( self : Optional[Any] , a_ : int )-> str: """simple docstring""" UpperCAmelCase_ : Any = n UpperCAmelCase_ : str = [None] * self.n ...
470
0
import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from tr...
717
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch lowerCamelCase__ = lo...
82
0
'''simple docstring''' import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin c...
48
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class A : '''simple docstring''' A__ = 42 A__ = None A__ = N...
15
0
import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, ...
715
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def lowercase (_lowerCAmelCase ): if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("""Undefined for non-integers""" ) elif precision < 1: ra...
573
0
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : str = { """configuration_informer""": [ """INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", ...
543
"""simple docstring""" import random def a__ ( snake_case__ , snake_case__ , snake_case__ = False ) -> dict: lowerCamelCase = {i: [] for i in range(snake_case__ )} # if probability is greater or equal than 1, then generate a complete graph if prob...
543
1
"""simple docstring""" import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import...
704
"""simple docstring""" import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import ...
533
0
import qiskit def UpperCamelCase ( snake_case__ : int , snake_case__ : int ) -> qiskit.result.counts.Counts: UpperCamelCase : Optional[int] = qiskit.Aer.get_backend('aer_simulator' ) # Create a Quantum Circuit acting on the q register Upper...
40
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def UpperCamelCase ( snake_case__ :...
40
1
lowerCamelCase__ = range(2, 20 + 1) lowerCamelCase__ = [10**k for k in range(ks[-1] + 1)] lowerCamelCase__ = {} def A(__a: Tuple , __a: Optional[Any] , __a: List[str] , __a: Optional[int] ): lowerCAmelCase_ = sum(a_i[j] for j in range(_SCREAMING_SNAKE_C...
701
import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class __magic_name__ (__lowercase ): lowerCamelCase__ = '''''' lowerCamelCase__ = ( None # protocol passed in prefix to the ur...
226
0
from math import isqrt def __lowerCAmelCase ( A ): UpperCAmelCase_ = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , lowercase_ , lowercase_ ): UpperCAmelCase_ = False ...
162
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging A_ : List[Any] = logging.get_logg...
456
0
from __future__ import annotations def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase, ): if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError("You cannot supply more or less than 2 values" ) elif electron_conc < 0: raise ValueError("...
703
from abc import ABC, abstractmethod from argparse import ArgumentParser class __lowerCAmelCase ( lowerCAmelCase): @staticmethod @abstractmethod def SCREAMING_SNAKE_CASE ( _lowerCAmelCase: ArgumentParser ): raise NotImplementedError() ...
453
0
'''simple docstring''' import math from numpy import inf from scipy.integrate import quad def _UpperCamelCase (_lowerCamelCase : float )-> float: '''simple docstring''' if num <= 0: raise ValueError('''math domain error''' ) return quad(_lowerCame...
24
import argparse import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator...
136
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase = { """configuration_swinv2""": ["""SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Swinv2Config"""], } try: if not is_torch_available(): raise OptionalDependen...
719
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : list , UpperCAmelCase_ : int ) -> Tuple: # Checks if the entire collection has been sorted if len(UpperCAmelCase_ ) <= 1 or n <= 1: return insert_next(UpperCA...
431
0
'''simple docstring''' import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import...
566
'''simple docstring''' import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_availabl...
566
1
import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class lowerCAmelCase__ ( ctypes.Structure ): # _fields is a specific attr expected by ctypes __A : Optional[Any] = [('size', ...
182
import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class lowerCAmelCase__ ( unittest.TestCase , UpperCamelCase ): def _lowercase ( self : List[Any]): A__ : Optional[Any] = load_tool("t...
182
1
def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : int ) -> int: if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): raise ValueError('''multiplicative_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''m...
443
import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : int = 8 ) -> str: SCREAMING_SNAKE_CASE_ : str =ascii_letters + digits + punctuation retu...
443
1
from __future__ import annotations def a_ ( __snake_case , __snake_case ) -> list[int]: '''simple docstring''' UpperCamelCase_ = 0 UpperCamelCase_ = len(__snake_case ) - 1 while i < j: if nums[i] + nums[j] == target: ...
559
def a_ ( __snake_case ) -> list[int]: '''simple docstring''' UpperCamelCase_ = [0 for i in range(len(__snake_case ) )] # initialize interval's left pointer and right pointer UpperCamelCase_ , UpperCamelCase_ = 0, 0 for i in r...
559
1
'''simple docstring''' # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING, Dict, Optional import numpy as np import pyarrow as pa from .. import config from ..utils.logging import get_logger from ..utils.py_utils import map_nested from .formatting import Tenso...
126
'''simple docstring''' import re from filelock import FileLock try: import nltk _A: Optional[int] = True except (ImportError, ModuleNotFoundError): _A: Dict = False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punkt"...
126
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : Union[str, Any] = logging.get_logger(__name_...
708
"""simple docstring""" import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTest...
158
0
'''simple docstring''' import os def __snake_case (): """simple docstring""" lowerCamelCase_ : Dict = os.path.dirname(os.path.realpath(__UpperCAmelCase ) ) lowerCamelCase_ : Optional[Any] = os.path.join(__UpperCAmelCase , '''triangle.txt''' ) ...
501
'''simple docstring''' import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient __lowerCamelCase : str = WebClient(token=os.environ["""CI_SLACK_BOT_TOKEN"""]) def ...
501
1
'''simple docstring''' import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) UpperCamelCase_ = logging.g...
88
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase_ = {"""configuration_plbart""": ["""PLBART_PRETRAINED_CONFIG_ARCHI...
88
1
'''simple docstring''' def _lowercase ( UpperCamelCase__ : Optional[Any] ): __A : Dict = 1 __A : Dict = 2 while i * i <= n: __A : Any = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors...
365
'''simple docstring''' from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass cla...
365
1
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerFor...
287
'''simple docstring''' import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ...
287
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ : str = { "configuration_layoutlmv3": [ "LAYOUT...
85
A = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' A = [{'type': 'code', 'content': INSTALL_CONT...
544
0
import re from filelock import FileLock try: import nltk SCREAMING_SNAKE_CASE__:Any = True except (ImportError, ModuleNotFoundError): SCREAMING_SNAKE_CASE__:Optional[Any] = False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punkt""", quiet=True) def _lowerCamelCa...
708
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class snake_case__ ( snake_case_ ): def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: ...
67
0
"""simple docstring""" from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class _UpperCamelCase ( __snake_case): def A (self , lowerCamelCase__ ): ...
574
"""simple docstring""" from __future__ import annotations import time lowerCamelCase__ = list[tuple[int, int]] lowerCamelCase__ = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0,...
574
1
import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...t...
720
"""simple docstring""" def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : List[str] = len(_UpperCamelCase ) __lowerCAmelCase : Tuple = [[0] * n for i in range(_UpperCamelCase )] for i in range(_UpperCamelCase ): __lower...
549
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: ...
132
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available SCREAMING_SNAKE_CASE = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyN...
199
0
'''simple docstring''' import csv import tweepy # Twitter API credentials snake_case_ = """""" snake_case_ = """""" snake_case_ = """""" snake_case_ = """""" def __lowercase (_SCREAMING_SNAKE_CASE :str ): SCREAMING_SNAKE_CASE : str = tweepy.OAut...
706
'''simple docstring''' from __future__ import annotations def __lowercase (_SCREAMING_SNAKE_CASE :int | str ): SCREAMING_SNAKE_CASE : int = str(_SCREAMING_SNAKE_CASE ) return n == n[::-1] def __lowercase (_SCREAMING_SNAKE_CASE :int = 1_00_00_0...
355
0
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ....
136
"""simple docstring""" from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE__ ( metaclass=_SCREAMING_SNAKE_CASE ): snake_case = ["speech"] def __init__( self : Optional[int] , *SCREAMING_SNAKE_CASE_ : Tuple , **SCREAMING_SNAKE_...
129
0
'''simple docstring''' from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import lo...
514
'''simple docstring''' import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_avai...
514
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase :List[str] = logging.get_logger(__name__) lowerCamelCase :List[str] = { 'facebook/wav2vec2-base-960h': 'https://huggingface.co/facebook/wav2vec2-base...
487
import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature ...
487
1
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...tes...
644
'''simple docstring''' def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : int = 10, SCREAMING_SNAKE_CASE__ : int = 22 ) -> int: UpperCAmelCase_ : Optional[int] = range(1, SCREAMING_SNAKE_CASE__ ) UpperCAmelCase_ : List[Any] = ra...
644
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_feature...
357
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless require...
306
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE : str = { "configuration_pegasus_x": ["PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP", "PegasusXConfig"], } try: if...
700
'''simple docstring''' from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuan...
238
0
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def A_ ( _lowerCAmelCase : dict , _lowerCAmelCase : str , _lowerCAmelCase : set , _lowerCAmelCase : set , _lowerCAmelCase : dict , ...
44
'''simple docstring''' import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingSt...
44
1
import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def UpperCamelCase__ ( ) -> None: print('Making key files...' ) make_key_files('rsa' , 1024 ) print('Key files ge...
699
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmen...
699
1
'''simple docstring''' from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def snake_case_ ( __snake_case : List[Any]) -> Optional[int]: return ConvertCommand( args.model_type , args.tf_checkpoint , args....
274
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE = {"""configuration_deit""": ["""DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """D...
199
0
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : Any = logging.get_logger(__name__) class lowercase_ ( a__ ): __UpperCAmelCase = 'encoder-decoder' __UpperCAmelCase = True ...
223
'''simple docstring''' # Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config ...
223
1
import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def __UpperCamelCase ( lowercase__ : Optional[Any] )...
600
import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def A__( __lowerCAmelCase ): _snake_case : Dict = [ 'decoder.version', 'decoder.output_projection.weight', '_float...
304
0
'''simple docstring''' import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from tra...
714
'''simple docstring''' from pathlib import Path import fire from tqdm import tqdm def lowerCAmelCase_ ( a : int="ro" , a : Tuple="en" , a : Union[str, Any]="wmt16" , a : List[Any]=None ): try: import datasets except (Modul...
126
0
'''simple docstring''' import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common imp...
309
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices snake_case = logging.get_logger(__name__) snake_case = { '''google/bit-50''': '''https://huggin...
309
1
'''simple docstring''' # XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path _lowerCAmelCase = Path(__file__).resolve().parents[3] / "src" sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa i...
245
'''simple docstring''' from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import flo...
245
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "ut/deta": "https://huggingface.co/ut/deta/resolve/main/config.json", } class ...
619
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( _a ): snake_case : Optional[Any] = """encoder-decoder""" snake_case : Optio...
619
1
from bisect import bisect from itertools import accumulate def _lowerCAmelCase ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ): '''simple docstring''' A_ : int = sorted(zip(_lowerCAmelCase ,_lowerCAmelCase ) ,key=lambda _lowerCAmelCase : x[0]...
708
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def _lowerCAmelCase ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,): '''simple docstring''' A_ , A_ : int = coefficient_matrix.shape...
481
0
'''simple docstring''' from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) lowercase__ : str...
390
'''simple docstring''' # Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def __lowerCamelCase ( _UpperCamelCase : Tuple , _UpperCamelCase : List[Any] , _UpperCamelCase : Dict , _UpperCamelCase : Tuple ): '''simple docstring''' UpperCAmelC...
390
1
'''simple docstring''' def __UpperCAmelCase ( ) -> list[list[int]]: return [list(range(1_000 - i , -1_000 - i , -1)) for i in range(1_000)] lowerCAmelCase__ = generate_large_matrix() lowerCAmelCase__ = ( [[4, 3, 2, ...
6
'''simple docstring''' import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_a...
6
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffu...
692
'''simple docstring''' lowerCAmelCase_ : Optional[Any] = '\n# Transformers 설치 방법\n! pip install transformers datasets\n# 마지막 릴리스 대신 소스에서 설치하려면, 위 명령을 주석으로 바꾸고 아래 명령을 해제하세요.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' lowerCAmelCase_ : List[Any] =...
692
1
"""simple docstring""" import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def lowercase_ ( _lowerCamelCase: i...
366
"""simple docstring""" import gc import threading import time import psutil import torch class _snake_case : def __init__( self : str ): __lowerCamelCase : Optional[Any] = psutil.Process() __lowerCamelCase : List[Any] = False ...
366
1
'''simple docstring''' import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def lowe...
366
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase : str = logging.get_logger(__name__) _lowerCAmelCase : ...
246
0
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.util...
717
import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import ...
643
0
from collections.abc import Sequence def __UpperCamelCase ( _A , _A = False ): if not arr: return 0 lowerCAmelCase_ = 0 if allow_empty_subarrays else float('''-inf''' ) lowerCAmelCase_ = 0.0 for num in arr: lowerCAmelCase_ ...
431
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureEx...
431
1
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.ut...
721
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import l...
219
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attenti...
92
'''simple docstring''' from __future__ import annotations def _lowerCAmelCase ( __magic_name__ : list[float] ) -> float: lowercase : Any =0.0_0 lowercase : Tuple =0 for resistor in resistors: if resistor <= 0: l...
92
1
_UpperCAmelCase = { 0: "0", 1: "1", 2: "2", 3: "3", 4: "4", 5: "5", 6: "6", 7: "7", 8: "8", 9: "9", 10: "a", 11: "b", 12: "c", 13: "d", 14: "e", 15: "f", } def _lowerCamelCase ( _a ): """simple docstring""" assert type(_a...
297
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _lowerCamelCase ( _a , _a , _a ): """simple docstring""" _lowerCamelCase ...
297
1
import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAIN...
348
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class __lowerCAmelCase ( unittest.TestCase ): lowerCamelCase_ : Tuple = inspect....
60
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Dict = { 'ksst...
704
from numpy import exp, pi, sqrt def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Dict , __UpperCamelCase : float = 0.0 , __UpperCamelCase : float = 1.0 ) -> int: """simple docstring""" return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu)...
55
0
import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data impo...
99
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
328
0
'''simple docstring''' import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem lowercase_ : Any = importlib.util.find_spec('''s3fs''') is not None if _has_safs: ...
715
'''simple docstring''' import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_visio...
653
0
import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin A_...
604
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a__ ( lowercase__ , lowercase__ , lowercase__=1_0_2_4 , lowercase__=1_0_2_4 , lowercase__=False , ...
54
0
"""simple docstring""" import re import subprocess import sys a_ : Optional[int] = subprocess.check_output('''git merge-base main HEAD'''.split()).decode('''utf-8''') a_ : Optional[int] = ( subprocess.check_output(F"""git diff --diff-filter=d --name-only {fork_point_sha}...
711
"""simple docstring""" import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from ...
263
0
from __future__ import annotations import math def _snake_case ( __snake_case ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes ...
10
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _lowerCamelCase ( UpperCamelCase , UpperCamelCase ): """simple docstring""" @re...
590
0
SCREAMING_SNAKE_CASE = [ (1_0_0_0, 'M'), (9_0_0, 'CM'), (5_0_0, 'D'), (4_0_0, 'CD'), (1_0_0, 'C'), (9_0, 'XC'), (5_0, 'L'), (4_0, 'XL'), (1_0, 'X'), (9, 'IX'), (5, 'V'), (4, 'IV'), (1, 'I'), ] def a (lowerCAmelCas...
719
from math import ceil, sqrt def a (lowerCAmelCase__ = 1_000_000 ): __a = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: __a = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) ...
209
0
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowercase_ = get_tests_dir("fixtures/test_sentencepie...
11
'''simple docstring''' from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase (__A = "laptop"): """simple docstring""" _a = F'''https://www.amazon.in/laptop/s?k={product}''' _a = { ...
11
1
def _UpperCamelCase ( lowerCAmelCase_ ) ->list[list]: UpperCAmelCase = current_set.copy() for row_index, row in enumerate(lowerCAmelCase_ ): UpperCAmelCase = row[0] for column_index, column in enumerate(lowerCAmelCase_ ): if magnitude == 0: ...
710
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configur...
627
0
from __future__ import annotations import typing from collections.abc import Iterable import numpy as np lowercase_ : str = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 lowercase_ : Optional[int] = typing.Union[np.floataa, int, float] # noqa: UP007 def ...
64
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/config.json""", ...
379
0
from datasets.utils.patching import _PatchedModuleObj, patch_submodule from . import _test_patching def UpperCAmelCase__ (): """simple docstring""" import os as original_os from os import path as original_path from os import rename as original_rename from os.path import dirname ...
715
"""simple docstring""" import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_si...
28
0
def _a ( __UpperCamelCase : int ): return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
233
"""simple docstring""" def __A ( a_ :Tuple , a_ :Union[str, Any] , a_ :int=False) -> List[str]: if isinstance(a_ , a_) and isinstance(a_ , a_): __a : List[str] = len(set_a.intersection(a_)) if alternative...
52
0
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class snake_case : '''simple docstring''' snake_case_ : Optional[Union[str, Path]] = None snake_case_ : bool ...
198
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, ...
198
1
from __future__ import annotations import os from typing import Any import requests SCREAMING_SNAKE_CASE : Union[str, Any] = "https://api.github.com" # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user SCREAMING_SNAKE_CASE : List[Any] ...
635
from __future__ import annotations def lowerCAmelCase_ ( _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , ): if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError("""You cannot supp...
635
1
import random def A ( __UpperCAmelCase ) -> bool: '''simple docstring''' UpperCAmelCase_ = num - 1 UpperCAmelCase_ = 0 while s % 2 == 0: UpperCAmelCase_ = s // 2 t += 1 for _ in range(5 ): Up...
561
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def A ( ) -> Optional[int]: '''simple docstring''' raise RuntimeError('''CUDA out of memory.''' ) ...
561
1
"""simple docstring""" import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSe...
289
import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger snake_case : Optional[int] = get_logger(__name__) snake_case : Union[str, Any] = R"\n Args:\n input_ids (`jnp.ndarray` ...
124
0
'''simple docstring''' import argparse lowerCAmelCase_ = 'docs/source/_static/js/custom.js' def A__ ( A : Dict): '''simple docstring''' with open(A , encoding="utf-8" , newline="\n") as f: UpperCamelCase : Union[str, Any] = f.readlines()...
435
'''simple docstring''' class UpperCAmelCase_ : """simple docstring""" def __init__( self , lowerCamelCase ) -> Dict: '''simple docstring''' UpperCamelCase : Union[str, Any] = arr.split("," ) def SCREAMING_SNAKE_CASE__ ( ...
435
1
import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedul...
2
import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class A_ ( __a , ...
428
0
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
319
'''simple docstring''' import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup __lowerCAmelCase = logging.get_logger(__name__) ...
319
1
'''simple docstring''' import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers i...
685
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'uclanlp/visualbert-vqa': 'https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json', 'uclanlp/visualbert-vqa-pre': 'https://h...
685
1
"""simple docstring""" from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_avail...
715
"""simple docstring""" import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import Pretraine...
165
0
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass c...
515
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : int = logging.get_logger(__name__) lowercase__ : Any = { "MIT/ast-finetuned-audioset-10-10-0.4593": ( "https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/m...
515
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging A__: Dict = logging.get_logger(__name__) class A__ ( UpperCAmelCase__ ): __UpperCamelCase : Optional[int] = "timm_backbone" ...
700
'''simple docstring''' from functools import lru_cache def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : int ) -> set: _a : Any =2 _a : Tuple =set() while i * i <= n: if n % i: i ...
506
0