code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch('socket.socket' ) @patch('builtins.open' ) def UpperCAmelCase_ ( __UpperCAmelCase : int , __UpperCAmelCase : int ) -> Union[str, Any]: # ===== initialization ===== SCREA...
31
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _a ( UpperCAmelCase__ ): """simple docstring""" def _UpperCAmelCase ( self , _UpperCAmelCase ) -> Dict: with open(_UpperC...
23
0
import functools import logging import os import sys import threading from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional import huggingface_hub.utils as hf_hub_u...
719
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class _SCREAMING_SNAKE_CASE ( __UpperCamelCase ): _A : Any = (KDPMaDiscreteScheduler,) _A : Dict = 10 def A_...
530
0
'''simple docstring''' __UpperCAmelCase = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformer...
90
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = { """xlm-roberta-base""": """http...
558
0
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase ( snake_case_ ): SCREAMING_SNAKE_CASE__ = '''ClapFeatureExtractor''' SCREAMING_SNAKE_CASE__ = ('''RobertaTokenizer''', '''RobertaTokenizerFast''') ...
664
def UpperCAmelCase__ ( _SCREAMING_SNAKE_CASE : Optional[Any] )->Any: # noqa: E741 _lowerCAmelCase = len(_SCREAMING_SNAKE_CASE ) _lowerCAmelCase = 0 _lowerCAmelCase = [0] * n _lowerCAmelCase = [False] * n _lowerCAmelCase = [False] * n def d...
664
1
import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging _lowerCAmelCase = logging.get_logger(__name__) def lowercase ( _a=None ,_a=None ) -> List[Any]: return field...
137
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { """google/bit-50""": """https://hu...
137
1
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def _SCREAMING_SNAKE_CASE ( snake_case , snake_case , snake_case , snake_case ) -> Tuple: _UpperCAmelCase = s.rsplit(snake_cas...
175
import itertools import string from collections.abc import Generator, Iterable def _SCREAMING_SNAKE_CASE ( snake_case , snake_case ) -> Generator[tuple[str, ...], None, None]: _UpperCAmelCase = iter(snake_case ) while True: _U...
175
1
from ...processing_utils import ProcessorMixin class lowercase ( _UpperCAmelCase ): lowerCamelCase : List[str] = '''SpeechT5FeatureExtractor''' lowerCamelCase : Optional[Any] = '''SpeechT5Tokenizer''' def __init__( self : int , _lowercase : L...
35
import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu f...
198
0
import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class _a : _UpperCamelCase: Dict = None def _snake_case ( self ) -> List[Any]: lowerCAmelCase : Dict = ...
711
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : Optional[int] =logging.get_logger(__name__) lowerCAmelCase : Optional[int] ={ 'transfo-xl-wt103': 'https://huggingface.co/transfo-xl-wt103/resolve/main/config.json', }...
693
0
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from util...
656
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def ...
656
1
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester fro...
719
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTes...
569
0
import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_ima...
423
from __future__ import annotations from collections.abc import MutableSequence class a__ : def __init__( self : Optional[Any] , A_ : int , A_ : MutableSequence[float] ) -> None: """simple docstring"...
423
1
import os from pathlib import Path def snake_case_ ( ): '''simple docstring''' from torch.utils.cpp_extension import load _lowerCAmelCase =Path(lowercase__ ).resolve().parent.parent.parent / """kernels""" / """deformable_detr""" _lowerCAmelCas...
706
from typing import TYPE_CHECKING from ...utils import _LazyModule __SCREAMING_SNAKE_CASE : Optional[Any] = {'''tokenization_wav2vec2_phoneme''': ['''Wav2Vec2PhonemeCTCTokenizer''']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys ...
149
0
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from ...
606
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING __a : Tuple = logging.get_logger(__name__) ...
606
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transform...
709
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration __UpperCAmelCase : Optional[int] = 500000 __UpperCAmelCase , __UpperCAmelCase : Any = os.path.split(__file__) __UpperCAmelCase : int = os.path...
643
0
def _UpperCAmelCase (UpperCamelCase__ : int ): if length <= 0 or not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError("Length must be a positive integer." ) return [n * (2 * n - 1) for n in range(UpperCamelCase__ )] if __name__ == "__main__":...
503
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def _UpperCAmelCase (UpperCamelCase__ : U...
503
1
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, ...
620
from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time UpperCamelCase__ : int = Lock() def _UpperCAmelCase ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE ...
620
1
'''simple docstring''' import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class _SCREAMING_SNA...
316
'''simple docstring''' def lowerCamelCase_ ( A_ , A_ ): __lowerCamelCase = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __lowerCamelCase = n - k # Calculate C(n,k) for i in range(A_ ): result *= n - i result //= i + 1 retur...
316
1
"""simple docstring""" import gc import threading import time import psutil import torch class SCREAMING_SNAKE_CASE_ : '''simple docstring''' def __init__( self) -> Optional[int]: '''simple docstring''' snake_case__ : Union[str, Any] = psut...
150
"""simple docstring""" import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers....
150
1
'''simple docstring''' import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMix...
245
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attenti...
245
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class UpperCAmelCase ( unittest.T...
718
import warnings from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType clas...
181
0
import pprint import requests SCREAMING_SNAKE_CASE = 'https://zenquotes.io/api' def a (): return requests.get(API_ENDPOINT_URL + """/today""" ).json() def a (): return requests.get(API_ENDPOINT_URL + """/random""" ).json() if __name__ == "__main__": ...
99
def _lowercase( __a : list[int] ): a__ =len(__a ) for i in range(__a ): for j in range(i + 1 , __a ): if numbers[j] < numbers[i]: a__ , a__ =numbers[j], numbers[i] return numbers ...
20
0
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipeline...
243
from __future__ import annotations def lowercase_ (A : list[int] ): return len(set(A ) ) == len(A ) if __name__ == "__main__": import doctest doctest.testmod()
243
1
import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common imp...
540
from __future__ import annotations def _UpperCAmelCase ( a__): '''simple docstring''' if len(a__) == 0: return [] a_ , a_ : List[Any] = min(a__), max(a__) a_ : Tuple = int(max_value - min_value) + 1 a_ : list[list] = ...
540
1
"""simple docstring""" lowerCAmelCase__ = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' lowe...
716
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_...
544
0
import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Tuple ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = [ """safet...
6
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Enco...
6
1
'''simple docstring''' import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCL...
665
'''simple docstring''' from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar a_ = TypeVar('T') class __SCREAMING_SNAKE_CASE ( Generic[T] ): snake_case_ = 42 # Cache store of keys snake_cas...
665
1
'''simple docstring''' import math import sys def A_ ( SCREAMING_SNAKE_CASE_ ) ->int: if number != int(SCREAMING_SNAKE_CASE_ ): raise ValueError("""the value of input must be a natural number""" ) if number < 0: raise ValueError("""the value of input must not be a negative n...
451
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case = { """configuration_git""": ["""GIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GitConfig""", """GitVisionConfig"""], """processing_git""": ["""GitProc...
451
1
from __future__ import annotations class lowerCAmelCase__ : def __init__( self : Dict , __UpperCamelCase : Optional[int]=None ) -> Any: A = data A = None def __repr__( self : int ) -> Optional[int]: ...
709
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor...
224
0
import argparse from collections import defaultdict def lowerCAmelCase_ ( __a , __a , __a , __a , __a ) -> Optional[int]: """simple docstring""" lowerCamelCase__: str =F"""{file}_{class_name}_{test_name}""" done_test[_id] ...
59
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import ...
59
1
import math def lowerCamelCase__ ( _a , _a): if ( not isinstance(_a , (int, float)) or power_factor < -1 or power_factor > 1 ): raise ValueError("power_factor must be a valid float value between -1 and 1.") return apparent_power * power_factor def lowerCamelCase__ ( _a ...
193
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__A ) , 'Tatoeba directory doe...
193
1
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, ...
3
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_a...
313
0
import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version(""">=""", FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_pla...
522
import random from typing import Any def UpperCAmelCase ( lowercase ): """simple docstring""" for _ in range(len(lowercase ) ): __lowercase = random.randint(0 , len(lowercase ) - 1 ) __lowercase = random....
522
1
import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers....
324
from __future__ import annotations from functools import lru_cache from math import ceil __magic_name__: Tuple = 100 __magic_name__: Any = set(range(3, NUM_PRIMES, 2)) primes.add(2) __magic_name__: int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: contin...
324
1
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration lowerCamelCase_ = HfArgumentParser(InitializationArguments) lowerCamelCase_ = parser.parse_args() # Load codeparrot tokenizer trained for Python code to...
714
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBer...
86
0
from math import factorial SCREAMING_SNAKE_CASE :dict[str, int] = {str(digit): factorial(digit) for digit in range(10)} def UpperCAmelCase ( a_ ) -> int: """simple docstring""" if not isinstance(a_ , a_ ): raise TypeError("Parameter number must...
55
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remot...
406
0
import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructure...
704
import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings lowerCAmelCase = logging.getLogge...
429
0
"""simple docstring""" from typing import Any class UpperCAmelCase : """simple docstring""" def __init__( self , _UpperCAmelCase ): lowercase__: List[str] = data lowercase__: Optional[Any] = None def __repr__( self ): return F"""Node({se...
586
"""simple docstring""" def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase = 1_0 , __UpperCAmelCase = 2_2 ) -> int: lowercase__: Optional[Any] = range(1 , __UpperCAmelCase ) lowercase__: Any = range(1 , __UpperCAmelCase ) return sum( 1 for power in po...
586
1
"""simple docstring""" import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, ...
397
"""simple docstring""" from typing import Any def lowercase__ ( snake_case_ :list , snake_case_ :list , snake_case_ :dict , snake_case_ :dict , snake_case_ :dict , ): _validation( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_...
397
1
'''simple docstring''' from __future__ import annotations import bisect def UpperCAmelCase__ ( UpperCAmelCase_ : list[int] , UpperCAmelCase_ : int , UpperCAmelCase_ : int = 0 , UpperCAmelCase_ : int = -1 ) -> int: if hi...
13
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer lowerCAmelCase__ ...
596
0
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def lowercase__ ( __A: str ,__A: Union[str, Any]=None ): '''simple docstring''' __magic_name__ : Union[str, Any] =...
721
from sklearn.metrics import mean_squared_error import datasets __lowerCamelCase : List[str] = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel,...
501
0
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import SeqaSeqTrainer from seqaseq_training_args import SeqaSeqTrainingArguments import transformers from transformers import ( AutoConfig, ...
19
"""simple docstring""" from math import sqrt def lowercase (snake_case__ : int ) -> int: '''simple docstring''' lowerCAmelCase = 0 for i in range(1 , int(sqrt(snake_case__ ) + 1 ) ): if n % i == 0 and i != sqrt(snake_case__ ): ...
169
0
"""simple docstring""" __lowercase = {str(digit): digit**5 for digit in range(10)} def lowercase ( A_ )-> int: '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(A_ ) ) def lowercase ( )-> int: '''simpl...
135
"""simple docstring""" import numpy as np def lowercase ( A_ , A_ , A_ , A_ , A_ )-> Tuple: '''simple docstring''' a : List[str] = int(np.ceil((x_end - xa) / h ) ) a : Optional[int] = np.zeros((n ...
135
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _UpperCamelCase = { """configuration_groupvit""": [ """GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GroupViTConfig""", ...
453
"""simple docstring""" from __future__ import annotations _UpperCamelCase = [] def SCREAMING_SNAKE_CASE ( lowercase__ , lowercase__ , lowercase__ ) -> bool: for i in range(len(lowercase__ ) ): if board[row][i] == 1: return False for i in range(len(lowercas...
453
1
"""simple docstring""" def __A ( a_ : int = 1 , a_ : int = 10_00 )-> int: '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = 1 SCREAMING_SNAKE_CASE : List[str] = 0 for divide_by_number in range(a_ , digit + 1 ): SCREAMIN...
18
"""simple docstring""" def __A ( a_ : int )-> list[int]: '''simple docstring''' if num <= 0: raise ValueError('''Input must be a positive integer''' ) SCREAMING_SNAKE_CASE : Optional[int] = [True] * (num + 1) SCREAMING_SNAKE_CASE : Optiona...
18
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.s...
428
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "encoder-decoder" lowerCamelCase_ = ...
6
0
from __future__ import annotations def lowerCAmelCase_ ( __a , __a ) -> int: """simple docstring""" if len(__a ) < k or k < 0: raise ValueError("Invalid Input" ) lowerCamelCase__: Dict =sum(array[:k] ) for i in range(len(__a ) - k ): lowerCamelCase__: Opti...
437
import os import sys import unittest __A = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_mod...
437
1
from __future__ import annotations def lowercase ( __A : list[list[int]] ) -> bool: '''simple docstring''' snake_case : Dict = len(__A ) # We need to create solution object to save path. snake_case : Union[str, Any] = [[0 for _ in rang...
36
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4...
72
0
"""simple docstring""" # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won'...
282
"""simple docstring""" import gc import threading import time import psutil import torch class _UpperCamelCase : '''simple docstring''' def __init__( self ): __lowerCAmelCase = psutil.Process() __lowerCAmelCase = False def ...
282
1
import fcntl import os import socket import torch import torch.distributed as dist def __SCREAMING_SNAKE_CASE ( *a__ : Union[str, Any] ) -> Union[str, Any]: with open(a_ ,"""r""" ) as fh: fcntl.flock(a_ ,fcntl.LOCK_EX ) try: print(*a_ ) finally: fcntl...
17
"""simple docstring""" import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class __lowercase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = (DDPMScheduler,) ...
52
0
'''simple docstring''' from collections import defaultdict class lowerCamelCase : '''simple docstring''' def __init__( self : Optional[Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Optional[int] ) -> Union[str, Any]: '''simple docstring''' A__ ...
721
'''simple docstring''' import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __snake_case : List[Any] = logging.get_logger(__name__) class lowerCamelCase ( lowercase_...
687
0
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging a :int = logging.get_logger(__name__) a :Optional[Any] = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-trans...
680
"""simple docstring""" import torch from diffusers import StableDiffusionPipeline UpperCamelCase__ = 'path-to-your-trained-model' UpperCamelCase__ = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('cuda') UpperCamelCase__ = 'A photo of sks dog in a buck...
110
0
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A__ : List[str] =logging.get_logger(__name__) A__ : List[Any] ={ 'asapp/sew-tiny-100k': 'https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config....
499
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def A_ ( __SCREAMING_SNAKE_CASE : Dict ) -> Optional[int]: """simple docstring""" if not...
499
1
lowerCamelCase_ = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' lowerCamelCase_ = [{'''type''': '...
513
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) lowerCa...
513
1
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder _UpperCamelCase: Optional[int] =datasets.utils.logging.get_logger(__name__) class __lowercase( folder_based_builder.FolderBasedBuilderConfig ): ...
585
from PIL import Image def _a ( __SCREAMING_SNAKE_CASE : Image ): """simple docstring""" _lowerCAmelCase , _lowerCAmelCase = image.size _lowerCAmelCase = 0 _lowerCAmelCase = image.load() for i in range(__SCREAMING_SNAKE_CASE ): for j in range(_...
585
1
from __future__ import annotations def a ( snake_case__: list , snake_case__: int , snake_case__: int , snake_case__: int ): '''simple docstring''' lowercase_ = [] lowercase_ , lowercase_ = input_list[low:mid], input_list[mid : high + 1] ...
97
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmen...
604
0
import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available fr...
711
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) a : List[Any] = { 'configuration_mobilevit': ['MOBILEVIT_PRETRAINED_CONFIG_ARCHI...
593
0
"""simple docstring""" import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is...
633
'''simple docstring''' from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torc...
523
0
import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMSchedule...
365
import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters __snake_case : str = (7_20, 12_80) # Height, Width __snake_case : Dict = (0.4, 0.6) # if height or width lower than this scale, drop it. __snake_case ...
365
1
from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig __UpperCamelCase : List[Any] = { 'susnato/ernie-m-base_pytorch': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json', 'susnato/ernie-m-large_pytorc...
519
import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class _UpperCamelCase ( tf.keras.optimizers.schedules.LearningRateSched...
519
1
'''simple docstring''' import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_...
568
'''simple docstring''' import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin ...
568
1
'''simple docstring''' def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return sum(e for e in range(3 , lowercase_ ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F"""{solution() = }""")
72
import shutil import tempfile import unittest from unittest.mock import patch from transformers import ( DefaultFlowCallback, IntervalStrategy, PrinterCallback, ProgressCallback, Trainer, TrainerCallback, TrainingArguments, is_torch_available, ) from transformers.testing_utils import...
464
0
'''simple docstring''' def UpperCAmelCase_ ( __lowercase : int = 1000 ) -> int: '''simple docstring''' return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
119
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __SCREAMING_SNAKE_CASE :List[str] = { '''configuration_swiftformer''': [ '''SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', ...
119
1
SCREAMING_SNAKE_CASE__ : dict[tuple[int, int, int], int] = {} def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> int: '''simple docstring''' # if we are absent twice, or late 3 consecutive days, # ...
79
'''simple docstring''' import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def a ( UpperCamelCase_ : Any ) -> List[str]: snake_case__ =os.path.join(args.tf_model_dir , 'parameters.json' ) ...
538
0
import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import re...
717
import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=5) -> Optional[int]: # Adapted from https://github.com/pytorch/fa...
155
0
"""simple docstring""" import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...t...
93
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandin...
505
0
import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :Dataset , SCREAMING_SNAKE_CASE :Dict[str, str] ) -> ...
240
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :Optional[Any]=() , SCREAMING_SNAKE...
240
1
import inspect from typing import Callable, List, Optional, Union import torch from transformers import ( CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, WhisperForConditionalGeneration, WhisperProcessor, ) from diffusers import ( AutoencoderKL, DDIMScheduler, Diff...
16
"""simple docstring""" from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase = logging.get_logger(__name__) lowercase = { '''snap-research/efficientformer-l1-300''': ( '''https://huggingface.co/...
573
0
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class lowercase : """simple docstring""" a__ = 42 a__ = None a__ = None def ...
720
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowerCAmelCase__ = { """configuration_blip""": [ """BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BlipConfig""...
648
0
"""simple docstring""" def A ( __snake_case: float , __snake_case: int ) -> Dict: """simple docstring""" if digit_amount > 0: return round(number - int(__UpperCAmelCase ) , __UpperCAmelCase ) return number - int(__UpperCAmel...
545
import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class __SCREAMING_SNAKE_CASE ( UpperCamelCase): """simple docstring""" __UpperCAmelCase = (DDIMParallelScheduler,) __UpperCAmelCase = (("eta", 0.0)...
576
0
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = ArgumentParser( des...
704
def _SCREAMING_SNAKE_CASE ( lowercase : list[int] , lowercase : list[int] ): '''simple docstring''' lowerCamelCase_ = len(lowercase ) print('The following activities are selected:' ) # The first activity is always selected lowerC...
651
0
import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.mode...
551
def UpperCamelCase_( _A :Union[str, Any] )-> List[str]: UpperCamelCase__ = [0] * len(_A ) UpperCamelCase__ = [] UpperCamelCase__ = [] UpperCamelCase__ = 0 for values in graph.values(): for i in values: indegree[i] += 1 for...
551
1
# coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applic...
504
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin...
504
1
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modelin...
98
"""simple docstring""" import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers impo...
373
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: a_ :Dict =...
243
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import...
243
1
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from...
136
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline _lowercase : str =argparse.ArgumentParser("Stable Diffusion script with intel optimization", add_help=False) parser.add_argument("--dpm",...
136
1
from ....configuration_utils import PretrainedConfig from ....utils import logging __UpperCAmelCase : Dict = logging.get_logger(__name__) __UpperCAmelCase : str = { "speechbrain/m-ctc-t-large": "https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json", ...
155
def A__ ( SCREAMING_SNAKE_CASE__ = 1000) -> int: __snake_case , __snake_case: Dict = 1, 1 __snake_case: int = 2 while True: __snake_case: str = 0 __snake_case: Any = fa + fa __snake_case , __snake_case: Tuple ...
155
1
'''simple docstring''' import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE ) class snake_case__ ( __SCREAMING_SNAKE_CASE ...
638
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load...
117
0
'''simple docstring''' import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE( ...
163
'''simple docstring''' import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoPr...
163
1
"""simple docstring""" import os import pytest from attr import dataclass _snake_case = "us-east-1" # defaults region @dataclass class _a : a_ : List[str] = 42 a_ : Optional[Any] = 'arn:aws:iam::558105141721:role/sagemaker_exe...
510
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowerCAmelCase = { 'configuration_wav2vec2': ['WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Wav2Vec2Co...
585
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """google/mobilen...
349
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_visio...
349
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCamelCase :Dict = { 'configuration_xlm_roberta_xl': [ 'XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaXLConfig', ...
222
"""simple docstring""" import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRobertaT...
222
1
class _UpperCamelCase : '''simple docstring''' def __init__( self : Union[str, Any] , a : Any ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = arr.split("," ) def __UpperCamelCase ( self ...
717
from __future__ import annotations def lowerCamelCase__ ( _a , _a): if b == 0: return (1, 0) ((SCREAMING_SNAKE_CASE) ,(SCREAMING_SNAKE_CASE)) : Tuple = extended_euclid(_a , a % b) SCREAMING_SNAKE_CASE : Dict = a // b return (y, x - k * y) def ...
193
0
import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class UpperCAmelCase_ ( lowerCamelCase__ ): '''simple docstring''' UpperCamelCase__ : Dict = ['''image_processor''', '''tokenizer'''] UpperCamelCase__ ...
148
'''simple docstring''' import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, B...
597
0
import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_commo...
705
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) _SCREAMING_SNAKE_CASE : Optional[Any] = {'configuration_beit': ['BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BeitConfig',...
206
0
def UpperCAmelCase ( a_ ) -> list: """simple docstring""" if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __A = gray_code_sequence_string(a_ ) # # convert them to integers for i ...
55
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowercase ( ...
313
0
import json import os import re import unicodedata from json.encoder import INFINITY from typing import Any, Dict, List, Optional, Tuple, Union import numpy as np import regex from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncodin...
706
from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : int = logging.get_logger(__name__) __magic_name__ : Optional[Any] = { '''uclanlp/visualbert-vqa''': '''https://huggingface.co/uclanlp/visualbert-v...
410
0
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _lowerCAmelCase ( __UpperCAmelCase , unittest.TestCase ): SCREAMING_SN...
621
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed...
621
1
'''simple docstring''' import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available ...
700
'''simple docstring''' import numpy as np def UpperCamelCase_ ( A__ : np.array ): '''simple docstring''' return 1 / (1 + np.exp(-vector )) def UpperCamelCase_ ( A__ : np.array ): '''simple do...
398
0
"""simple docstring""" from abc import ABC, abstractmethod from argparse import ArgumentParser class a__ ( __magic_name__ ): @staticmethod @abstractmethod def a_ ( UpperCamelCase_ : ArgumentParser): """simple docstring""" raise NotImplementedError() ...
77
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torc...
77
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(snake_case_ ) , "Tatoe...
601
import string import numpy def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: int , __lowerCamelCase: int ): '''simple docstring''' return b if a == 0 else greatest_common_divisor(b % a , __lowerCamelCase ) class __lowerCamelCase : """simple docstring"...
601
1
import os from datetime import datetime as dt from github import Github a__ = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''enhancement''', '''new pipeline/model''', '''new scheduler''', '''wip''', ] def __Uppe...
14
def _lowerCamelCase ( __A : int ) -> str: _UpperCAmelCase : Tuple = int(__A ) if decimal in (0, 1): # Exit cases for the recursion return str(__A ) _UpperCAmelCase , _UpperCAmelCase : int = divmod(__A , 2 ) ...
485
0
"""simple docstring""" class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = graph ...
708
"""simple docstring""" import string import numpy def A ( snake_case :int , snake_case :int ) -> int: return b if a == 0 else greatest_common_divisor(b % a , snake_case ) class __lowerCAmelCase : lowercase = string.ascii_uppercase + string.digits #...
293
0
from math import factorial def lowerCAmelCase_ ( lowercase: int = 100 ) -> int: '''simple docstring''' return sum(int(__UpperCamelCase ) for x in str(factorial(__UpperCamelCase ) ) ) if __name__ == "__main__": print(solution(int(input('''Enter the Number: ''').strip())))
271
'''simple docstring''' import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel SCREA...
301
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by a...
718
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, ...
149
0
from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("To use the rich extension, install rich with `pip install rich`")
632
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCamelCase : Tuple = { """configuration_mctct""": ["""MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MCTCTConfig"""], """feature_extraction_mctct""": ["""MCTCTFeatureExtract...
297
0
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import evaluate import numpy as np from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForSequenceClassification, Aut...
700
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { "bert-base-uncased": "https://huggingface.co/bert-base-uncas...
526
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiec...
133
'''simple docstring''' from __future__ import annotations from scipy.special import comb # type: ignore class UpperCAmelCase_ : def __init__( self : Union[str, Any] , UpperCAmelCase__ : list[tuple[float, float]] ) -> Optional[int]: lowerC...
133
1
'''simple docstring''' import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def _snake_case ( A , A=() , A=None , A="no" , A="2...
98
'''simple docstring''' from __future__ import annotations class a__ : '''simple docstring''' def __init__( self , lowerCamelCase_ = 0 ) -> List[Any]: lowerCAmelCase__ = key def _...
98
1
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavave...
53
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _lowercase: Any = {'''configuration_swin''': ['''SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwinConfig''', '''SwinOnnxConfig''']} try: if not is_torch_available(): ...
192
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...imag...
630
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) lowerCAmelCase : int = { """configuration_trocr""...
630
1