code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import...
35
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def a ( A__ ) -> Tuple: ...
35
1
import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils import TOKEN, USER, get_tests_d...
35
import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def a ( A__ , A__ , A__ ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SN...
35
1
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a_ :Optional[int] = logging.get_logger(__name__) def a ( A__ ) -> List[str]: '''simple docstring''' ...
35
from sklearn.metrics import recall_score import datasets a_ :int = '\nRecall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation:\nRecall = TP / (TP + FN)\nWhere TP is the true positives and FN is the false negatives.\n' a...
35
1
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": a_ :int = argparse.ArgumentParser() parser.add_argument('--dump_path', default=None, type=str, required=True,...
35
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available a_ :List[Any] = logging.getLogger(__name__) @dataclass class ...
35
1
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class lowercase ( _UpperCAmelCase , _UpperCAmelCase ): @register_to_config def __init__( ...
35
import os def a ( A__ = "matrix.txt" ) -> int: '''simple docstring''' with open(os.path.join(os.path.dirname(A__ ) , A__ ) ) as in_file: SCREAMING_SNAKE_CASE__ : Optional[Any] = in_file.read() SCREAMING_SNAKE_CASE__ ...
35
1
import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeat...
35
from math import factorial def a ( A__ = 2_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... SCREAMING_SNAKE_CASE__ : Dict =...
35
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowercase ( _UpperCAmelCase ): lowerCamelCase : Optional[Any] = ['''image_processor''', '''tokenizer'''] lowerCamelCase : List[Any] = '''ViTImageProce...
35
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_tim...
35
1
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a_ :Optional[int] = get_logger(__name__) class lowercase ( enum.Enum ): lowerCamelCase : str = '''all_checks''' lowerCamelCase : str ...
35
def a ( A__ ) -> int: '''simple docstring''' if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(A__ , A__ ): raise TypeError('''Input value must be a \'int\' type''' ) return bin(A__...
35
1
import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_utils import GenerationTes...
35
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STAN...
35
1
from __future__ import annotations def a ( A__ , A__ , A__ ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance < 0...
35
from __future__ import annotations from typing import Any class lowercase : def __init__( self : int , _lowercase : int ): SCREAMING_SNAKE_CASE__ : List[str] = num_of_nodes SCREAMING_SNAKE_CASE__ : list[list[int]] ...
35
1
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch a_ :Tuple = l...
35
from typing import TYPE_CHECKING from ...utils import _LazyModule a_ :Tuple = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys a_ :Optional[int] = _LazyMod...
35
1
from cva import destroyAllWindows, imread, imshow, waitKey def a ( A__ ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Any = img.shape[0], img.shape[1] # converting each pixel's color to its negative for...
35
def a ( A__ ) -> str: '''simple docstring''' return "".join([hex(A__ )[2:].zfill(2 ).upper() for byte in list(A__ )] ) def a ( A__ ) -> bytes: '''simple docstring''' if (len(A__ ) % 2) != 0: raise ValueE...
35
1
import itertools import json import os import unittest from transformers import AddedToken, RobertaTokenizer, RobertaTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import T...
35
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class lowercase ( unittest.TestCase ): lowerCamelCase : List[Any] = inspect...
35
1
def a ( A__ ) -> list: '''simple docstring''' if any(not isinstance(A__ , A__ ) or x < 0 for x in sequence ): raise TypeError('''Sequence must be list of non-negative integers''' ) for _ in range(len(A__ ) ): for i,...
35
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ :List[str] = { 'configuration_groupvit': [ 'GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GroupViTConfig', 'GroupViTOnnxConfig', 'Grou...
35
1
import fire from utils import calculate_rouge, save_json def a ( A__ , A__ , A__=None , **A__ ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = [x.strip() for x in open(A__ ).readlines()] SCREAMING_SNAKE...
35
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class lowercase ( _UpperCAmelCase ): def lowercase__ ( self : Optional[int] ): return [ {"col_1": 3, "col_2": "a"}, ...
35
1
def a ( A__ ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = [1] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : int = 0, 0, 0 SCREAMING_SNAKE_CASE__ : Tuple ...
35
import pickle import numpy as np from matplotlib import pyplot as plt class lowercase : def __init__( self : List[str] , _lowercase : Tuple , _lowercase : List[Any] , _lowercase : Tuple , _lowercase : Any , _lowercase : Optional[int] , ...
35
1
def a ( A__ ) -> float: '''simple docstring''' if edge <= 0 or not isinstance(A__ , A__ ): raise ValueError('''Length must be a positive.''' ) return 3 * ((2_5 + 1_0 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def a ( A__ ) ->...
35
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tens...
35
1
def a ( A__ ) -> int: '''simple docstring''' if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(A__ , A__ ): raise TypeError('''Input value must be a \'int\' type''' ) return bin(A__...
35
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers cl...
35
1
def a ( A__ ) -> list: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = False while is_sorted is False: # Until all the indices are traversed keep looping SCREAMING_SNAKE_CASE__ : Optional[Any] = True for ...
35
from __future__ import annotations def a ( A__ , A__ , A__ ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance < 0...
35
1
from math import sqrt def a ( A__ ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = 0 for i in range(1 , int(sqrt(A__ ) + 1 ) ): if n % i == 0 and i != sqrt(A__ ): total += ...
35
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer a_ :Tuple = logging.get_logger(__name__) a_ :Optional[An...
35
1
from math import factorial def a ( A__ = 2_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... SCREAMING_SNAKE_CASE__ : Dict =...
35
import random def a ( A__ ) -> bool: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = num - 1 SCREAMING_SNAKE_CASE__ : Optional[int] = 0 while s % 2 == 0: SCREAMING_SNAKE_CASE__ : Optional[Any] = ...
35
1
def a ( A__ , A__ , A__ ) -> float: '''simple docstring''' if principal <= 0: raise Exception('''Principal borrowed must be > 0''' ) if rate_per_annum < 0: raise Exception('''Rate of interest must be >= 0''' ) if years_...
35
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def a ( A__ ) -> List[Any]: '''simple docstring''' return 1 / (1 + np.exp(-z )) def ...
35
1
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() a_ :str = logging.get_logger(__name__) a_ :Optional[int] = 'https://openaipubl...
35
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def a ( A__ ) -> Tuple: ...
35
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ :List[str] = { 'configuration_groupvit': [ 'GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GroupViTConfig', 'GroupViTOnnxConfig', 'Grou...
35
import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def a ( A__ , A__ , A__ ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SN...
35
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer a_ :Tuple = logging.get_logger(__name__) a_ :int = {'vo...
35
from sklearn.metrics import recall_score import datasets a_ :int = '\nRecall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation:\nRecall = TP / (TP + FN)\nWhere TP is the true positives and FN is the false negatives.\n' a...
35
1
def a ( A__ ) -> str: '''simple docstring''' if not all(char in '''01''' for char in bin_string ): raise ValueError('''Non-binary value was passed to the function''' ) if not bin_string: raise ValueError('''Empty string was passed to the fun...
35
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available a_ :List[Any] = logging.getLogger(__name__) @dataclass class ...
35
1
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def a ( ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = ArgumentParser( desc...
35
import os def a ( A__ = "matrix.txt" ) -> int: '''simple docstring''' with open(os.path.join(os.path.dirname(A__ ) , A__ ) ) as in_file: SCREAMING_SNAKE_CASE__ : Optional[Any] = in_file.read() SCREAMING_SNAKE_CASE__ ...
35
1
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ :str = logging.get_logger(__name__) a_ :Tuple = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class lowercase ( _UpperCAmelCase ): lowerCamelCase : Tuple = '''ct...
35
from math import factorial def a ( A__ = 2_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... SCREAMING_SNAKE_CASE__ : Dict =...
35
1
def a ( A__ , A__ ) -> str: '''simple docstring''' if not isinstance(A__ , A__ ): raise ValueError('''iterations must be defined as integers''' ) if not isinstance(A__ , A__ ) or not number >= 1: raise ValueErro...
35
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_tim...
35
1
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ :str = logging.get_logger(__name__) a_ :Any = { 'weiweishi/roc-bert-base-zh': 'https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json', } class lowercase ( _UpperCAmelCase ...
35
def a ( A__ ) -> int: '''simple docstring''' if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(A__ , A__ ): raise TypeError('''Input value must be a \'int\' type''' ) return bin(A__...
35
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ :List[Any] = logging.get_logger(__name__) a_ :Dict = { 'facebook/deit-base-disti...
35
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STAN...
35
1
from __future__ import annotations def a ( A__ , A__ , A__ ) -> tuple[float, list[float]]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = list(range(len(A__ ) ) ) SCREAMING_SNAKE_CASE__ : Any ...
35
from __future__ import annotations from typing import Any class lowercase : def __init__( self : int , _lowercase : int ): SCREAMING_SNAKE_CASE__ : List[str] = num_of_nodes SCREAMING_SNAKE_CASE__ : list[list[int]] ...
35
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel ...
35
from typing import TYPE_CHECKING from ...utils import _LazyModule a_ :Tuple = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys a_ :Optional[int] = _LazyMod...
35
1
from __future__ import annotations def a ( A__ , A__ , A__ , ) -> tuple[str, float]: '''simple docstring''' if (stress, tangential_force, area).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) e...
35
def a ( A__ ) -> str: '''simple docstring''' return "".join([hex(A__ )[2:].zfill(2 ).upper() for byte in list(A__ )] ) def a ( A__ ) -> bytes: '''simple docstring''' if (len(A__ ) % 2) != 0: raise ValueE...
35
1
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import Bnb...
35
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class lowercase ( unittest.TestCase ): lowerCamelCase : List[Any] = inspect...
35
1
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ :int = logging.get_logger(__name__) a_ :Tuple = { 'facebook/s2t-small-librispeech-asr': ( 'https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json' ), # See all Sp...
35
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ :List[str] = { 'configuration_groupvit': [ 'GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GroupViTConfig', 'GroupViTOnnxConfig', 'Grou...
35
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torc...
35
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class lowercase ( _UpperCAmelCase ): def lowercase__ ( self : Optional[int] ): return [ {"col_1": 3, "col_2": "a"}, ...
35
1
import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfi...
35
import pickle import numpy as np from matplotlib import pyplot as plt class lowercase : def __init__( self : List[str] , _lowercase : Tuple , _lowercase : List[Any] , _lowercase : Tuple , _lowercase : Any , _lowercase : Optional[int] , ...
35
1
from functools import reduce a_ :Optional[Any] = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '6689664895044524452316...
35
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tens...
35
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig a_ :Optional[int] = logging.get_logger(__name__) a_ :Any = { 'Intel/dpt-large': 'https://huggingface.co/Intel/dpt-large/resolve/main/config.json', # See all DPT mod...
35
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers cl...
35
1
from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig a_ :Union[str, Any] = logging.get_logger(__name__) a_ :Union[str, Any] = 'T5Config' class lowercase ( _UpperCAmelCase ...
35
from __future__ import annotations def a ( A__ , A__ , A__ ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance < 0...
35
1
import pytest import datasets # Import fixture modules as plugins a_ :Union[str, Any] = ['tests.fixtures.files', 'tests.fixtures.hub', 'tests.fixtures.fsspec'] def a ( A__ , A__ ) -> Union[str, Any]: '''simple docstring''' for item in items: if a...
35
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer a_ :Tuple = logging.get_logger(__name__) a_ :Optional[An...
35
1
from sklearn.metrics import mean_squared_error import datasets a_ :Tuple = '\\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer,...
35
import random def a ( A__ ) -> bool: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = num - 1 SCREAMING_SNAKE_CASE__ : Optional[int] = 0 while s % 2 == 0: SCREAMING_SNAKE_CASE__ : Optional[Any] = ...
35
1
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as Proph...
35
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def a ( A__ ) -> List[Any]: '''simple docstring''' return 1 / (1 + np.exp(-z )) def ...
35
1
from ...processing_utils import ProcessorMixin class lowercase ( _UpperCAmelCase ): lowerCamelCase : List[str] = '''SpeechT5FeatureExtractor''' lowerCamelCase : Optional[Any] = '''SpeechT5Tokenizer''' def __init__( self : int , _lowercase : L...
35
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def a ( A__ ) -> Tuple: ...
35
1
import argparse import math import traceback import dateutil.parser as date_parser import requests def a ( A__ ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = {} SCREAMING_SNAKE_CASE__ : int = job['''started...
35
import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def a ( A__ , A__ , A__ ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SN...
35
1
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, Fla...
35
from sklearn.metrics import recall_score import datasets a_ :int = '\nRecall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation:\nRecall = TP / (TP + FN)\nWhere TP is the true positives and FN is the false negatives.\n' a...
35
1
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...feat...
35
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available a_ :List[Any] = logging.getLogger(__name__) @dataclass class ...
35
1
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENT...
35
import os def a ( A__ = "matrix.txt" ) -> int: '''simple docstring''' with open(os.path.join(os.path.dirname(A__ ) , A__ ) ) as in_file: SCREAMING_SNAKE_CASE__ : Optional[Any] = in_file.read() SCREAMING_SNAKE_CASE__ ...
35
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class...
35
from math import factorial def a ( A__ = 2_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... SCREAMING_SNAKE_CASE__ : Dict =...
35
1
import argparse import json import logging import os import shutil import sys import tempfile import unittest from unittest import mock import torch from accelerate.utils import write_basic_config from transformers.testing_utils import TestCasePlus, get_gpu_count, run_command, slow, torch_device from transformers....
35
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_tim...
35
1
def a ( A__ = 1_0_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = (n * (n + 1) // 2) ** 2 SCREAMING_SNAKE_CASE__ : int = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": p...
35
def a ( A__ ) -> int: '''simple docstring''' if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(A__ , A__ ): raise TypeError('''Input value must be a \'int\' type''' ) return bin(A__...
35
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING a_ :Optional[Any] = logging.get_logger(__name__) a_ :Dict = { 'SenseTime/deformable-detr': 'https://huggingface.co/sensetime/deformable-detr/resolve/main/config.j...
35
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STAN...
35
1
import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class lowercase ( _UpperCAmelCase ): lowerCamelCase : List[Any] = (DDIMParallelScheduler,) lowerCamelCase : Union[str, Any] = (('''eta''', 0.0), ('''num_inference_st...
35
from __future__ import annotations from typing import Any class lowercase : def __init__( self : int , _lowercase : int ): SCREAMING_SNAKE_CASE__ : List[str] = num_of_nodes SCREAMING_SNAKE_CASE__ : list[list[int]] ...
35
1
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, and t...
35
from typing import TYPE_CHECKING from ...utils import _LazyModule a_ :Tuple = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys a_ :Optional[int] = _LazyMod...
35
1
from __future__ import annotations def a ( A__ ) -> int: '''simple docstring''' for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(A__ ) )...
35
def a ( A__ ) -> str: '''simple docstring''' return "".join([hex(A__ )[2:].zfill(2 ).upper() for byte in list(A__ )] ) def a ( A__ ) -> bytes: '''simple docstring''' if (len(A__ ) % 2) != 0: raise ValueE...
35
1
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class lowercase ( _UpperCAmelCase ): def lowercase__ ( self : Optional[int] ): return [ {"col_1": 3, "col_2": "a"}, ...
35
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class lowercase ( unittest.TestCase ): lowerCamelCase : List[Any] = inspect...
35
1
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transform...
35
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ :List[str] = { 'configuration_groupvit': [ 'GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GroupViTConfig', 'GroupViTOnnxConfig', 'Grou...
35
1
import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import ...
35
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class lowercase ( _UpperCAmelCase ): def lowercase__ ( self : Optional[int] ): return [ {"col_1": 3, "col_2": "a"}, ...
35
1
# Imports import numpy as np class lowercase : def __init__( self : Dict , _lowercase : List[str]=None , _lowercase : int=None , _lowercase : int=None , _lowercase : Dict=None , _lowercase : Optional[int]=None ): self.set_mat...
35
import pickle import numpy as np from matplotlib import pyplot as plt class lowercase : def __init__( self : List[str] , _lowercase : Tuple , _lowercase : List[Any] , _lowercase : Tuple , _lowercase : Any , _lowercase : Optional[int] , ...
35
1
import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS...
35
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tens...
35
1
def a ( A__ , A__ , A__ ) -> int: '''simple docstring''' if exponent == 1: return base if exponent % 2 == 0: SCREAMING_SNAKE_CASE__ : Tuple = _modexpt(A__ , exponent // 2 , A__ ) % modulo_value ...
35
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers cl...
35
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyImgaImgPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import ...
35
from __future__ import annotations def a ( A__ , A__ , A__ ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance < 0...
35
1
import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common im...
35
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer a_ :Tuple = logging.get_logger(__name__) a_ :Optional[An...
35
1
from heapq import heappop, heappush import numpy as np def a ( A__ , A__ , A__ , A__ , ) -> tuple[float | int, list[tuple[int, int]]]: '''simple docstring''' SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] =...
35
import random def a ( A__ ) -> bool: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = num - 1 SCREAMING_SNAKE_CASE__ : Optional[int] = 0 while s % 2 == 0: SCREAMING_SNAKE_CASE__ : Optional[Any] = ...
35
1
from __future__ import annotations def a ( A__ ) -> list[int]: '''simple docstring''' if len(A__ ) == 0: return array SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : int = min(A__ ), max(A__ ) # Compute th...
35
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def a ( A__ ) -> List[Any]: '''simple docstring''' return 1 / (1 + np.exp(-z )) def ...
35
1
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_tim...
35
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def a ( A__ ) -> Tuple: ...
35
1
def a ( A__ = 1_0_0_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = 3 SCREAMING_SNAKE_CASE__ : Optional[int] = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a ...
35
import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def a ( A__ , A__ , A__ ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SN...
35
1
import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_to...
35
from sklearn.metrics import recall_score import datasets a_ :int = '\nRecall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation:\nRecall = TP / (TP + FN)\nWhere TP is the true positives and FN is the false negatives.\n' a...
35
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ :List[str] = logging.get_logger(__name__) a_ :List[str] = { 'facebook/data2vec-v...
35
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available a_ :List[Any] = logging.getLogger(__name__) @dataclass class ...
35
1
from argparse import ArgumentParser from .env import EnvironmentCommand def a ( ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = ArgumentParser('''Diffusers CLI tool''' , usage='''diffusers-cli <command> [<args>]''' ) S...
35
import os def a ( A__ = "matrix.txt" ) -> int: '''simple docstring''' with open(os.path.join(os.path.dirname(A__ ) , A__ ) ) as in_file: SCREAMING_SNAKE_CASE__ : Optional[Any] = in_file.read() SCREAMING_SNAKE_CASE__ ...
35
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, Di...
35
from math import factorial def a ( A__ = 2_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... SCREAMING_SNAKE_CASE__ : Dict =...
35
1
def a ( A__ = 1_0_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = set() SCREAMING_SNAKE_CASE__ : Tuple = 0 SCREAMING_SNAKE_CASE__ : Union[str, Any] = n + 1 # maximum limit for a in ra...
35
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_tim...
35
1
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ :Tuple = logging.get_logger(__name__) a_ :Any = { 'facebook/nllb-moe-54B': 'https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json', } class lowercase ( _UpperCAmelCase ): ...
35
def a ( A__ ) -> int: '''simple docstring''' if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(A__ , A__ ): raise TypeError('''Input value must be a \'int\' type''' ) return bin(A__...
35
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set...
35
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STAN...
35
1
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def a ( A__ ) -> List[Any]: '''simple docstring''' return 1 / (1 + np.exp(-z )) def ...
35
from __future__ import annotations from typing import Any class lowercase : def __init__( self : int , _lowercase : int ): SCREAMING_SNAKE_CASE__ : List[str] = num_of_nodes SCREAMING_SNAKE_CASE__ : list[list[int]] ...
35
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('>=', '4.25.0')): raise OptionalDependencyNotAvailable() except ...
35
from typing import TYPE_CHECKING from ...utils import _LazyModule a_ :Tuple = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys a_ :Optional[int] = _LazyMod...
35
1
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applic...
35
def a ( A__ ) -> str: '''simple docstring''' return "".join([hex(A__ )[2:].zfill(2 ).upper() for byte in list(A__ )] ) def a ( A__ ) -> bytes: '''simple docstring''' if (len(A__ ) % 2) != 0: raise ValueE...
35
1
from __future__ import annotations def a ( A__ ) -> float: '''simple docstring''' if not nums: raise ValueError('''List is empty''' ) return sum(A__ ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
35
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class lowercase ( unittest.TestCase ): lowerCamelCase : List[Any] = inspect...
35
1
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import ...
35
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ :List[str] = { 'configuration_groupvit': [ 'GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GroupViTConfig', 'GroupViTOnnxConfig', 'Grou...
35
1
import collections import os import re from pathlib import Path a_ :Union[str, Any] = 'src/transformers' # Matches is_xxx_available() a_ :int = re.compile(r'is\_([a-z_]*)_available()') # Catches a one-line _import_struct = {xxx} a_ :List[Any] = re.compile(r'^_import_structure\s+=\s+...
35
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class lowercase ( _UpperCAmelCase ): def lowercase__ ( self : Optional[int] ): return [ {"col_1": 3, "col_2": "a"}, ...
35
1
import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLik...
35
import pickle import numpy as np from matplotlib import pyplot as plt class lowercase : def __init__( self : List[str] , _lowercase : Tuple , _lowercase : List[Any] , _lowercase : Tuple , _lowercase : Any , _lowercase : Optional[int] , ...
35
1
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common i...
35
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tens...
35
1
from sklearn.metrics import recall_score import datasets a_ :int = '\nRecall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation:\nRecall = TP / (TP + FN)\nWhere TP is the true positives and FN is the false negatives.\n' a...
35
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers cl...
35
1
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers cl...
35
from __future__ import annotations def a ( A__ , A__ , A__ ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance < 0...
35
1
def a ( A__ , A__ , A__ , A__ , A__ , ) -> float: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise V...
35
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer a_ :Tuple = logging.get_logger(__name__) a_ :Optional[An...
35
1
import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin a_ :Union[str, Any] = '\nHugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenage...
35
import random def a ( A__ ) -> bool: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = num - 1 SCREAMING_SNAKE_CASE__ : Optional[int] = 0 while s % 2 == 0: SCREAMING_SNAKE_CASE__ : Optional[Any] = ...
35
1
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def a ( A__ ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = [ '''encoder.version''', ''...
35
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def a ( A__ ) -> List[Any]: '''simple docstring''' return 1 / (1 + np.exp(-z )) def ...
35
1
from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNe...
35
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def a ( A__ ) -> Tuple: ...
35
1
from ..utils import DummyObject, requires_backends class lowercase ( metaclass=_UpperCAmelCase ): lowerCamelCase : Dict = ['''speech'''] def __init__( self : Any , *_lowercase : Tuple , **_lowercase : Optional[Any] ): requires_back...
35
import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def a ( A__ , A__ , A__ ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SN...
35
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ :Dict = logging.get_logger(__name__) a_ :List[str] = { 'google/bigbird-roberta-base': 'https://huggingface.co/g...
35
from sklearn.metrics import recall_score import datasets a_ :int = '\nRecall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation:\nRecall = TP / (TP + FN)\nWhere TP is the true positives and FN is the false negatives.\n' a...
35
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...te...
35
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available a_ :List[Any] = logging.getLogger(__name__) @dataclass class ...
35
1
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseT...
35
import os def a ( A__ = "matrix.txt" ) -> int: '''simple docstring''' with open(os.path.join(os.path.dirname(A__ ) , A__ ) ) as in_file: SCREAMING_SNAKE_CASE__ : Optional[Any] = in_file.read() SCREAMING_SNAKE_CASE__ ...
35
1
import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesser...
35
from math import factorial def a ( A__ = 2_0 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... SCREAMING_SNAKE_CASE__ : Dict =...
35
1
import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def a ( A__ , A__=1 ) -> int: '''simple docstring''' if n_shave_prefix_segments >= 0: return ".".join(path.split('''.''' )[n_shave_prefix_s...
35
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_tim...
35
1
def a ( A__ ) -> str: '''simple docstring''' return "".join([hex(A__ )[2:].zfill(2 ).upper() for byte in list(A__ )] ) def a ( A__ ) -> bytes: '''simple docstring''' if (len(A__ ) % 2) != 0: raise ValueE...
35
def a ( A__ ) -> int: '''simple docstring''' if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(A__ , A__ ): raise TypeError('''Input value must be a \'int\' type''' ) return bin(A__...
35
1
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowercase ( enum.Enum ): lowerC...
35
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STAN...
35
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) a_ :List[str] = { 'configuration_swiftformer': [ 'SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SwiftFormerConfig', 'SwiftFormerOnnxConfig', ...
35
from __future__ import annotations from typing import Any class lowercase : def __init__( self : int , _lowercase : int ): SCREAMING_SNAKE_CASE__ : List[str] = num_of_nodes SCREAMING_SNAKE_CASE__ : list[list[int]] ...
35
1
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applic...
35
from typing import TYPE_CHECKING from ...utils import _LazyModule a_ :Tuple = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys a_ :Optional[int] = _LazyMod...
35
1
a_ :Dict = 8.3_1_4_4_6_2 # Unit - J mol-1 K-1 def a ( A__ , A__ , A__ ) -> float: '''simple docstring''' if moles < 0 or kelvin < 0 or volume < 0: raise ValueError('''Invalid inputs. Enter positive value.''' ) return moles * kel...
35
def a ( A__ ) -> str: '''simple docstring''' return "".join([hex(A__ )[2:].zfill(2 ).upper() for byte in list(A__ )] ) def a ( A__ ) -> bytes: '''simple docstring''' if (len(A__ ) % 2) != 0: raise ValueE...
35
1