code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accel...
94
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def lowercase_ ( __A : dict , __A : str , __A : set , __A : set , __A : dict , __A : dict , __A : ...
94
1
"""simple docstring""" import re def lowerCamelCase_ ( UpperCAmelCase_ ) ->bool: """simple docstring""" __UpperCAmelCase : str = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' )...
706
"""simple docstring""" from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar lowercase__ :int = TypeVar('T') class snake_case ( Generic[T] ): '''simple docstring''' def _...
374
0
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTes...
631
import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_...
631
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torc...
636
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import conv...
636
1
'''simple docstring''' from math import isqrt def __UpperCamelCase( _A : int ): '''simple docstring''' return all(number % divisor != 0 for divisor in range(2 , isqrt(_A ) + 1 ) ) def __UpperCamelCase( _A : int = 10**6 ): '''simple docstring''' Upp...
614
'''simple docstring''' from __future__ import annotations def __UpperCamelCase( _A : list[int] , _A : int , _A : int , _A : int ): '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[in...
614
1
"""simple docstring""" import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAM...
709
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A__ : List[str]= logging.get_logger(__name__) class __lowerCamelCase ( _a ): a : Optional[int] ="""timm_backbone""" def __init__( self , snak...
20
0
import timeit import numpy as np import datasets from datasets.arrow_writer import ArrowWriter from datasets.features.features import _ArrayXD def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Any: """simple docstring""" def wrapper(*lowercase_ , **lowercase_ ): ...
87
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils i...
621
0
"""simple docstring""" from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageRes...
19
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : Tuple = { '''configuration_pegasus_x''': ['''PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PegasusXConfig'''], } try: if not is_torch_available(): r...
19
1
'''simple docstring''' import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @datacla...
13
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ : List[str] = logging.get_logger(__name__) UpperCamelCase__ : str = { ...
614
0
from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging __lowerCamelCase : List[Any] = logging.get_logger(__name__) __lowerCamelCase : List[Any] = { "google/umt5-small"...
708
import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class __magic_name__ ( A__, unittest.Test...
457
0
from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch("socket.socket" ) @patch("builtins.open" ) def lowerCAmelCase_ ( __a , __a ) -> Dict: """simple docstring""" lowerCamelCase__: int =Mock() lowerCamelCase...
59
from __future__ import annotations SCREAMING_SNAKE_CASE_:Tuple = """#""" class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self ): A : dict = {} def _lowerCAmelCase ( self, lowerCamelCase__ ): A : List[A...
662
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase : str = logging.get_logger(__name__) __UpperCAmelCase : List[Any] = { 'microsoft/wavlm-base': 'https://huggingface.co/microsoft...
701
import pytest __UpperCAmelCase : int = "__dummy_dataset1__" __UpperCAmelCase : int = "\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\"\nURLS = {\"train\": REPO_URL + \"wikiann...
57
0
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES snake_case__ : int = logging.get_logger(__name__) snake_case__ : ...
23
from __future__ import annotations __A : str = list[tuple[int, int]] __A : Optional[int] = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [...
16
0
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def lowerCAmelCase__ ( a__ , a__=None ) ->str: '''simple docstring''' _UpperCamelCase = None if token is not None: _UpperCamel...
82
lowerCamelCase__ = '''Alexander Joslin''' import operator as op from .stack import Stack def lowerCAmelCase__ ( a__ ) ->int: '''simple docstring''' _UpperCamelCase = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _UpperCamelCase = Stack() _UpperCam...
82
1
from ...processing_utils import ProcessorMixin class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' a__ = """SpeechT5FeatureExtractor""" a__ = """SpeechT5Tokenizer""" def __init__( self : Union[str, Any] , Up...
529
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, requi...
378
0
from math import factorial, radians def A__ ( __A , __A = 18 , __A = 10 ): '''simple docstring''' _lowerCamelCase : List[Any] = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians _lowerCamelCase ...
15
from __future__ import annotations class __snake_case : '''simple docstring''' def __init__( self : Tuple , _UpperCamelCase : int = 0) ->str: """simple docstring""" _lowerCamelCase : Union[str, Any] = ke...
15
1
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_...
85
"""simple docstring""" def lowerCamelCase__ ( _lowerCamelCase : float , _lowerCamelCase : float ) -> float: if density <= 0: raise ValueError('Impossible fluid density' ) if bulk_modulus <= 0: raise ValueError('Impossible bulk modul...
549
0
from abc import ABC, abstractmethod from argparse import ArgumentParser class lowerCamelCase__ ( UpperCAmelCase ): """simple docstring""" @staticmethod @abstractmethod def snake_case__ ( snake_case ): '''simple docstring''' raise NotImp...
185
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __UpperCamelCase = { 'configuration_resnet': ['RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ResNetConfig', 'ResNetOnnxConfig'...
185
1
import re def _a ( lowerCAmelCase )-> list: return [char.split() for char in re.split(R'[^ a-z A-Z 0-9 \s]' , str_ )] def _a ( lowerCAmelCase )-> str: SCREAMING_SNAKE_CASE_ = split_input(str_ ) return "".join( [''.join([char.capit...
360
from math import isqrt def _a ( lowerCAmelCase )-> bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase ) + 1 ) ) def _a ( lowerCAmelCase = 10**6 )-> int: SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ ...
360
1
"""simple docstring""" def _UpperCAmelCase ( lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" lowerCAmelCase__ = """""" for i in table: res += inp[i - 1] return res def _UpperCAmelCase ( lowerCamelCase__ ): """simple doc...
704
"""simple docstring""" from __future__ import annotations from math import gcd def _UpperCAmelCase ( lowerCamelCase__ , lowerCamelCase__ = 2 , lowerCamelCase__ = 1 , lowerCamelCase__ = 3 , ): """simple docstring""" if num < 2: raise ValueError("""The input value ca...
674
0
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_a...
32
from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=A__ ): __A : str = ["""torch""", """scipy"""] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ): requires_backends(self , ['''torc...
32
1
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = {'''vocab_file''': '''vocab.json'''} ...
404
"""simple docstring""" from collections.abc import Sequence from queue import Queue class UpperCAmelCase : def __init__( self : Any , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase ...
404
1
def A__ ( SCREAMING_SNAKE_CASE_ : int = 10_00 ) -> int: """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = 1, 1 _UpperCAmelCase = 2 while True: _UpperCAmelCase = 0 _UpperCAmelCase = ...
32
from math import sqrt def A__ ( SCREAMING_SNAKE_CASE_ : int ) -> bool: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all mult...
32
1
from __future__ import annotations def A ( lowercase__ : list[int] ) -> int: if not nums: return 0 UpperCamelCase__ :Dict = nums[0] UpperCamelCase__ :Dict = 0 for num in nums[1:]: UpperCamelCase__ :Optional[Any] = ( max_excluding + ...
709
def A ( lowercase__ : List[str]=2_8123 ) -> Union[str, Any]: UpperCamelCase__ :Optional[Any] = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i Up...
383
0
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase : Optional[int] = { """configuration_mctct""": ["""MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MCTCTConfig"""], """feature_extraction_mctct""": ["""MCTCTFeatureExtr...
302
from cva import destroyAllWindows, imread, imshow, waitKey def A_ ( A__ ) -> Tuple: # getting number of pixels in the image a__ , a__ : Any = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(A__...
302
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A = { '''configuration_blenderbot''': [ '''BLENDERBOT_PRETRAINED_CONFIG_...
719
from __future__ import annotations import math def __UpperCAmelCase ( __A , __A ) -> float: '''simple docstring''' UpperCAmelCase__ = u for i in range(1 , __A ): UpperCAmelCase__ = temp * (...
277
0
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from tran...
316
'''simple docstring''' import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_...
435
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import lo...
701
'''simple docstring''' import unittest import numpy as np from transformers import MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING, TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING from transformers.pipelines import AudioClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipe...
517
0
"""simple docstring""" import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { """vocab_file""": """vocab.json""", """to...
178
"""simple docstring""" import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch _...
178
1
"""simple docstring""" import operator as op def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->Any: '''simple docstring''' a : str = [] a : List[str] = lambda _lowercase , _lowercase : int(x / y ...
31
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/lic...
31
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case = { """configuration_al...
67
'''simple docstring''' from collections import defaultdict def UpperCAmelCase ( A : int ): SCREAMING_SNAKE_CASE : List[Any] = 1 SCREAMING_SNAKE_CASE : Dict = True for v in tree[start]: if v not in visited:...
527
0
"""simple docstring""" def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ): '''simple docstring''' if number < 0 or shift_amount < 0: raise ValueError('both inputs must be positive integers' ) lowercase__ : Optional[Any] ...
645
"""simple docstring""" import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() _UpperCamelCase : Dict = logging.get_logger(__name__) _UpperCamelCase : List[Any] ...
645
1
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration UpperCAmelCase__ : str = 5_00_00 UpperCAmelCase__ : int = 50_00 UpperCAmelCase__ , UpperCAmelCase__ : Tuple = os.path.split(__fil...
313
"""simple docstring""" import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version(""">=""", FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp ...
409
0
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = ['image_processor', 'tokenizer'] UpperCamelCase = 'CLIPImageProcess...
718
from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class A: '''simple docstring''' UpperCamelCase = 42 UpperCamelCase = None UpperCamelCase = None lowerCamelCase : str ...
651
0
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_availabl...
175
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor,...
175
1
def __SCREAMING_SNAKE_CASE ( UpperCamelCase : float , UpperCamelCase : float ) -> float: """simple docstring""" if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / den...
708
def __SCREAMING_SNAKE_CASE ( UpperCamelCase : int ) -> int: """simple docstring""" a_ = 1 for i in range(1 , num + 1 ): fact *= i return fact def __SCREAMING_SNAKE_CASE ( UpperCamelCase : int ) -> int: """simple docstring""" a_ = 0 while numb...
403
0
'''simple docstring''' import socket def _lowercase ( ) -> Dict: """simple docstring""" __UpperCAmelCase : List[str] = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) __UpperCAmelCase : Dict = socket.getho...
168
'''simple docstring''' from ....configuration_utils import PretrainedConfig from ....utils import logging _a : Optional[int] = logging.get_logger(__name__) _a : List[str] = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van...
168
1
'''simple docstring''' from typing import Dict from .base import GenericTensor, Pipeline class snake_case__ ( SCREAMING_SNAKE_CASE_ ): def A_ ( self : Optional[int] , __a : Optional[int]=None , __a : Optional[int]=None , __a : Union[str, Any]=...
124
'''simple docstring''' import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) A__ : Union[str, Any] ...
124
1
import glob import os import random from string import ascii_lowercase, digits import cva a__ : List[str] = '' a__ : Optional[int] = '' a__ : int = '' a__ : List[Any] = 1 # (0 is vertical, 1 is horizontal) def UpperCAmelCase_ ( ...
188
from __future__ import annotations def snake_case( __magic_name__ , __magic_name__ ) -> list[list[int]]: '''simple docstring''' lowercase : list[list[int]] = [] lowercase : list[int] = [] lowercase ...
217
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import...
268
from ...configuration_utils import PretrainedConfig from ...utils import logging _a: List[str] = logging.get_logger(__name__) _a: Any = { """sayakpaul/vit-msn-base""": """https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json""", # See all ViT MSN models at https://h...
268
1
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 100 ): _snake_case = 1 ...
585
'''simple docstring''' from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeli...
585
1
import random def __lowercase ( UpperCAmelCase__ ): """simple docstring""" __lowerCAmelCase = num - 1 __lowerCAmelCase = 0 while s % 2 == 0: __lowerCAmelCase = s // 2 t += 1 for _ in range(5 ): __lower...
102
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAM...
102
1
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=lowerCAmelCase__ ) class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_...
182
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_...
182
1
import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( 'The ...
703
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def _A ( ): """simple docstring""" lowerCAmelCase__ = [randint(-1000 , 1000 ) for i in range(10 )] lowerCAmelCase__ ...
125
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ = { "configuration_owlvit": [ "OWLV...
631
def lowercase ( a , a , a , a ): '''simple docstring''' SCREAMING_SNAKE_CASE_ :int = [False] * len(a ) SCREAMING_SNAKE_CASE_ :List[Any] = [] queue.append(a ) SCREAMING_SNAKE_CASE_ :int = True while queue: SCREAMING_SNAKE_CASE_...
631
1
import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --h...
169
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def A (__A : Tuple , __A : List[Any]=None ) -> Optional[int]: """simple docstring""" ...
169
1
'''simple docstring''' import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Optional[Any] = { 'vocab...
649
'''simple docstring''' from typing import Dict, List, Optional from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase : str = logging.get_logger(__name__) lowercase : Optional[Any] = { 'nielsr/canine-s': 2_0_4_...
649
1
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_...
711
def __lowerCAmelCase ( UpperCamelCase ) -> bool: if not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(UpperCamelCase ) == 0: raise ValueError('''Input list must be a non empty lis...
470
0
import math import random from typing import Any from .hill_climbing import SearchProblem def __lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str] = True , UpperCAmelCase__ : Tuple = math.inf , UpperCAmelCase__ : Un...
272
from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class __snake_case ( ...
193
0
import math def _lowerCAmelCase ( _a : list , _a : int = 0 , _a : int = 0 ) -> list: lowerCAmelCase_ : List[str] = end or len(_a ) for i in range(_a , _a ): lowerCAmelCase_ : List[Any] = i lowerCAme...
440
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class lowercase__ ( unittest.TestCase ): __UpperCamelCase = inspect.getfil...
440
1
"""simple docstring""" def __snake_case ( _lowercase ): """simple docstring""" if n == 1 or not isinstance(_lowercase ,_lowercase ): return 0 elif n == 2: return 1 else: UpperCamelCase = [0, 1] for i in range(2 ,n + 1 ...
34
"""simple docstring""" import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_C...
34
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase_ = { """configuration_lilt""": ["""LILT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LiltConfig"""], } try: if not is_torch_available(): raise OptionalDependen...
470
import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope='''session''' ) def __lowerCAmelCase ( ) -> Optional[...
470
1
import math import sys def __lowerCAmelCase ( UpperCamelCase ) -> str: lowerCAmelCase__ : Dict = '''''' try: with open(UpperCamelCase , '''rb''' ) as binary_file: lowerCAmelCase__ : str = binary_file.read() for dat in data: lowerCAm...
678
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { """naver-clova-ix/donut-base""": """https://huggingface.co/naver-clova-ix/donut-base/resolve/main/config.json""", # See all Donut ...
678
1
import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlitea import sql...
712
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets __A =datasets.logging.get_logger(__name__) __A ='''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Ma...
313
0
'''simple docstring''' import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class UpperCAmelCase_ (tf.keras.optimizers.schedules.Learni...
13
'''simple docstring''' A__ : dict[tuple[int, int, int], int] = {} def UpperCAmelCase__ ( UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int ) -> int: # if we are absent twice, or late 3 consecutive days, ...
13
1
"""simple docstring""" import inspect import unittest class __UpperCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self : Tuple ) -> Any: '''simple docstring''' try: import diffusers # noqa: F401 except ImportError: ...
251
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2...
251
1
class A__ : """simple docstring""" def __init__( self : List[Any] , lowerCamelCase__ : str , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[str] ): a__ : str = name a__ : Optional[int] = value a__ : Dict = we...
37
import heapq def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : list[list] =[] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with...
21
0
from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_form...
719
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _SCREAMING_SNAKE_CASE : Dict = { '''configuration_longformer''': [ '''LONGFORMER_PRETRAINED_CONFIG_AR...
472
0
"""simple docstring""" SCREAMING_SNAKE_CASE_ = 9.80665 def __snake_case ( _lowercase ,_lowercase ,_lowercase = g ): """simple docstring""" if fluid_density <= 0: raise ValueError('''Impossible fluid density''' ) if volume < 0: raise Valu...
34
"""simple docstring""" import operator def __snake_case ( _lowercase ,_lowercase = False ,_lowercase = None ): """simple docstring""" UpperCamelCase = operator.lt if reverse else operator.gt UpperCamelCase = solution or [] if not arr: ...
34
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ : List[str] = { 'configuration_clipseg': [ 'CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CLIPSegConfig', 'CLIPSegTextConfig', 'CLIP...
484
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a_ : int = {'configuration_van': ['VAN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'VanConfig']} try: if not is_torch_available(): raise Op...
484
1
from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def UpperCamelCase ( lowercase_ = "isbn/0140328726" ) -> dict: '''simple docstring''' lowercase__ : Dict = olid.strip().strip("""/""" ) # Remove leading/trailing w...
12
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder lowerCamelCase__ : Any = datasets.utils.logging.get_logger(__name__) class _snake_case ( folder_based_builder.FolderBasedBuilderConfig ...
12
1
from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEn...
83
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'vocab_file': 'vocab.jso...
83
1
import argparse import hashlib # hashlib is only used inside the Test class import struct class __snake_case : def __init__( self ,a_ ): """simple docstring""" lowerCAmelCase__ = data lowerCAmelCase__ = [0X67452301, 0XEFCDAB89, 0X98BADCFE, 0X10325476, 0XC3D2E1F...
193
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { "microsoft/markuplm-base": "https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json", "microsoft/markuplm-large": "...
256
0
from __future__ import annotations from typing import Any def __lowercase( UpperCAmelCase__ ): """simple docstring""" create_state_space_tree(UpperCAmelCase__ , [] , 0 ) def __lowercase( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ): ...
710
from math import pi def __lowercase( UpperCAmelCase__ , UpperCAmelCase__ ): """simple docstring""" return 2 * pi * radius * (angle / 360) if __name__ == "__main__": print(arc_length(9_0, 1_0))
484
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer a__ : str =logging.get_logger(__name__...
399
'''simple docstring''' import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_att...
683
0
"""simple docstring""" from numpy import exp, pi, sqrt def lowerCamelCase ( _snake_case ,_snake_case = 0.0 ,_snake_case = 1.0 ): return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
254
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf UpperCamelCase__ = logging.ge...
254
1
'''simple docstring''' def _lowerCAmelCase ( _UpperCamelCase : str , _UpperCamelCase : List[Any] ) -> List[Any]: """simple docstring""" if discount_rate < 0: raise ValueError('Discount rate cannot be negative' ) if not cash_flows: raise ValueError(...
405
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a: Optional[int] = logging.get_logger(__name__) _a: Optional[Any] = { """...
162
0
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->str: A__ : Tu...
498
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : str, UpperCAmelCase__ : List[Any] ) ->List[Any]: A__ : Union[str, Any] = [1] for i in range(2, UpperCAmelCase__ ): factorials.append(factorials[-1] * i ) ...
498
1
"""simple docstring""" import os from collections.abc import Iterator def __magic_name__ ( lowercase = "." ): for dir_path, dir_names, filenames in os.walk(lowercase ): SCREAMING_SNAKE_CASE_: Any =[d for d in dir_names if d != """scripts""" and d[0] not in """._"""]...
409
"""simple docstring""" from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class a ( UpperCAmelCase__ ): def __init__( ...
409
1
"""simple docstring""" import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, ...
509
"""simple docstring""" from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class A_ ( _UpperCAmelCase ): """simple docstring""" lowercase : Tuple = ["image_processor", "tokenizer"] lowercase : str = "AutoImageProcessor" ...
509
1
'''simple docstring''' import argparse import copy def lowerCamelCase( SCREAMING_SNAKE_CASE_ ) -> Optional[int]: A_ = {} with open(__lowerCamelCase ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: A_ = [] _lis...
366
import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUE...
468
0
"""simple docstring""" import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE fr...
712
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArgu...
309
0
'''simple docstring''' import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_confi...
675
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a_ : Any = get_tests_dir("""fixture...
675
1
"""simple docstring""" import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) fr...
556
"""simple docstring""" import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class __a ( _lowerCAmelCase ): UpperCamelCase_ : Any = (EulerDiscreteScheduler,) UpperCamelCase...
556
1
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, re...
66
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase = logging.get_logger(__name__) __lowerCamelCase = { """kssteven/ibert-roberta-base""": """https:/...
204
0
'''simple docstring''' from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class lowercase ( _lowercase ): """simple docstring""" def __lt__( self , __snake_case): return self[-1]...
711
lowerCAmelCase__ = range(2, 2_0 + 1) lowerCAmelCase__ = [1_0**k for k in range(ks[-1] + 1)] lowerCAmelCase__ = {} def lowerCamelCase_ ( UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : A...
648
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class _lowerCAmelCase ( __a ): _lowercase =42 try: ...
290
from math import loga def lowerCamelCase__ ( __lowerCAmelCase : int ): """simple docstring""" if a < 0: raise ValueError("Input value must be a positive integer" ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): raise TypeError("Inp...
290
1
from functools import reduce lowerCamelCase__ = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''668966...
702
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_devic...
408
0
from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function _lowerCamelCase =1.0_5_4_5_7_1_8_1_7E-3_4 # unit of ℏ : J * s _lowerCamelCase =3E8 # unit of c : m * s^-1 def _a ( lowerCamelCase, low...
681
from functools import reduce _A = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "668966489504452445231617318564030987111217223831...
290
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models....
714
import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCamelCase_ ( lowercase , lowercase ): @register_to_config def __init__( self , *, lowerCamelCase_ = 4 , lowerC...
589
0
class UpperCAmelCase__ : """simple docstring""" def __init__( self : Optional[int] ) -> str: SCREAMING_SNAKE_CASE__ = {} def lowercase_ ( self : int ) -> List[Any]: print(self.vertex ) for i in self.vertex: ...
493
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __lowerCAmelCase ( __magic_name__ , __magic_name__ , __magic_name__=1_0_2_4 , __magic_name__=1_0_2_4 , __magic_name__=False , *...
226
0
import itertools import os import random import tempfile import unittest import numpy as np from transformers import TvltFeatureExtractor, is_datasets_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_t...
594
import random from .binary_exp_mod import bin_exp_mod def __lowerCamelCase ( __a : List[Any] , __a : Optional[Any]=1_000 ) -> Union[str, Any]: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd _lowercase =n - 1 _lowercase =0 whil...
594
1
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaToken...
505
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device ...
505
1
'''simple docstring''' import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def _lowerCAmelCase ( *_lowerCAmelCase )-> Optional[Any]: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): __UpperCAmelCase = ...
617
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A: Tuple = logging.get_logger(__name__) class UpperCAmelCase ( UpperCAmelCase_ ): _A : List[Any] = """timm_backbone""" def __init__( self ,...
617
1
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedule...
143
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_param...
143
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer __A = logging.get_logger(__name__) __A =...
437
from __future__ import annotations def lowerCAmelCase_ ( __a , __a , __a , ) -> tuple: """simple docstring""" if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError("You cannot supply more or less than 2 values" ) elif electron_conc < 0: ...
437
1
import math import random def _A ( __magic_name__ , __magic_name__ = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value _snake_case = 0.02 def _A ( __magic_name__ , __magic_name__ ): lowercase__ = ...
655
import math import random def _A ( __magic_name__ , __magic_name__ = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value _snake_case = 0.02 def _A ( __magic_name__ , __magic_name__ ): lowercase__ = ...
655
1
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score ...
590
from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging UpperCAmelCase_ : str = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Union[tf.Tensor, np.ndarray] ) -> List[int]: ...
590
1
import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def lowercase ( _lowerCAmelCase , _lowerCAmelCase=1 ): if n_shave_prefix_segments >= 0: return ".".join(path.split(""".""" )[n_shave_prefix_segments:] ) ...
392
'''simple docstring''' import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tok...
28
0
"""simple docstring""" from __future__ import annotations from decimal import Decimal from numpy import array def UpperCAmelCase__ ( lowerCAmelCase__ :list[list[float]] ) -> list[list[float]]: '''simple docstring''' lowercase = Decimal ...
197
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ ( lowerCAmelCase__ :list[float] , lowerCAmelCase__ :list[float] ) -> float: '''simple docstring''' lowercase = sorted(numsa + numsa ) lowercase , l...
197
1
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Opti...
354
from manim import * class __a ( SCREAMING_SNAKE_CASE ): def UpperCamelCase ( self : Tuple)-> Dict: __lowerCAmelCase =Rectangle(height=0.5 , width=0.5) __lowerCAmelCase =Rectangle(height=0.4_6 , width=0.4_6).set_stroke(width=0) _...
354
1
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class lowercase : def __init__( self ) -> Any: """simple docstring""" UpperCamelCase = '' UpperCamelCase = '' UpperCamelCase = [] UpperCamelCase = 0 ...
3
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput _UpperCAmelCase : str = "scheduler_config.json" class lowercase ( _SC...
3
1
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging UpperCAmelCase__ = logging.get_logger(_...
351
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available SCREAMING_SNAKE_CASE = {"tokenization_herbert": ["HerbertTokenizer"]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependen...
579
0
from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class A ( ...
706
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_uti...
559
0
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class _SCREAMING_SNAKE_CASE ( _lowerCAmelCase ): a_ : Optional[int] = "M-CLIP" def __init__(self , UpperCAmelCase=1_0_2_4 , UpperCAmelCase=7_6_8 , **UpperCAmelCase): ...
132
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONT...
282
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available A : Optional[Any] = { "configuration_groupvit": [ "GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "GroupViTConfig", "GroupViTOnnxConfig", ...
356
import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings A : Tuple = r"\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n can be used to control the model outputs. Read the...
356
1
import warnings from .generation import TFGenerationMixin class lowerCAmelCase_ ( a__ ): # warning at import time warnings.warn( "Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will " "be removed in Transformer...
40
'''simple docstring''' import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def _lowerCAmelCase ( lowercase : str , lowercase : str , **lowercase : Tuple ) ->Tuple: """simple docstring""" lowercase...
161
0
'''simple docstring''' import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEI...
710
'''simple docstring''' import requests from bsa import BeautifulSoup def lowerCAmelCase ( UpperCamelCase__ : str = "AAPL" ): """simple docstring""" __UpperCAmelCase = f"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" __UpperCAmelCase = Beautif...
654
0
from __future__ import annotations import typing from collections.abc import Iterable import numpy as np __SCREAMING_SNAKE_CASE =typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 __SCREAMING_SNAKE_CASE =typing.Union[np.floataa, int, float] # noqa: UP007 def ...
234
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils im...
484
0
'''simple docstring''' def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not nums: # Makes sure that the list is not empty raise ValueError("List is empty" ) _snake_case = sum(SCREAMING_SNAKE_CASE__ ) / len(SCREAMING_SNAKE_CASE__ ...
368
'''simple docstring''' import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProce...
368
1