code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule lowerCAmelCase = {"""processing_wav2vec2_with_lm""": ["""Wav2Vec2ProcessorWithLM"""]} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys ...
174
"""simple docstring""" # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-...
174
1
from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class __A( __lowerCamelCase ): """simple docs...
714
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBer...
86
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.mo...
80
'''simple docstring''' from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMix...
523
0
'''simple docstring''' import torch from torch import nn class UpperCAmelCase_ ( nn.Module ): def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_=1 , lowercase_=False): super().__init__() snake_case_ : Any ...
718
'''simple docstring''' from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase_ ( snake_case__ ): UpperCAmelCase_ = """ClapFeatureExtractor""" UpperCAmelCase_ = ("""RobertaTokenizer""", """RobertaTokenizerFast""") ...
92
0
"""simple docstring""" import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( ...
528
"""simple docstring""" import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def _lowerCamelCase( ): raise RuntimeError("CUDA out of memory." ) class snake_case__ ( ...
528
1
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { '''Salesforce/blip-vqa-base''': '''https://huggingface.co/Salesforce...
702
'''simple docstring''' def _SCREAMING_SNAKE_CASE ( A : float , A : list[float] ) -> float: """simple docstring""" if discount_rate < 0: raise ValueError('Discount rate cannot be negative' ) if not cash_flows: raise ValueE...
61
0
"""simple docstring""" from __future__ import annotations import collections import pprint from pathlib import Path def lowercase (SCREAMING_SNAKE_CASE_ : str ) -> str: return "".join(sorted(SCREAMING_SNAKE_CASE_ ) ) def lowercase (SCREAMI...
247
"""simple docstring""" import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets __UpperCamelCase = '''\ @inproceedings{popovic-2015-chrf, title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation", author = "Popovi{\'c}, Maja", ...
247
1
"""simple docstring""" def lowerCAmelCase_ ( ): '''simple docstring''' __lowerCamelCase : List[str] =[31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31] __lowerCamelCase : List[Any] =6 __lowerCamelCase...
718
"""simple docstring""" import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _UpperCamelCase = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE_ ( snake_case__ ): """simple docstring""" def _...
363
0
"""simple docstring""" import unittest from transformers import MobileBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_c...
82
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) SCREAMING_SNAKE_CASE__ : List[str] = { "configuration_trocr": ["TROCR_PRETRAINED_CONFIG_ARCHIVE_...
85
0
from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling...
582
from math import factorial __UpperCAmelCase = {str(digit): factorial(digit) for digit in range(10)} def _lowerCamelCase ( A_ : int ) -> int: '''simple docstring''' if not isinstance(A_ , A_ ): raise TypeError("Parameter number must be int" ) if number < 0: ...
582
1
def UpperCamelCase_ ( __a ) -> list: a__ : Union[str, Any] = [0] * len(__a ) for i in range(1 , len(__a ) ): # use last results for better performance - dynamic programming a__ : Dict = prefix_result[i - 1] while j > 0 an...
37
from __future__ import annotations from math import pi def _lowercase ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_): """simple docstring""" if (inductance, frequency, reactance).count(0) != 1: raise ValueError("""One and only one argument must be 0""") if...
648
0
import requests from bsa import BeautifulSoup def lowerCAmelCase_ ( lowercase: str , lowercase: dict ) -> str: '''simple docstring''' _UpperCamelCase: Union[str, Any] = BeautifulSoup(requests.get(lowercase , params=lowercase ).content , '''html.parser...
264
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_i...
264
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase : Tuple = logging.get_logger(__name__) lowerCamelCase :...
405
def lowerCamelCase_ ( UpperCamelCase__ : int, UpperCamelCase__ : int ): '''simple docstring''' if b == 0: return 1 if (b % 2) == 0: return actual_power(UpperCamelCase__, int(b / 2 ) ) * actual_power(UpperCamelCase_...
240
0
def _a ( __lowercase , __lowercase ) -> str: """simple docstring""" return "\n".join( F"""{number} * {i} = {number * i}""" for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_ter...
567
def _a ( __lowercase ) -> int: """simple docstring""" if not isinstance(__lowercase , __lowercase ): raise ValueError('Input must be an integer' ) if input_num <= 0: raise ValueError('Input must be positive' ) return sum( ...
567
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging A_ = logging.get_logger(__name__) A_ = '''▁''' A_ = {'''vocab_file''': ...
393
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(a ) , """Tatoeba directo...
403
0
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_f...
253
from collections.abc import Generator from math import sin def __a ( __UpperCAmelCase : bytes ) -> bytes: """simple docstring""" if len(__UpperCAmelCase ) != 32: raise ValueError("Input must be of length 32" ) lowerCamelCase_ : Optional[Any...
253
1
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu __lowerCAmelCase : Any ...
509
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Union[str, Any] = { "RWKV/rwkv-4-169m-pile": "https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/...
509
1
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from ...
713
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQu...
636
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __snake_case = { """configuration_altclip""": [ """ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """AltCLIPConfig""",...
178
import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def UpperCamelCase ( lowercase_ ) -> Any: '''simple...
12
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCAmelCase_ : int = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except Optiona...
718
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase_ : Union[str, Any] = logging.get_logger(__nam...
204
0
"""simple docstring""" import functools def _snake_case ( __snake_case : str , __snake_case : str ): """simple docstring""" _lowerCamelCase : int = len(__snake_case ) _lowerCamelCase : List[Any] = len(__snake_case ) ...
88
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def _snake_case ( __snake_case : float , __snake_case : float , __snake_case : bool = False ): ...
88
1
import math lowerCAmelCase__ = 1_0 lowerCAmelCase__ = 7 lowerCAmelCase__ = BALLS_PER_COLOUR * NUM_COLOURS def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: int = 2_0 ) -> str: '''simple docstring''' A__ = math.comb(SCREAMING_SNAKE_CASE_ , SCREAMING...
626
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf...
626
1
import json import sys def __lowerCAmelCase ( _A ,_A ): """simple docstring""" with open(_A ,encoding="""utf-8""" ) as f: _lowercase = json.load(_A ) _lowercase = ["""<details>""", """<summary>Show updated benchmarks!</summar...
398
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_...
398
1
"""simple docstring""" import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTes...
702
"""simple docstring""" import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _R...
442
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( Diffusion...
566
'''simple docstring''' from __future__ import annotations def lowercase__ ( __UpperCamelCase : list[int] ): '''simple docstring''' if len(__UpperCamelCase ) == 0: return array __lowercase , __lowercase = min(__UpperCamelCase ), max(__UpperCam...
566
1
import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_av...
718
import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import...
234
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ : Dict = {"configuration_fnet": ["FNET_PRETRAINED_CONFIG_ARCHIVE_MAP"...
21
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { "microsoft/markuplm-base": "https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json", "microsoft/markuplm-large": "...
256
0
import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor f...
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotA...
627
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import...
541
'''simple docstring''' def __UpperCAmelCase ( A : list ) -> list: if len(A ) <= 1: return lst UpperCAmelCase_ : List[str] = 1 while i < len(A ): if lst[i - 1] <= lst[i]: i += 1 else: UpperCAmelCase_ , UpperC...
541
1
import os # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_doctest_list.py lowerCamelCase__ = """.""" if __name__ == "__main__": lowerCamelCase__ = os.path.join(REPO_PATH, """utils/documenta...
291
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available from ...utils import OptionalDependencyNotAvailable lowerCamelCase__ = {"""configuration_dpt""": ["""DPT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DPTConfi...
291
1
import argparse import os import re SCREAMING_SNAKE_CASE = 'src/transformers' # Pattern that looks at the indentation in a line. SCREAMING_SNAKE_CASE = re.compile(r'^(\s*)\S') # Pattern that matches `"key":" and puts `key` in group 0. SCREAMING_SNAKE_CASE = ...
99
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE = {'configuration_wavlm': ['WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WavLMConfig']} try: if not is_torch_available(): raise OptionalDependency...
99
1
import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFe...
706
from __future__ import annotations from scipy.special import comb # type: ignore class lowercase : """simple docstring""" def __init__( self : Optional[int] , a_ : list[tuple[float, float]] ): """simple docstring""" lowerCamelCase__ =...
235
0
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTeste...
117
from __future__ import annotations UpperCAmelCase__ = list[tuple[int, int]] UpperCAmelCase__ = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0,...
117
1
"""simple docstring""" import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from...
492
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretra...
492
1
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_c...
10
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs ...
666
0
_snake_case = [ '''DownloadConfig''', '''DownloadManager''', '''DownloadMode''', '''StreamingDownloadManager''', ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadMana...
170
from maths.prime_factors import prime_factors def __lowerCamelCase ( _lowercase ) -> int: if not isinstance(_lowercase , _lowercase ): UpperCamelCase = F'Input value of [number={number}] must be an integer' raise TypeError(_lowercase ) if number < 1: ...
170
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : Union[str, Any] = { "configuration_upernet": ["UperNetConfig"], } try: if not is_torch_available(): ra...
567
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokeniz...
567
1
'''simple docstring''' import math import os import sys def a ( lowerCamelCase__ ): '''simple docstring''' A_ : List[Any] = """""" try: with open(lowerCamelCase__ , """rb""" ) as binary_file: A_ : Dict = binary_file.read() ...
708
'''simple docstring''' import argparse import importlib from pathlib import Path # Test all the extensions added in the setup lowerCamelCase :Any = [ '''kernels/rwkv/wkv_cuda.cu''', '''kernels/rwkv/wkv_op.cpp''', '''kernels/deformable_detr/ms_deform_attn.h''', '''kernels/...
686
0
'''simple docstring''' from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class SCREAMING_SNAKE_CASE ...
8
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tok...
538
0
def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' if not head: return True # split the list to two parts lowerCamelCase : int = head.next, head while fast and fast.next: lowerCamelCase : List[str] = fast.next.next lowerC...
714
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
231
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig _SCREAMING_SNAKE_CASE : Optional[Any] = { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/config.json""", """alber...
344
import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch _SCREAMING_SNAKE_CASE : Any = """sshleifer/bart-tiny...
344
1
from math import ceil def __lowercase ( __lowerCAmelCase : int = 1_0_0_1 ): a__ = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): a__ = 2 * i + 1 a__ = 2 * i a__ = total + 4 * odd**2 - 6 * e...
720
def __lowercase ( __lowerCAmelCase : int ): a__ = generate_pascal_triangle(__lowerCAmelCase ) for row_idx in range(__lowerCAmelCase ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=' ...
657
0
'''simple docstring''' def _lowercase ( __A ,__A ): '''simple docstring''' __UpperCamelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def _lowercase ( __A ,__A ,__A ): '''simple docstring''' ...
601
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/lice...
601
1
"""simple docstring""" from PIL import Image def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase ): __lowercase : Tuple = (2_59 * (level + 2_55)) / (2_55 * (2_59 - level)) def contrast(__UpperCamelCase ) -> int: return int(1_28 + factor...
710
"""simple docstring""" def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : Tuple = len(__UpperCamelCase ) for _ in range(__UpperCamelCase ): for i in range(_ % 2 , arr_size - 1 , 2 ): if arr[i + 1] < arr[i]: __lowercase...
523
0
'''simple docstring''' from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, St...
94
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = ["image_processor", "tokenizer"] lowercase_ = "CLIPImageProc...
59
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics f...
709
"""simple docstring""" from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase : Tuple = logging.get_logger(__name__) UpperCamelCase : List[str] = { "snap-research/efficientformer-l1-300": ( "https://...
293
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, Stab...
699
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.uti...
699
1
"""simple docstring""" import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTest...
121
"""simple docstring""" import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def __a ( _lowercase ): """simple docstring""" lowerCamelCase__ : Any = os.path.join(args.tf_model_dir ...
121
1
'''simple docstring''' from math import sqrt def _lowerCAmelCase ( lowercase : int = 1_0_0_0_0_0_0 ) ->int: """simple docstring""" lowercase__ = 0 lowercase__ = 0 lowercase__ = 42 whil...
161
'''simple docstring''' import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def _lowerCAmelCase ( lowercase : str , lowercase : str , **lowercase : Tuple ) ->Tuple: """simple docstring""" lowercase...
161
1
"""simple docstring""" def snake_case__ ( _SCREAMING_SNAKE_CASE ) ->bool: if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True UpperCAmelCase__ = 4 UpperCAmelCase__ = (1 << p) - 1 for _ in range(p - 2 ): UpperCAmelCase__ = ...
422
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_...
422
1
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractio...
40
"""simple docstring""" from __future__ import annotations from collections.abc import Callable SCREAMING_SNAKE_CASE_ = list[list[float | int]] def lowercase__ ( lowerCAmelCase : Matrix , lowerCAmelCase : Matrix ) -> Matrix: ...
373
0
"""simple docstring""" import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warni...
705
"""simple docstring""" import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration...
101
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Optional[int] = { """Salesforce/blip-vqa-base""": """https:/...
197
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.uti...
197
1
"""simple docstring""" from __future__ import annotations _SCREAMING_SNAKE_CASE = """Muhammad Umer Farooq""" _SCREAMING_SNAKE_CASE = """MIT""" _SCREAMING_SNAKE_CASE = """1.0.0""" _SCREAMING_SNAKE_CASE = """Muhammad Umer Farooq""" _SCREAMING_SNAKE_CASE = ...
705
"""simple docstring""" from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __UpperCamelCase ( ) -> tuple[list[int], int]: """simple docstring""" __snake_case = [randint(-10_00 , ...
614
0
"""simple docstring""" import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __magic_name__ : Optional[int] = Mapping[str, np.ndarray] __magic_name__ : A...
281
"""simple docstring""" __magic_name__ : Optional[Any] = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/' def a_ ( lowercase__ :bytes ): # Make sure the supplied data is a bytes-like object if not isinstance(lowercase__, lowercase__ ): ...
281
1
from manim import * class _lowerCamelCase( _a ): def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : str = Rectangle(height=0.5, width=0.5) _lowercase : List[Any] = Rectangle(height=0.4_6, width...
354
import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor fro...
354
1
def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if len(__lowerCamelCase ) != len(__lowerCamelCase ): raise ValueError("The length of profit and weight must be same." ) if max_weight <= 0: raise ValueError("max_we...
249
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_col...
249
1
'''simple docstring''' import json import os import re import unicodedata from json.encoder import INFINITY from typing import Any, Dict, List, Optional, Tuple, Union import numpy as np import regex from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base impor...
156
'''simple docstring''' import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultip...
156
1
'''simple docstring''' import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConf...
208
'''simple docstring''' UpperCamelCase ="Input must be a string of 8 numbers plus letter" UpperCamelCase ="TRWAGMYFPDXBNJZSQVHLCKE" def snake_case ( a_ : str ) -> bool: """simple docstring""" if not isinstance(a_ , a_ ): ...
208
1
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class lowercase ( __SCREAMING_SNAKE_CASE ): lowercase_ : Optional[int] =(UnCLIPScheduler,) def A__ ( self ,**A__): lowercase ...
721
from __future__ import annotations def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < 2: raise ValueError('''Monogons and Digons are not polygons in the Euclidean space''' ) if any(i <= 0 for i in nums ): raise ValueError('''All ...
633
0
_a : Tuple = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _a : List[Any] = ["a", "b", "c", "d", "e"] def UpperCamelCase__ ( _A: Dict , _A: Any , _A: List[Any] ): '''simple docstring''' __lowerCame...
479
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase__ : Union[str, Any] = { "configuration_m2m_100": ["M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP", "M2M100Config", ...
698
0
'''simple docstring''' import colorsys from PIL import Image # type: ignore def __UpperCAmelCase ( a_: float, a_: float, a_: int ): _UpperCAmelCase : List[Any] = x _UpperCAmelCase : Optional[Any] = y for step in range(a_ ): #...
257
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { 'google/bigbird-roberta-base': 'https:/...
257
1
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ...
65
from functools import reduce __a = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856...
30
0
"""simple docstring""" import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast a : Union[str, Any] = datasets.utils.logging.get_logger(__name__) @dataclas...
704
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _UpperCamelCase ( __UpperCamelCase ): '''simple docstring''' def A__ ( self , __lowercase ): with open(__lowercase , encoding="...
422
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConformerConfig, WavaVecaConformerForCTC, WavaVecaConformerForPreTraining, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ...
300
import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: List[str] , lowerCAmelCase: str , lowerCAmelCase: str )...
300
1
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging A_ : Tuple = logging.get_logger(__name__) def ...
32
import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from utils_ner im...
32
1
from __future__ import annotations def lowercase__ ( A_: list[list[int]] ) -> int: """simple docstring""" for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in rang...
68
'''simple docstring''' def _UpperCAmelCase ( __A : int ): a_ : Optional[Any] = [] a_ : Optional[Any] = [] a_ : List[str] = { '''^''': 3, '''*''': 2, '''/''': 2, '''%''': 2, '''+''': 1, ...
466
0
import json import logging import math import os import sys from dataclasses import dataclass, field from typing import Optional from datasets import Dataset, load_dataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_MASKED_LM_MAPPING, AutoConfig, ...
291
def UpperCamelCase ( snake_case__ : str ,snake_case__ : int ): '''simple docstring''' __snake_case :list[list[str]] = [[] for _ in range(snake_case__ )] __snake_case :Union[str, Any] = key - 1 if key <= 0:...
291
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase__ =logging.get_logger(__name__) UpperCAmelCase_...
616
"""simple docstring""" import os import re import shutil import sys import tempfile import unittest import black UpperCAmelCase__ =os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa:...
616
1
import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask snake_case = logging.getLogger(__name__) class __A ( snake_case__ ): '''simple docstring''' def __init__( self , ...
587
import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller snake_case = 3 def UpperCamelCase_ ( lowerCAmelCase__ ): """simple docstring""" print("Generating primitive root of p" ) while True: _lowerCAmelCase ...
587
1
"""simple docstring""" def SCREAMING_SNAKE_CASE ( snake_case): __snake_case = set() # edges = list of graph's edges __snake_case = get_edges(snake_case) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his ex...
564
"""simple docstring""" import os import numpy import onnx def SCREAMING_SNAKE_CASE ( snake_case, snake_case): __snake_case = a.name __snake_case = b.name __snake_case = '''''' __snake_case = '''''' __snake_case = a == b ...
564
1
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWi...
718
import random def UpperCamelCase ( __lowercase : int ): '''simple docstring''' A_ : Tuple = num - 1 A_ : Optional[Any] = 0 while s % 2 == 0: A_ : Optional[int] = s // 2 t += 1 for _ in range(5 ): A_ ...
70
0
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, Ju...
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) a : Optional[Any] ...
633
0
'''simple docstring''' # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline...
703
'''simple docstring''' from typing import Any, Dict, Optional import torch import torch.nn.functional as F from torch import nn from ..utils import maybe_allow_in_graph from .activations import get_activation from .attention_processor import Attention from .embeddings import CombinedTimestepLabelEmbeddings ...
692
0
"""simple docstring""" import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSample...
299
"""simple docstring""" def lowercase_ ( __UpperCAmelCase ) -> list: lowerCAmelCase__ : List[Any] = len(__UpperCAmelCase ) for i in range(1 , __UpperCAmelCase ): lowerCAmelCase__ : List[Any] = collection[i] lowerCAmelCase...
299
1
from dataclasses import dataclass, field from typing import Optional @dataclass class a : __lowerCAmelCase : Optional[str] = field( default="""codeparrot/codeparrot""" , metadata={"""help""": """Model name or path of model to be trained."""} ) ...
219
import unittest from transformers import DonutProcessor A__ = '''naver-clova-ix/donut-base''' class a ( unittest.TestCase ): def __lowerCamelCase ( self :Optional[int] ): snake_case__ : str = DonutProcessor.from_pretrained(__lowercase )...
219
1
from typing import Any def A__ ( lowerCamelCase ) -> list[Any]: if not input_list: return [] UpperCamelCase_: Optional[Any] = [input_list.count(lowerCamelCase ) for value in input_list] UpperCamelCase_: Tuple = max(lowerCamelCase ) # Gets ...
548
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUID...
548
1
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface ...
714
"""simple docstring""" # Algorithm for the pigeonhole sorting def _A (__a ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = min(__a ) # min() finds the minimum value SCREAMING_SNAKE_CASE_ : int = max(__a ) # max...
176
0
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class _lowerCAmelCase : """simple docstring""" def __init__( self : str): '''simple docstring''' snake_case__ = """""" snake_case__ ...
654
import glob import os import random from string import ascii_lowercase, digits import cva a__ = """""" a__ = """""" a__ = """""" a__ = 1 # (0 is vertical, 1 is horizontal) def _UpperCAmelCase ( ): snake_case__ , snake_case__ = get_dataset(a , a ) print("""P...
654
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() __Uppe...
721
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.u...
582
0
'''simple docstring''' from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging lowerCamelCase : Dict = logging.get_logger(__name__) lowerCamelCase : Union[str, Any] = { 'google/umt5-...
460
'''simple docstring''' def _SCREAMING_SNAKE_CASE (A ) -> bool: """simple docstring""" if not isinstance(A , A ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(A ) == 0: raise ValueError('''Input li...
460
1
"""simple docstring""" from datetime import datetime as dt import os from github import Github _a : List[Any] = [ """good first issue""", """good second issue""", """good difficult issue""", """feature request""", """new model""", """wip""", ] def a__ ( ): """simp...
87
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_c...
87
1
'''simple docstring''' import os import re import shutil import sys import tempfile import unittest import black lowerCamelCase_ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_copies # noqa: E402 #...
418
'''simple docstring''' from heapq import heappop, heappush import numpy as np def SCREAMING_SNAKE_CASE_ ( __A : np.ndarray , __A : tuple[int, int] , __A : tuple[int, int] , __A : bool , ) -> tuple[float | int, list[tuple[int, int]]]: _SCREA...
418
1
import heapq as hq import math from collections.abc import Iterator class _a : def __init__( self: Optional[Any] , UpperCamelCase_: int ) -> Tuple: """simple docstring""" lowercase__ = str(id_ ) low...
710
from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...t...
429
0
'''simple docstring''' from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class lowerCAmelCase ( nn.Module): def __init__( self , __SCREAMING_SNAKE_CASE = 16 , __SCREAMING_SNAKE_CASE = 88 , __SCREAMING_SNAK...
24
import argparse import logging import pickle from collections import Counter logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) _lowerCamelCase = logging.getLogger(__name__) if __name__ == "__main__": _lowerCamelC...
6
0
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=log...
711
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig lowercase = { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/config.json''', ...
41
0
import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepie...
97
import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_commo...
605
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer SCREAMING_SNAKE_CASE__ ...
714
'''simple docstring''' from math import sqrt def lowercase__ ( __UpperCamelCase )-> int: UpperCamelCase = 0 for i in range(1 , int(sqrt(__UpperCamelCase ) + 1 ) ): if n % i == 0 and i != sqrt(__UpperCamelCase )...
35
0
import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version('>=', FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpo...
16
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tens...
35
0
'''simple docstring''' def _A ( A ,A ,A ,A ,A ,A ) -> Any: if index == r: for j in range(lowercase_ ): print(data[j] ,end=" " ) print(" " ) return # When no more elements are there to put in data[] if i >= n: return # cur...
720
'''simple docstring''' import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelT...
425
0
"""simple docstring""" from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging _UpperCamelCase = logging.get_logger(__name__) def _A( lowerCAmelCase ): if isinstance(lowerCAmelCase , np.ndarray ): return list(te...
363
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"vocab_file": "se...
363
1
from collections.abc import Iterable from typing import Generic, TypeVar lowerCAmelCase__ = TypeVar("_T") class _a ( Generic[_T] ): """simple docstring""" def __init__( self , lowerCAmelCase_ = None ): _lowercase =list(iterable or [] ) _lowercase =[] ...
594
import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) class _a ( lowerCamelCase_ ): """simple docstring""" __SCREAMING_SNAKE_CASE ...
594
1
def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Optional[int] ) -> Any: UpperCAmelCase_ = len(__UpperCamelCase ) for i in range(length - 1 ): UpperCAmelCase_ = i for k in range(i + 1 , __UpperCamelCase ): if collection[k] < collecti...
144
def SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase) -> float: if density <= 0: raise ValueError("Impossible fluid density") if bulk_modulus <= 0: raise ValueError("Impossible bulk modulus") return (bulk_modulus / density) ** 0.5 if __name__ == "_...
515
0
"""simple docstring""" from __future__ import annotations import math import random from typing import Any class lowerCamelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' __UpperCAmelCase : list[An...
299
"""simple docstring""" def lowerCamelCase ( _UpperCamelCase : int = 3 , _UpperCamelCase : int = 7 , _UpperCamelCase : int = 1_0_0_0_0_0_0 ) -> int: '''simple docstring''' __UpperCAmelCase : Dict = 0 __UpperCAmelCase : Optional[int] ...
299
1
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_co...
496
'''simple docstring''' from __future__ import annotations def __a ( A__ , A__ = None , A__ = None , A__ = False , ) -> tuple[int, float, str]: lowerCAmelCase = cipher_alphabet or [chr(A__ ) for i in range(97 , 123 ...
649
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCamelCase : Union[str, Any] = { "configuration_time_series_transformer": [ "TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", ...
514
'''simple docstring''' from ..utils import DummyObject, requires_backends class _snake_case ( metaclass=a_ ): SCREAMING_SNAKE_CASE : List[str] = ['''torch''', '''torchsde'''] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): '''si...
514
1
'''simple docstring''' from __future__ import annotations def UpperCamelCase_ ( A__ ): # This function is recursive a_ = len(A__ ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Els...
263
'''simple docstring''' def UpperCamelCase_ ( A__ , A__ , A__ ): if len(A__ ) != len(A__ ): raise ValueError("""The length of profit and weight must be same.""" ) if max_weight <= 0: raise ValueError("""max_weight must greater than zero.""" ) if any(p < 0 for p i...
263
1
'''simple docstring''' __snake_case = range(2, 20 + 1) __snake_case = [10**k for k in range(ks[-1] + 1)] __snake_case = {} def A_ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ->List[str]: lowercase_ = sum(a_i[j] for j...
603
'''simple docstring''' # Lint as: python3 import itertools import os import re __snake_case = re.compile(r"""([A-Z]+)([A-Z][a-z])""") __snake_case = re.compile(r"""([a-z\d])([A-Z])""") __snake_case = re.compile(r"""(?<!_)_(?!_)""") __snake_case = re.compile(r"""(_{2,})""") __snake_case ...
603
1
'''simple docstring''' def lowerCAmelCase__ ( ): for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE__ ): __a : List[Any] = 1 __a : Tuple = 2 while i * i <= ...
597
'''simple docstring''' from __future__ import annotations import numpy as np def A_ ( _lowerCamelCase : list[float] ): return np.maximum(0 , _lowerCamelCase ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
309
0
'''simple docstring''' def __UpperCamelCase ( _UpperCAmelCase ): __UpperCAmelCase : str = min(_UpperCAmelCase ) # min() finds the minimum value __UpperCAmelCase : Any = max(_UpperCAmelCase ) # max() finds the maximum value __UpperCAmelCase : Op...
716
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase__ : str...
329
0