code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def a__ ( ) -> str: __lowerCAmelCase: int = ArgumentParser( d...
346
"""simple docstring""" import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.to...
346
1
from __future__ import annotations def lowerCAmelCase ( snake_case__ : float , snake_case__ : float , snake_case__ : float , )-> tuple[str, float]: if (stress, tangential_force, area).count(0 ) != 1: raise ValueError("You ...
715
import math def lowerCAmelCase ( snake_case__ : float , snake_case__ : float )-> float: return math.pow(snake_case__ , 2 ) - a def lowerCAmelCase ( snake_case__ : float )-> float: return 2 * x def ...
608
0
"""simple docstring""" def __A ( )-> int: '''simple docstring''' return [ a * b * (10_00 - a - b) for a in range(1 , 9_99 ) for b in range(_lowerCAmelCase , 9_99 ) if (a * a + b * b == (10_00 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solu...
698
'''simple docstring''' from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def _A ( _lowerCAmelCase = "isbn/0140328726" ): """simple docstring""" __lowercase =olid.strip().strip('/' ) # Remove leading/trail...
474
0
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() a__ = logging.get_logger('''transformers.models.speecht5''') def snake_case__ ( a ,...
706
'''simple docstring''' from __future__ import annotations class __magic_name__: def __init__( self : Dict , __UpperCamelCase : str , __UpperCamelCase : str ): '''simple docstring''' snake_case__ , snake_case__ = text, pattern ...
566
0
import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import Iter...
615
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a : int = { """configuration_bridgetower""": [ """BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BridgeTowerConfig""", """Bridge...
613
0
"""simple docstring""" import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax...
705
"""simple docstring""" def A__ ( A__ ) -> str: '''simple docstring''' _UpperCAmelCase = "" for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def A__ ( A__ ) -> dict[str, str]: ...
579
0
'''simple docstring''' from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils impor...
451
'''simple docstring''' from typing import List import numpy as np def UpperCAmelCase_ ( __lowerCamelCase : dict ): lowercase_ :Dict = {key: len(__lowerCamelCase ) for key, value in gen_kwargs.items() if isinstance(__lowerCamelCase ,__lowerCamelCase ...
172
0
"""simple docstring""" import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __A = get_tests_dir("""fixtures/test_sentencepiec...
560
"""simple docstring""" import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeli...
560
1
from manim import * class a ( UpperCAmelCase_ ): def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : Any = Rectangle(height=0.5 , width=0.5 ) _UpperCAmelCase : Dict = Rect...
300
'''simple docstring''' import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel f...
51
0
'''simple docstring''' def lowercase__ ( __UpperCamelCase , __UpperCamelCase )-> Any: print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(_lowerCamelCase ): for j in range(_lowerCamelCase ): ...
717
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/license...
35
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING _SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : List[Any] = { "ut/deta": "https://hugg...
550
def UpperCAmelCase__ (UpperCamelCase_ = 10_00 ): """simple docstring""" return sum(2 * a * ((a - 1) // 2) for a in range(3 ,n + 1 ) ) if __name__ == "__main__": print(solution())
550
1
"""simple docstring""" def lowercase__ ( ) -> Tuple: """simple docstring""" _UpperCamelCase : Tuple = 0 for i in range(1 ,1_001 ): total += i**i return str(_lowercase )[-10:] if __name__ == "__main__": print(solution())
712
"""simple docstring""" def lowercase__ ( lowercase_ ) -> set: """simple docstring""" _UpperCamelCase : Union[str, Any] = set() # edges = list of graph's edges _UpperCamelCase : Union[str, Any] = get_edges(lowercase_ ) # Whi...
51
0
'''simple docstring''' import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, requi...
667
import math def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [True] * n _lowerCAmelCase : Optional[int] = False _lowerCAmelCase : Tuple = False _lowerCAmelCase : O...
500
0
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline lowerCAmelCase__ : List[Any] = datasets.utils.logging.get...
711
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_...
699
0
from __future__ import annotations def A__ ( SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : int ) -> list[int]: """simple docstring""" _UpperCAmelCase = 0 _UpperCAmelCase = len(SCREAMING_SNAKE_CASE_ ) - 1 whi...
32
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _SCREAMING_SNAKE_CASE = HfArgumentParser(InitializationArguments) _SCREAMING_SNAKE_CASE = parser.parse_args() # Load codeparrot tokenize...
401
0
import copy import random from transformers import CLIPTokenizer class lowerCAmelCase_ ( lowercase_ ): def __init__( self : Optional[Any] , *UpperCAmelCase_ : Any , **UpperCAmelCase_ : int ) -> Optional[Any]: '''simple docstring'''...
416
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase__ : Optional[int] = get_tests_dir('fixtures...
416
1
from __future__ import annotations def _lowercase ( SCREAMING_SNAKE_CASE_ : List[str] ): """simple docstring""" UpperCamelCase = 2 UpperCamelCase = [] while i * i <= n: if n % i: i += 1 else: ...
386
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerat...
73
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """junnyu/roformer_chinese_small""": """https://huggingface.co/junnyu/rofo...
715
import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def __lowerCAmelCase ( ...
286
0
import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepie...
583
import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor snake_case__ = logging.get_logger(__name__) class UpperCAmelCase ( __lowerCamelCase ): def __init__( self : Optional[Any] , *lowerCAmelCase : ...
583
1
from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def UpperCAmelCase ( lowercase__ : int ): '''simple docstring''' a__ = int(number**0.5 ) return number == sq * sq def UpperCAmelCase ( lowercase__ : in...
412
import operator as op def UpperCAmelCase ( lowercase__ : str ): '''simple docstring''' a__ = [] a__ = lambda lowercase__ , lowercase__ : int(x / y ) # noqa: E731 integer division operation a__ = { """^""": op.pow, ...
412
1
def __snake_case ( __UpperCamelCase : str ,__UpperCamelCase : int ): """simple docstring""" return [sentence[i : i + ngram_size] for i in range(len(__UpperCamelCase ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod(...
86
class _a : """simple docstring""" def __init__( self : Union[str, Any] , UpperCAmelCase : int , UpperCAmelCase : Any , UpperCAmelCase : Dict ): A_ = None A_ = None A_ ...
86
1
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import...
506
'''simple docstring''' import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": A__: List[str] = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_path'''...
506
1
'''simple docstring''' import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import t...
374
'''simple docstring''' import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, ...
374
1
"""simple docstring""" import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowerCAmelCase__ ( A_ ): def __init__( self : Tuple , _lowerCamelCa...
430
"""simple docstring""" def _UpperCAmelCase ( __lowerCamelCase : int ) -> str: _snake_case = int(__lowerCamelCase ) if decimal in (0, 1): # Exit cases for the recursion return str(__lowerCamelCase ) _snake_case , _snake_case = divmod(__lowerCamelCase , ...
430
1
from __future__ import annotations import math class A__ : def __init__( self , __magic_name__ ): lowerCamelCase : str = size # approximate the overall size of segment tree with given value lowerCamelCase : Dict = [0 for i in ...
681
import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _a ( lowerCamelCase ): # vision encoder if "img_encoder.pos_embed" in name: lowerCamelCase : Tuple = name.replace("""img_encoder.pos_...
681
1
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase__ : List[str] =logging.get_lo...
711
import argparse import torch from transformers import ( UniSpeechSatConfig, UniSpeechSatForAudioFrameClassification, UniSpeechSatForSequenceClassification, UniSpeechSatForXVector, WavaVecaFeatureExtractor, logging, ) logging.set_verbosity_info() UpperCAmelCase__ : int =l...
269
0
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blen...
80
'''simple docstring''' import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase : Tuple = "src/transformers" # Th...
405
0
import os import sys __SCREAMING_SNAKE_CASE : int = os.path.join(os.path.dirname(__file__), 'src') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassi...
580
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ...
580
1
"""simple docstring""" def lowercase ( __UpperCamelCase = 50 ) -> int: __magic_name__ = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_le...
490
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_chan...
314
0
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu snake_case_ : Union[str, Any] ...
191
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_...
191
1
'''simple docstring''' from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsComma...
28
'''simple docstring''' import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformer...
28
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class lowerCamelCase ( _a ): UpperCAmelCase__ : Optional[Any] = (PNDMScheduler,) UpperCAmelCase__ : Optional[int] = (("""num...
720
import os import unittest from transformers.models.phobert.tokenization_phobert import VOCAB_FILES_NAMES, PhobertTokenizer from ...test_tokenization_common import TokenizerTesterMixin class lowerCamelCase ( A_ , unittest.TestCase ): UpperCAmelCase__ : str ...
294
0
"""simple docstring""" from math import isclose, sqrt def __UpperCAmelCase ( lowercase ,lowercase ,lowercase ): """simple docstring""" _UpperCAmelCase = point_y / 4 / point_x _UpperCAmelCase = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) _UpperCAm...
277
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class a ( lowerCAmelCase_ ): _snake_case : Dict = CustomTokenizer pass
277
1
'''simple docstring''' def __lowercase (_SCREAMING_SNAKE_CASE :list ): def merge(_SCREAMING_SNAKE_CASE :list , _SCREAMING_SNAKE_CASE :list ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop...
355
'''simple docstring''' import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def __lowercase (_SCREAMING_SNAKE_CASE :List[Any] ): return x + 2 class a__ ( unittest.TestCase ): def lowercase__ ...
355
1
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCAmelCase__ ( _a : str , _a : List[str] , _a : int ): snake_case...
568
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFl...
568
1
import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONA...
712
'''simple docstring''' from __future__ import annotations from math import ceil, floor, sqrt def UpperCAmelCase_ ( lowerCamelCase_ = 2_0_0_0_0_0_0 ): """simple docstring""" lowerCAmelCase__ : list[int] = [0] lowerCAmelCase__ : int for idx in range(1 , ceil(sqrt(targ...
568
0
def __lowercase ( __lowerCAmelCase : List[str] , __lowerCAmelCase : int , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[int] ): global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1...
335
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : List[str] = { '''unc-nlp/lxmert-base-uncased''': '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json''', } class ...
335
1
'''simple docstring''' import logging import os import threading import time try: import warnings except ImportError: _lowerCAmelCase = None try: import msvcrt except ImportError: _lowerCAmelCase = None try: import fcntl except ImportError: _...
718
'''simple docstring''' def _lowerCAmelCase ( lowercase : int ) ->int: """simple docstring""" if not isinstance(lowercase , lowercase ): raise ValueError('''multiplicative_persistence() only accepts integral values''' ) if num < 0: ...
318
0
"""simple docstring""" import logging import numpy as np import pytest from scipy.linalg import eigh logging.basicConfig(level=logging.INFO, format="%(message)s") def __snake_case ( SCREAMING_SNAKE_CASE__ : np.ndarray ) -> np.ndarray: '''simple docstring''' return input_arr...
289
"""simple docstring""" import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append(".") def __snake_case ( SCREAMING_SNAKE_CASE__ : Optional[int] ) -> List[str]: '''simple ...
289
1
import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_comm...
298
from __future__ import annotations def snake_case_ ( SCREAMING_SNAKE_CASE_ ) -> list[int]: lowercase__ : List[str] = [True] * limit lowercase__ : Union[str, Any] = False lowercase__ : List[str] = False ...
298
1
"""simple docstring""" import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") __SCREAMING_SNAKE_CASE ="https://www.google.com/search?q=" + " ".join(sys.argv[1:]) __SCREAMI...
425
"""simple docstring""" from __future__ import annotations def lowercase__( __SCREAMING_SNAKE_CASE : float , __SCREAMING_SNAKE_CASE : float , __SCREAMING_SNAKE_CASE : float , ): if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('You cannot...
425
1
"""simple docstring""" from collections import defaultdict def A_ ( UpperCAmelCase__ ) -> int: a : Any = 1 a : Optional[Any] = True for v in tree[start]: if v not in visited: ret += dfs(UpperCAmelCase__ ) if ret % 2 == ...
509
"""simple docstring""" import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from fl...
509
1
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDepende...
286
'''simple docstring''' import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transforme...
286
1
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class __lowerCAmelCase ( lowerCAmelCase_ ): """...
284
import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets __lowerCAmelCase : Tuple = "\n@inproceedings{xu-etal-2016-optimizing,\n title = {Optimizing Statistical Machine Translation for Text Simplification}...
284
1
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsm...
257
import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() __lowerCAmelCase = logging.g...
684
0
"""simple docstring""" from __future__ import annotations a_ = list[list[int]] # assigning initial values to the grid a_ = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, ...
714
"""simple docstring""" def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : List[Any] = len(__UpperCamelCase ) for i in range(length - 1 ): __lowercase : Optional[Any] = i for k in range(i + 1 , __UpperCamelCase ): if col...
523
0
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_availa...
108
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ : Optional[Any] = logging.get_logger(__name__) lowercase_ : Optional[int] = { '''microsoft/trocr-base-handwritten''': ( '''https://huggingface.co/micr...
588
0
'''simple docstring''' from __future__ import annotations class UpperCAmelCase : '''simple docstring''' def __init__( self , lowercase__ ) -> Tuple: SCREAMING_SNAKE_CASE : Dict = TypeError( 'Matrices...
179
'''simple docstring''' from __future__ import annotations def __lowerCAmelCase ( a_ , a_ = None ) -> list[list[str]]: '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = word_bank or [] # create a table ...
179
1
from ....configuration_utils import PretrainedConfig from ....utils import logging a : Union[str, Any] = logging.get_logger(__name__) a : Any = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/mai...
63
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDe...
63
1
'''simple docstring''' import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils...
718
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer A_ : Any = logging.get_logger(__nam...
419
0
class a : """simple docstring""" def __init__( self : Optional[Any] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = {} def UpperCAmelCase ( self : str ) -> None: print(self.vertex ) fo...
63
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from tra...
176
0
"""simple docstring""" import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline fro...
710
"""simple docstring""" import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.m...
600
0
"""simple docstring""" import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def SCREAMING_SNAKE_CASE__ ( snake_case : Any , snake_case : str , snake_case : ...
438
'''simple docstring''' import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging a_ ...
296
0
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, ...
712
from __future__ import annotations import numpy as np def lowerCamelCase_ ( UpperCamelCase__ : np.ndarray ) -> tuple[np.ndarray, np.ndarray]: """simple docstring""" __lowerCamelCase , __lowerCamelCase = np.shape(UpperCamelCase__ ) if rows !...
167
0
"""simple docstring""" import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class a_ ( UpperCAmelCase_ ): UpperCamelCase_ : Optional[Any] = (EulerDiscreteScheduler,) UpperCamelCase_...
644
'''simple docstring''' import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from t...
51
0
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( snake_case : int = 4_000_000 ) -> int: """simple docstring""" a : List[Any] = [] a , a : Tuple = 0, 1 while b <= n: if b % 2 == 0: even_...
610
'''simple docstring''' import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets UpperCamelCase : int = """\ @inproceedings{kakwani2020indicnlpsuite, title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Mu...
610
1
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.ut...
219
'''simple docstring''' import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import Backbone...
418
0
'''simple docstring''' import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets A = '\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saura...
449
'''simple docstring''' import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency A = { 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03...
449
1
"""simple docstring""" from __future__ import annotations import numpy as np def a ( __snake_case : list[float] ): '''simple docstring''' return np.maximum(0, __snake_case ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
608
"""simple docstring""" __lowerCamelCase = { "A": ["B", "C", "E"], "B": ["A", "D", "E"], "C": ["A", "F", "G"], "D": ["B"], "E": ["A", "B", "D"], "F": ["C"], "G": ["C"], } def a ( __snake_case : dict, __snake_case : str, __snake_case : Unio...
608
1
"""simple docstring""" from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class UpperCamelCase_ ( __SCREAMING_SNAKE_CASE): """simple docstring""" snake_case__ : str = "EncodecFeatureEx...
705
"""simple docstring""" import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_ut...
553
0
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=a_ ) class __SCREAMING_SNAKE_CASE( a_ ): # `task` is not a ClassVar since we want it to be part of the `asdi...
328
import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class __SCREAMING_SNAKE_CASE( a_...
328
1
import logging from transformers.configuration_utils import PretrainedConfig lowercase : str = logging.getLogger(__name__) class a__ ( lowerCAmelCase__ ): _A = "masked_bert" def __init__( self : Any , A_ : int=3...
712
from collections import defaultdict from math import ceil, sqrt def UpperCAmelCase_ ( _UpperCAmelCase = 1_0_0_0_0_0_0 , _UpperCAmelCase = 1_0 ): lowerCamelCase_: defaultdict = defaultdict(_UpperCAmelCase ) for outer_width in range(3 , (t_limit // 4) + 2 ...
584
0
import tensorflow as tf from ...tf_utils import shape_list class A__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase=1 , lowercase=False , **lowercase) ->...
302
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils...
300
0
"""simple docstring""" import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import Tokeniz...
480
"""simple docstring""" import unittest import torch from torch import nn from diffusers.models.activations import get_activation class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def lowerCamelCase(self ): A_ : Optional[int] = get_activa...
480
1
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer,...
53
import json import logging import math import os import sys from dataclasses import dataclass, field from typing import Optional from datasets import Dataset, load_dataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_MASKED_LM_MAPPING, AutoConfig, AutoModelForMaskedLM...
53
1
from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging A = logging.get_logger(__name__) def __UpperCAmelCase ( __A ) -> List[int]: '''simple docstring''' if isinstance(lowerCamelCa...
710
from __future__ import annotations class lowercase__ : def __init__( self : int , _lowercase : list[list[int]] ): """simple docstring""" UpperCAmelCase__ = TypeError( "Matrices must be formed ...
277
0
"""simple docstring""" import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tok...
473
'''simple docstring''' # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[de...
536
0
'''simple docstring''' import math import sys import cva import numpy as np def _UpperCAmelCase ( _UpperCamelCase : np.ndarray, _UpperCamelCase : float ) -> np.ndarray: # For applying gaussian function for each element in matrix. A_ = math.sqrt(_UpperC...
174
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common ...
174
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE_ = {'configuration_plbart': ['PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PLBartConfig']} ...
300
from __future__ import annotations def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: int , lowerCAmelCase: int ) -> list[list[int]]: _UpperCAmelCase : list[list[int]] = [] create_all_state(1 , lowerCAmelCase , lowerCAmelCase , [] , lowerCAmelCase ) return result ...
300
1
from itertools import product def lowerCamelCase ( UpperCamelCase : int , UpperCamelCase : int ) -> list[int]: _lowerCamelCase = sides_number _lowerCamelCase = max_face_number * dice_number _lowerCamelCase = [0] * ...
234
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def lowerCamelCase ( ) -> tuple[list[int], int]: _lowerCamelCase = [randint(-10_00 , 10_00 ) for i in range(10 )] _lowerCamelCase ...
234
1
import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common i...
66
import random def lowerCamelCase_ ( UpperCamelCase__ : list, UpperCamelCase__ : List[Any] ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = [], [], [] for element in data: i...
240
0
"""simple docstring""" def lowercase__ ( lowerCAmelCase__ : int = 1_0_0_0_0_0_0 ) -> int: '''simple docstring''' a__ : Optional[Any] = set(range(3 , lowerCAmelCase__ , 2 ) ) primes.add(2 ) for p in range(3 , lowerCAmelCase__ , 2 ): if ...
714
"""simple docstring""" import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVeca...
251
0
def lowerCamelCase__ ( _lowercase = 10 , _lowercase = 22 ): '''simple docstring''' UpperCAmelCase_ : Tuple = range(1 , _lowercase ) UpperCAmelCase_ : Optional[int] = range(1 , _lowercase ) return sum( 1 for power in powers for base in base...
30
'''simple docstring''' from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_ima...
474
0
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class UpperCamelCase_ ( datasets.BeamBasedBuilder ): def lowerCAmelCase ( self ) -> Dict: ...
541
from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get...
541
1
import argparse import gc import json import os import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator...
298
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration A_: Union[str, Any] = 5_0000 A_: str = 5000 A_ , A_: int = os.path.split(__file__) A_: str = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py',...
398
0
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2...
74
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common imp...
74
1
"""simple docstring""" import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig UpperCamelCase__ = logging.get_logger(__name__) class a__ : def __init_...
227
"""simple docstring""" import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # n...
227
1
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common ...
717
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_t...
5
0
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowercase_ : Optional[Any] = False class _lowerCamelCase ( unittest.TestCase ...
64
'''simple docstring''' from functools import reduce __lowerCamelCase = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290...
467
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import...
708
'''simple docstring''' import argparse import collections import json import os import re import string import sys import numpy as np lowerCamelCase_ : Union[str, Any] = re.compile(r'''\b(a|an|the)\b''', re.UNICODE) lowerCamelCase_ : int = None def __magic_name__( ): ...
265
0
"""simple docstring""" from random import shuffle import tensorflow as tf from numpy import array def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : List[Any] ): lowerCAmelCase = int(__UpperCAmelCase ) assert noofclusters < len(__UpperCAmelCase ) # Find ...
4
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class _lowerCamelCase ( unittest.TestCase ): def _lowerCAmelCase ( self : Optiona...
299
0
from __future__ import annotations class UpperCamelCase: def __init__( self : Tuple , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : str ) -> Union[str, Any]: '''simple docstring''' __snake_case , __snake_case =...
473
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...
473
1
import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import Mode...
455
'''simple docstring''' import os from typing import List, Optional, Union from ...tokenization_utils import PreTrainedTokenizer from ...tokenization_utils_base import AddedToken from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase...
90
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( a : list ) ->list: if len(a ) == 0: return [] snake_case , snake_case = min(a ), max(a ) snake_case = int(max_value - min_value ) + 1 snake_case = [[] fo...
44
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassificatio...
44
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Union[str, Any] = { 'configuration_git': ['GIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GitConfig', 'GitVisionConfig'], 'processing_...
3
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_...
463
0
import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class UpperCAmelCase__ ( snake_case ): """simple docstring""" lowerCAmelCase__ : Union[str, Any] = (DDPMScheduler,) def _UpperCAmelCase ( self: Dict , **__lowerCAmel...
286
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/spiece.model""") @require_sente...
286
1
from __future__ import annotations from typing import TypedDict class _snake_case ( UpperCAmelCase_ ): __lowerCAmelCase : str __lowerCAmelCase : int def UpperCamelCase ( lowercase_ ) -> list[str]: '''simple docstring''' if not isinstance(lowercase_ ...
12
'''simple docstring''' import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse('9.1.0'): __magic_name__ : str ={ 'linear': PIL.Image.Resampling.BILINEAR, 'bilinear': PIL.Ima...
664
0
from ..utils import DummyObject, requires_backends class lowerCamelCase (metaclass=__lowerCamelCase ): """simple docstring""" UpperCAmelCase_ = ["torch", "torchsde"] def __init__( self : Optional[int], *_UpperCAmelCa...
157
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase : Union[str, Any] = { '''configuration_x_clip''': [ '''XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XCLIPConfig''', ...
157
1
'''simple docstring''' from __future__ import annotations UpperCamelCase : Optional[int] = 10 def A__ ( __lowerCAmelCase : list[int] ): lowerCamelCase__ = 1 lowerCamelCase__ = max(__lowerCAmelCase ) while placement <= max_digit: ...
50
from abc import ABC, abstractmethod from typing import List, Optional class snake_case_ (lowerCamelCase_ ): def __init__( self :Optional[Any] ) -> Dict: # test for the above condition self.test() def lowerCamelCase__( self :Tuple ) -> int: ...
335
0
"""simple docstring""" from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): imp...
494
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { '''microsoft/trocr-base-handwritten''': ( '''https://huggingface.co/microsoft/trocr-base-handwritten/resol...
494
1
"""simple docstring""" import copy import random from transformers import CLIPTokenizer class UpperCAmelCase_ ( snake_case ): def __init__( self , *UpperCamelCase_ , **UpperCamelCase_ ) -> List[Any]: super().__init__(*UpperCamelCase_ , ...
76
'''simple docstring''' import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_...
128
0
"""simple docstring""" import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transforme...
701
"""simple docstring""" import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class _lowerCame...
507
0
import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency A_ = { """E""": 1_2.7_0, """T""": 9.06, """A""": 8.17, """O""": 7.51, """I""": 6.97, """N""": 6.75, """S""": 6.33, """H""": 6.09, """R""": 5.99, """D""": 4.25, """L""": 4.03, ...
393
'''simple docstring''' import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVeca...
501
0
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimeste...
707
def _lowerCAmelCase ( __lowerCAmelCase ) -> float: """simple docstring""" if edge <= 0 or not isinstance(__lowerCAmelCase , __lowerCAmelCase ): raise ValueError('''Length must be a positive.''' ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) *...
219
0
import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( 'The ...
101
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ : Tuple = logging.get_logger(__name__) UpperCamelCase__ : Option...
105
0
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def _UpperCAmelCase ( a : Tuple ): snake_case__ = args.pruning_method snake_case__ = args.threshold snake_case__ = args.mo...
99
from collections.abc import Callable def _UpperCAmelCase ( a : Callable[[float], float] , a : float , a : float ): snake_case__ = a snake_case__ = b if function(a ) == 0: # one of the a or b is a root for the function return a ...
99
1
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""vocab_file""": """vocab.json"""} lowercase__ = ...
630
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-singl...
630
1
"""simple docstring""" import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, Wa...
701
"""simple docstring""" def _A ( _a : int | float | str ): """simple docstring""" try: A = float(_a ) except ValueError: raise ValueError("""Please enter a valid number""" ) A = decimal - ...
255
0
"""simple docstring""" import os from argparse import ArgumentParser from typing import List import torch.utils.data from datasets import Dataset, IterableDataset from datasets.distributed import split_dataset_by_node snake_case_ : Optional[int] = 4 snake_case_ : int = 3 class snak...
595
"""simple docstring""" import logging import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_...
595
1
'''simple docstring''' from typing import Union import fire import torch from tqdm import tqdm def __A ( a_ : str ,a_ : str = "cpu" ,a_ : Union[str, None] = None ): lowerCAmelCase : Optional[int] = torch.load(a_ ,map_location=a_ ) for...
551
'''simple docstring''' from __future__ import annotations lowerCAmelCase = [] def __A ( a_ : list[list[int]] ,a_ : int ,a_ : int ): for i in range(len(a_ ) ): if board[row][i] == 1: return False for i in range(len(a_ ) ): if b...
551
1