code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from abc import ABC, abstractmethod from argparse import ArgumentParser class snake_case__ ( lowerCAmelCase_ ): """simple docstring""" @staticmethod @abstractmethod def lowercase_ ( _snake_case : ArgumentParser ) ->Optional[Any]: raise NotImplementedEr...
478
import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class snake_case__ : """simple docstring""" _SCREAMING_SNAKE_CASE = None def lowercase_ ( self : Optional[int] ) ->Optional[int]: ...
478
1
"""simple docstring""" # HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's ...
112
"""simple docstring""" def __lowercase ( lowerCamelCase_ : Tuple ): SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = { "^": 3, "*": 2, "/": 2, "%": 2, "+": 1, "-...
112
1
"""simple docstring""" import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class A__ ( unittest.TestCase): """simple docstring""" ...
222
"""simple docstring""" from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class a ( __snake_case ): ...
549
0
"""simple docstring""" class lowerCAmelCase : def __init__( self , a__ ): _UpperCAmelCase = len(a__ ) _UpperCAmelCase = [0] * len_array if len_array > 0: _UpperCAmelCase = array[0] ...
494
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available from .timesteps import ( fastaa_timesteps, smartaa_timestep...
494
1
"""simple docstring""" import sys __A = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "668966489504452...
134
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch...
134
1
from math import factorial def __A(lowerCAmelCase = 1_0_0 ) -> int: """simple docstring""" return sum(map(lowerCAmelCase , str(factorial(lowerCAmelCase ) ) ) ) if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
202
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_...
202
1
import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAdde...
145
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator...
145
1
"""simple docstring""" import json import logging import os import socket import git import numpy as np import torch logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - PID: %(process)d - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO, ) SCREAMING_SNAKE_CASE ...
283
"""simple docstring""" import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import lo...
283
1
from ....configuration_utils import PretrainedConfig from ....utils import logging __a : Union[str, Any] = logging.get_logger(__name__) # TODO: upload to AWS __a : Dict = { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolv...
637
import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline _UpperCAmelCase = { """n_samples""": 64, """horizon""": 32, """num_inference_steps""": 20, """n_guide_steps""": 2, # can set to 0 for faster sampling, does not use value network...
558
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel...
720
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotSmallConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_...
417
0
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization...
66
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOn...
66
1
'''simple docstring''' def __lowerCamelCase ( __lowerCAmelCase : int , __lowerCAmelCase : float , __lowerCAmelCase : float ) -> float: return round(float(moles / volume ) * nfactor ) def __lowerCamelCase ( __lowerCAmelCase : ...
709
'''simple docstring''' from __future__ import annotations def __lowerCamelCase ( __lowerCAmelCase : list ) -> list: if len(__lowerCAmelCase ) == 0: return [] snake_case , snake_case = min(__lowerCAmelCase ), max(__lowerCAmelCase ) sn...
517
0
'''simple docstring''' import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_util...
692
'''simple docstring''' import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a =(CMStochasticIterativeScheduler,) __a =10 def...
692
1
"""simple docstring""" import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def UpperCamelCase ( _lowerCAmelCase : List[Any] ): __a = [ """decoder.version""", """decoder.output_p...
706
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING __A = logging.get_logger(__name__) ...
173
0
def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): return int((input_a, input_a).count(0 ) == 0 ) def lowerCAmelCase_ ( ): assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , ...
81
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicab...
333
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A : List[str] = logging.get_logger(__name__) A : List[Any] = { 'facebook/vit-mae-base': 'https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json'...
700
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils impor...
304
0
'''simple docstring''' import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRober...
51
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer lowercase_ = logging.get_logger(__name__) ...
235
0
'''simple docstring''' import argparse import collections import torch from flax import traverse_util from tax import checkpoints from transformers import TaConfig, TaEncoderModel, TaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def __low...
179
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_atten...
179
1
import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_byt...
86
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __snake_case ( __UpperCamelCase : str ,__UpperCamelCase : str ,__UpperCamelCase : List[str] ): """simple docstring""" A_ = { "en": "Machine learning is great, isn't i...
86
1
import requests from bsa import BeautifulSoup def lowerCAmelCase_ ( lowerCamelCase = "AAPL" ): __magic_name__ : Any =F"https://in.finance.yahoo.com/quote/{symbol}?s={symbol}" __magic_name__ : List[str] =BeautifulSoup(requests.get(lowerCamelCase ).text , """html.pa...
367
from ... import PretrainedConfig UpperCAmelCase_ : List[str] = { "sijunhe/nezha-cn-base": "https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json", } class __A ( UpperCamelCase__ ): UpperCamelCase = NEZHA_PRETRAINED_CONFIG_ARCH...
367
1
'''simple docstring''' import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision...
78
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class lowerCAmelCase ( __a , unittest.TestCase ): '''simple docstring''' _A : Un...
149
0
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_C...
715
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs ...
647
0
import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDataset...
14
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_s...
682
0
'''simple docstring''' import numpy as np from PIL import Image def __A ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = np.array(lowerCamelCase_ ) if arr.shape[0] != arr.shape[1]: raise ValueError("""The...
703
'''simple docstring''' from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class UpperCamelCase__ : """simple docstring""" SC...
79
0
"""simple docstring""" from __future__ import annotations class lowerCAmelCase__ : """simple docstring""" def __init__( self : Any , lowercase__ : str , lowercase__ : str ): __lowercase : Tuple = text, pattern ...
575
"""simple docstring""" import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import ja...
123
0
'''simple docstring''' def _lowerCamelCase( UpperCamelCase__ : int = 4_000_000 ) -> int: A : Dict = [0, 1] A : Any = 0 while fib[i] <= n: fib.append(fib[i] + fib[i + 1] ) if fib[i + 2] > n: break i +...
537
'''simple docstring''' import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTenso...
537
1
import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor lowercase_: str = logging.get_logger(__name__) class lowercase__ (__snake_case ): """simple docstring""" def __init__( self : in...
648
from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceCla...
648
1
from graphs.minimum_spanning_tree_kruskal import kruskal def __lowerCamelCase ( ) -> Tuple: _UpperCAmelCase = 9 _UpperCAmelCase = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4]...
129
import sys def __lowerCamelCase ( _lowerCAmelCase ) -> Tuple: _UpperCAmelCase = len(_lowerCAmelCase ) _UpperCAmelCase = [[0 for x in range(_lowerCAmelCase )] for x in range(_lowerCAmelCase )] _UpperCAmelCase = [[0 for x in range(_lowerCAmelCase )] for x in range(_lowerCAmelCase ...
129
1
def __A ( _lowercase ): '''simple docstring''' _A = len(_lowercase ) for i in range(_lowercase ): for j in range(i + 1 , _lowercase ): if numbers[j] < numbers[i]: _A ,_A = numbers[j], numbers[...
484
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if ...
484
1
import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration SCREAMING_SNAKE_CASE_ = [ # tf -> hf ('/', '.'), ('layer_', 'layers.'), ('kernel', 'weight'), ('b...
467
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput SCREAMING_SNAKE_CASE_ = 'scheduler_config.json' class a ( UpperCAmelCase ): ...
467
1
from math import factorial def UpperCamelCase( __UpperCamelCase : int = 20 ): lowerCAmelCase_ : List[str] = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... lowerCAmelCase_ : List[Any] = n // 2 return int(facto...
171
'''simple docstring''' def A_( A : list[int]): UpperCamelCase = [] if len(A) == 1: return [nums.copy()] for _ in range(len(A)): UpperCamelCase = nums.pop(0) UpperCamelCase = permute(A) for perm ...
3
0
"""simple docstring""" from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url ...
719
"""simple docstring""" def snake_case__ ( _snake_case : str ): """simple docstring""" UpperCamelCase__ = 0 # if input_string is "aba" than new_input_string become "a|b|a" UpperCamelCase__ = "" UpperCamelCase__ = "" # app...
304
0
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : str , __UpperCamelCase : str ): '''simple docstring''' snake_case_ : Optional[int] = len(__UpperCamelCase ) snake_case_ : List[Any] ...
58
_lowercase = '0.18.2' from .configuration_utils import ConfigMixin from .utils import ( OptionalDependencyNotAvailable, is_flax_available, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, ...
306
0
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _snake_case (__lowercase , __lowercase , __lowercase): UpperCamelCase...
705
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging ...
618
0
"""simple docstring""" import math def _snake_case ( lowercase__ ): assert isinstance(__A , __A ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return T...
630
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from tra...
607
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import Au...
172
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow lowerCAmelCase__ = logging.getLogger() @unittest.s...
172
1
import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, ...
668
def lowerCAmelCase_ (lowercase__ : float , lowercase__ : int ) -> float: '''simple docstring''' if digit_amount > 0: return round(number - int(lowercase__ ) , lowercase__ ) return number - int(lowercase__ ) if __name_...
668
1
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch...
399
'''simple docstring''' import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTeste...
399
1
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_co...
0
"""simple docstring""" from math import factorial class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : str , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int ) -> Optional[int]: ...
580
0
from maths.prime_check import is_prime def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ): lowerCAmelCase : str = F"""Input value of [...
693
from ..utils import DummyObject, requires_backends class _a ( metaclass=snake_case_ ): _UpperCamelCase: List[Any] = ["keras_nlp"] def __init__( self , *lowercase_ , **lowercase_ ) -> Tuple: requires_backends(self , ...
693
1
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation Uppe...
120
'''simple docstring''' import mpmath # for roots of unity import numpy as np class a : '''simple docstring''' def __init__( self , lowerCamelCase_=None , lowerCamelCase_=None ) -> Tuple: # Input as list _a : Optional[int] = list(poly_a or [0] ...
120
1
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore A__: Tuple = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" A__: Tuple = [file for file in filep...
221
def lowerCAmelCase_ ( A_): UpperCamelCase__: Union[str, Any] = "" for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def lowerCAmelCase_ ( A_): UpperCamelCase__...
221
1
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, Da...
411
from __future__ import annotations from collections.abc import MutableSequence class _lowerCAmelCase : '''simple docstring''' def __init__( self : Optional[int] , UpperCamelCase : int , UpperCamelCase : MutableSequence[float] ): '''simple ...
411
1
"""simple docstring""" from __future__ import annotations __A : Optional[Any] = [] def lowercase ( _SCREAMING_SNAKE_CASE : list[list[int]] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ): '''si...
95
"""simple docstring""" import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, s...
95
1
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_avail...
631
from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time SCREAMING_SNAKE_CASE__ = Lock() def lowercase ( a , a , a , a , a , a , a ): '''simple docstring''' global process_lock # we perfor...
631
1
import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, r...
322
from __future__ import annotations def lowerCamelCase_ ( _a : int | float | str , _a : int | float | str ): '''simple docstring''' if nth_term == "": return [""] UpperCAmelCase_ : Tuple = int(_a ) UpperCAmelCase_ : Opti...
322
1
from __future__ import annotations from collections.abc import Callable _SCREAMING_SNAKE_CASE = list[list[float | int]] def snake_case ( snake_case__ :Tuple , snake_case__ :List[Any]) -> Matrix: _A = len(snake_case__) _A = [[0 for _ in rang...
401
import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_avail...
315
0
def lowerCAmelCase_ ( ): '''simple docstring''' __lowerCamelCase : int =0 for i in range(1 , 1001 ): total += i**i return str(SCREAMING_SNAKE_CASE )[-10:] if __name__ == "__main__": print(s...
721
"""simple docstring""" import math from datetime import datetime, timedelta def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE : int ): '''simple docstring''' __lowerCamelCase : Any =year % 19 __lowerCamelCase ...
363
0
'''simple docstring''' import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from...
577
'''simple docstring''' def _snake_case ( A_ : list ): """simple docstring""" for i in range(len(A_ ) - 1 , 0 , -1 ): a_ : List[str] = False for j in range(A_ , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: a_ , a_ : L...
577
1
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __snake_case ( UpperCamelCase_ ): '''simple docstring''' lowerCAmelCase__ = ["""image_processor""", """tokenizer"""] lowerCAmelCase__ = """AutoIma...
712
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: ...
155
0
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin...
19
"""simple docstring""" from collections.abc import Callable import numpy as np def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case, __snake_case ) -> np.array: """simple docstring""" _UpperCamelCase = int(np.cei...
19
1
'''simple docstring''' class _lowerCAmelCase : '''simple docstring''' def __init__( self : Dict ) -> Any: '''simple docstring''' _lowercase : List[Any] = 0 _lowercase : str = 0 ...
411
'''simple docstring''' import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic...
411
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series impor...
315
from __future__ import annotations from collections.abc import Callable _A : Tuple = list[list[float | int]] def _a ( UpperCAmelCase , UpperCAmelCase ) -> Matrix: """simple docstring""" lowerCamelCase__ : int = len(UpperCAmelCase ) lowe...
315
1
"""simple docstring""" import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( __SCREAMING_SNAKE_CASE ): A__ = (KDPMaDiscreteScheduler,) A__ = 10 d...
701
"""simple docstring""" import operator as op def _lowerCAmelCase ( lowerCamelCase__ : Tuple ) -> List[str]: _SCREAMING_SNAKE_CASE : Optional[int] = [] _SCREAMING_SNAKE_CASE : str = lambda lowerCamelCase__, lowerCamelCase__ : int(x / ...
295
0
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_ver...
612
from abc import ABC, abstractmethod from argparse import ArgumentParser class lowerCAmelCase__ ( __lowercase ): @staticmethod @abstractmethod def A_ ( a ) -> Union[str, Any]: '''simple docstring''' raise NotImplementedError() ...
612
1
def _lowerCAmelCase ( __lowerCamelCase : int ) -> Tuple: """simple docstring""" if n == 1 or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): return 0 elif n == 2: return 1 else: __SCREAMING_SNAKE_CASE : Optional[int] = [0, 1] for i in range(2 , n...
702
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput...
447
0
from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): _SCREAMING_SNAKE_CASE : Dict = cva....
249
import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): _SCREAMING_SNAKE_CASE : Tuple = AutoConfig.from_pretrained(__lowerC...
249
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, ...
712
"""simple docstring""" import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import Tok...
598
0
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def A__ ( snake_case_ : str , snake_case_ : str , **snake_case_ : Optional[int] ): SCREAMING_SNAKE_CASE__: Any= AutoConfig.from_pretrained(snake_case_ , **snake_case_ ...
64
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.u...
252
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a__: Tuple = logging.get_logger(__name__) a__: Tuple ...
700
def UpperCamelCase__( UpperCamelCase__ : int )->list: A__ = int(UpperCamelCase__ ) if n_element < 1: A__ = ValueError('''a should be a positive number''' ) raise my_error A__ = [1] A__ , A__ ...
212
0
import heapq import sys import numpy as np UpperCAmelCase_ = tuple[int, int] class __UpperCamelCase : def __init__( self ): _UpperCAmelCase = [] _UpperCAmelCase = set() def UpperCamelCase( self ): if not se...
32
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_uti...
578
0
"""simple docstring""" def snake_case ( A__ ,A__ ): UpperCAmelCase_ : Optional[Any] = len(A__ ) + 1 UpperCAmelCase_ : Union[str, Any] = len(A__ ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches w...
706
"""simple docstring""" from itertools import count def snake_case ( A__ = 50 ): UpperCAmelCase_ : Any = [1] * min_block_length for n in count(A__ ): fill_count_functions.append(1 ) for block_length in range(A__ ,n + 1 ): for block_start in range(n - ...
463
0
"""simple docstring""" import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version UpperCAmelCase = version.parse(importlib_metadata.version("""nltk""")) if NLTK_VERSION >= version.Version("""3.6.4"""): from nltk import word...
535
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { """shi-labs/nat-m...
535
1
"""simple docstring""" import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfAr...
109
"""simple docstring""" import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers ...
109
1
from graphs.minimum_spanning_tree_kruskal import kruskal def _SCREAMING_SNAKE_CASE ( ) -> Any: _UpperCAmelCase = 9 _UpperCAmelCase = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], ...
518
import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MA...
518
1
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def SCREAMING_SNAKE_CASE__ ( lowercase ) -> None: snake_case : List[str] = analyze_text(lowercase ) snake_case : Optional[int] = list(""" ""...
709
import contextlib import importlib import io import unittest import transformers # Try to import everything from transformers to ensure every object can be loaded. from transformers import * # noqa F406 from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, require_tf, require_torch from...
684
0
"""simple docstring""" import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_s...
52
'''simple docstring''' from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class a_ ( _lowerCAmelCase ): def lowercase__ ( self : Tuple , lowercase : float )...
172
0
'''simple docstring''' import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _UpperCamelCase ( unitte...
454
'''simple docstring''' import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import...
454
1
'''simple docstring''' from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_p...
22
"""simple docstring""" import inspect import unittest from transformers import ViTMSNConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import Co...
473
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a__ = { '''configuration_transfo_xl''': ['''TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TransfoXLConfig'''], '''tokenization_tran...
566
'''simple docstring''' def snake_case__ ( a ) -> int: '''simple docstring''' if n == 1 or not isinstance(a , a ): return 0 elif n == 2: return 1 else: snake_case__ = [0, 1] for i in range(2 , n + 1 ): sequence.append(s...
566
1
'''simple docstring''' import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common i...
350
'''simple docstring''' import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class a_ ( snake_case ): UpperCAmelCase : str = (CMStochasticIterativeScheduler,) UpperCAmelCase : int ...
350
1
from __future__ import annotations import typing from collections import Counter def __lowerCamelCase ( _lowerCAmelCase ) -> Union[str, Any]: _UpperCAmelCase = Counter() for base in range(1 , max_perimeter + 1 ): for perpendicular in range(snake_case_ , max_perimeter + 1 ): ...
703
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCAmelCase = { "configuration_time_series_transformer": [ "TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "TimeSeriesTransformerConfig", ...
129
0
'''simple docstring''' import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_com...
452
'''simple docstring''' import argparse import json import re from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileNetVaConfig, MobileNetVaForImageClassification, MobileNetVaImageProcessor, load_tf_we...
452
1
from torch import nn def __lowercase ( _SCREAMING_SNAKE_CASE ) -> Optional[Any]: '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": ...
116
import argparse import torch from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_from_original_stable_diffusion_ckpt if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", ...
116
1
'''simple docstring''' SCREAMING_SNAKE_CASE = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) SCREAMING_SNAKE_CASE =...
94
"""simple docstring""" from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar a_ : Optional[Any] = TypeVar('''T''') class __lowercase( Generic[T] ): '''simple docstring''' __a : deque[T] # Cache store of keys ...
594
0
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_dataset...
181
import math def _lowercase ( SCREAMING_SNAKE_CASE_ : int ): """simple docstring""" UpperCamelCase = [True] * n UpperCamelCase = False UpperCamelCase = False UpperCamelCase = True for i in range(3...
181
1
"""simple docstring""" UpperCAmelCase : int = range(2, 20 + 1) UpperCAmelCase : Any = [10**k for k in range(ks[-1] + 1)] UpperCAmelCase : dict[int, dict[int, list[list[int]]]] = {} def lowerCamelCase ( _UpperCamelCase : List[Any] , _UpperCamelCase : ...
139
# Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def lowerCamelCase( a__ ,a__ ,a__ ,a__): _SCREAMING_SNAKE_CASE ={ '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', ...
691
0
'''simple docstring''' def UpperCAmelCase ( A : list ): if len(_lowerCamelCase ) <= 1: return lst SCREAMING_SNAKE_CASE : Tuple = 1 while i < len(_lowerCamelCase ): if lst[i - 1] <= lst[i]: ...
719
'''simple docstring''' import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester ...
464
0
"""simple docstring""" import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, ...
95
'''simple docstring''' from collections import Counter from timeit import timeit def _lowerCAmelCase ( lowercase = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(""" """ , """""" ).lower() ).values() ) < 2 def _lowerCAmelCase ( lowe...
689
0
"""simple docstring""" import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils i...
63
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, req...
63
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor _A = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE_ ( snake_case ): def __init__( self , *lowercase , **lowercase ) -...
158
'''simple docstring''' def A_ ( __SCREAMING_SNAKE_CASE : int ) -> bool: if num < 0: return False __SCREAMING_SNAKE_CASE : int = num __SCREAMING_SNAKE_CASE : int = 0 while num > 0: __SCREAMING_SNAKE_CASE : ...
158
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ : Dict = {"configuration_wavlm": ["WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "WavLMConfig"]} try: if not is_torch_available(): rai...
367
from ... import PretrainedConfig UpperCAmelCase_ : List[str] = { "sijunhe/nezha-cn-base": "https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json", } class __A ( UpperCamelCase__ ): UpperCamelCase = NEZHA_PRETRAINED_CONFIG_ARCH...
367
1
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.I...
61
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from tra...
517
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig ...
700
"""simple docstring""" import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn...
562
0
'''simple docstring''' def __snake_case( _lowerCAmelCase ) -> list: snake_case__ : Optional[int] = [0] * len(snake_case__ ) for i in range(1 , len(snake_case__ ) ): # use last results for better performance - dynamic programming ...
374
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def UpperCamelCase ( snake_case__ : np.ndarray , snake_case__ : np.ndarray ) -> float: return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(snake_case__ , snake_case_...
40
0
"""simple docstring""" import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline __A : List[Any] ...
700
"""simple docstring""" from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def lowercase ( ): lowercase_ : Union[str, Any] = { '''repo_name''': ['''test_repo1''', ...
141
0
lowerCamelCase__ = 8.3_14_45_98 def _lowerCamelCase( __snake_case , __snake_case ) -> float: if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) ...
524
import os import shutil import sys import tempfile import unittest from pathlib import Path import pytest import transformers from transformers import ( BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoTokenizer, BertConfig, BertTokenizer, BertTokenizerFast, CTRL...
524
1
from bisect import bisect from itertools import accumulate def __magic_name__ ( __a : str , __a : Tuple , __a : int , __a : str ): '''simple docstring''' UpperCamelCase__ = sorted(zip(__a , __a ) , key=lambda __a : x[0] / x[1] , reverse=__a ) UpperCamelCase__...
719
from __future__ import annotations from typing import TypedDict class __A( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 42 def __magic_name__ ( __a : str ): '''simple docstring''' if not...
86
0
'''simple docstring''' import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig A_ = logging.get_logger(__name__) class UpperCAmelCase : '''simple docstring''' def __init...
42
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTo...
577
0
'''simple docstring''' import importlib.metadata import operator import re import sys from typing import Optional from packaging import version __snake_case = { '''<''': operator.lt, '''<=''': operator.le, '''==''': operator.eq, '''!=''': operator.ne, '''>=''': operator.ge, '''>'...
280
'''simple docstring''' import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) __snake_case = logging.getLogger() def a ( __a ) ...
280
1
"""simple docstring""" import numpy as np from PIL import Image def _A ( __lowercase , __lowercase , __lowercase ): """simple docstring""" lowerCamelCase__ = np.array(__lowercase ) if arr.shape[0] != arr.shape[1]: raise...
129
"""simple docstring""" def _A ( __lowercase , __lowercase ): """simple docstring""" return numa ^ numa < 0 if __name__ == "__main__": import doctest doctest.testmod()
129
1
'''simple docstring''' import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProc...
347
'''simple docstring''' from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusio...
347
1
'''simple docstring''' def a ( _UpperCAmelCase ) -> bool: """simple docstring""" if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError('check_bouncy() accepts only integer arguments' ) a_ = str(_UpperCAmelCase ) a_ ...
697
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase =logging.get_logger(__name__) __lowerCAmelCase ={ "SCUT-DLVCLab/lilt-roberta-en-base": ( "https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json...
697
1
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTeste...
701
from graphs.minimum_spanning_tree_kruskal import kruskal def UpperCAmelCase_ ( ): lowerCamelCase_: str = 9 lowerCamelCase_: Tuple = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2]...
584
0
'''simple docstring''' def __magic_name__ ( __UpperCAmelCase ) -> list[int]: '''simple docstring''' snake_case_ = [0 for i in range(len(__UpperCAmelCase ) )] # initialize interval's left pointer and right pointer snake_case_ ,snake_case...
640
'''simple docstring''' import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import Tokenize...
640
1
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> bool: '''simple docstring''' return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(__UpperCAmelCase ) ) def __magic_name_...
638
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { 'configuration_trajectory_transformer': [ 'TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrajectoryTransformerConfig', ], } try: ...
638
1
from string import ascii_uppercase __lowerCamelCase : Optional[int] = {str(ord(c) - 55): c for c in ascii_uppercase} def SCREAMING_SNAKE_CASE ( snake_case_ : int , snake_case_ : int ): if isinstance(snake_case_ , snake_case_ ): raise TypeError("int() can't convert ...
297
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import...
297
1
'''simple docstring''' from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from .....
489
'''simple docstring''' from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import f...
489
1
'''simple docstring''' from __future__ import annotations lowercase__ : Dict = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] lowercase__ : Optional[int] = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def a__ ( lowercase : list[float] ...
98
'''simple docstring''' import numpy as np from cva import COLOR_BGR2GRAY, cvtColor, imread from numpy import array, uinta from PIL import Image from digital_image_processing import change_contrast as cc from digital_image_processing import convert_to_negative as cn from digital_image_pr...
207
0
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING a_ = logging.get_logger(__name__) a_ = { "SenseTime/deformable-detr": "https://huggingface.co/sensetime/defor...
621
"""simple docstring""" import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base impor...
621
1