code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transform...
45
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import D...
45
1
def A ( lowercase__ : int ) -> Optional[Any]: stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def A ( lowercase__ : Union[str, Any] , lowercase__ : Dict , lowercase__ : str ) -> List[str]: if i >= h: return # If first element is smaller than the last the...
45
import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class lowerCAmelCase_ ( lowercase ): """simple docstring""" de...
45
1
import pprint import requests UpperCamelCase = "https://zenquotes.io/api" def A ( ) -> list: return requests.get(API_ENDPOINT_URL + """/today""" ).json() def A ( ) -> list: return requests.get(API_ENDPOINT_URL + """/random""" ).json() if __name__ == "__main__": ...
45
def A ( lowercase__ : int ) -> bool: if num < 0: return False UpperCamelCase__ :int = num UpperCamelCase__ :int = 0 while num > 0: UpperCamelCase__ :Optional[int] = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main__":...
45
1
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO ) UpperCame...
45
from __future__ import annotations def A ( lowercase__ : list[int] ) -> bool: return len(set(lowercase__ ) ) == len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
45
1
def A ( lowercase__ : int , lowercase__ : float , lowercase__ : float ) -> float: return round(float(moles / volume ) * nfactor ) def A ( lowercase__ : float , lowercase__ : float , lowercase__ : float ) -> float: return round(float((moles * 0.0821 * temperature) / (...
45
from __future__ import annotations class lowerCAmelCase_ : """simple docstring""" def __init__( self :List[Any] , lowerCamelCase__ :int = 0 ): UpperCamelCase__ :List[str] = key def __a ( self :Optional[Any] , lowerCamelCase__...
45
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices Uppe...
45
import random def A ( lowercase__ : Dict , lowercase__ : str , lowercase__ : Optional[Any] ) -> int: UpperCamelCase__ :List[Any] = a[left_index] UpperCamelCase__ :Dict = left_index + 1 for j in range(left_index + 1 , lowercase__ ): if a[j] < pivot: UpperCamelC...
45
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase = { "configuration_xlm_roberta_xl": [ "XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMRobertaXLConfig", "XLMRobertaXLOnnxConfig", ]...
45
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "shi-labs/dinat-mini-in1k-224": "https:/...
45
1
from dataclasses import dataclass, field from typing import Optional @dataclass class lowerCAmelCase_ : """simple docstring""" _snake_case : Optional[str] = field( default="""codeparrot/codeparrot""" , metadata={"""help""": """Model name or path of mo...
45
def A ( lowercase__ : int , lowercase__ : int ) -> int: return int(input_a == input_a == 0 ) def A ( ) -> None: print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(f"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) p...
45
1
from collections.abc import Sequence def A ( lowercase__ : Sequence[int] | None = None ) -> int: if nums is None or not nums: raise ValueError("""Input sequence should not be empty""" ) UpperCamelCase__ :int = nums[0] for i in range(1 , len(lowercase__ ) ): UpperCamelCase__...
45
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): impo...
45
1
import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import Backbone...
45
import math def A ( lowercase__ : Tuple , lowercase__ : Union[str, Any] ) -> Optional[Any]: if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(lowercase__ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0:...
45
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "facebook/convnextv2-tiny-1k-224": "http...
45
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_te...
45
1
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( lowercase ): """simple docstring""" _snake_case : List[Any] = """ClapFeatureExtractor""" _snake_case : int = ("""RobertaTo...
45
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCamelCase = False class lowerCAmelCase_ ( unittest.TestCase ...
45
1
from __future__ import annotations from fractions import Fraction def A ( lowercase__ : int , lowercase__ : int ) -> bool: return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def A ( lowercase__ : int ) -> list[str]: UpperCamelCa...
45
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property ...
45
1
def A ( lowercase__ : int ) -> Optional[Any]: UpperCamelCase__ :List[str] = len(lowercase__ ) for i in range(length - 1 ): UpperCamelCase__ :Tuple = i for k in range(i + 1 , lowercase__ ): if collection[k] < collection[least]: UpperCamelCase__ :str = k ...
45
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import I...
45
1
import numpy as np import pandas as pd from sklearn.preprocessing import Normalizer from sklearn.svm import SVR from statsmodels.tsa.statespace.sarimax import SARIMAX def A ( lowercase__ : list , lowercase__ : list , lowercase__ : list , lowercase__ : list , lowercase__ : list ) -> f...
45
from __future__ import annotations def A ( lowercase__ : int ) -> list[int]: UpperCamelCase__ :Union[str, Any] = [True] * limit UpperCamelCase__ :int = False UpperCamelCase__ :Optional[Any] = False UpperCamelCase__ :str = True for i in range(3 , int...
45
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class lowerCAmelCase_ ( lowercase ): """simple docstring""" _snake_case : int = (PNDMScheduler,) _snake_case : Any = (("""num_inf...
45
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTe...
45
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "facebook/data...
45
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def A ( lowercase__ : dict ) -> tuple: return (data["data"], d...
45
1
import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCa...
45
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def A ( lowercase__ : Optional[int] ) -> Optional[Any]: UpperCamelCase__ :Union[str, Any] = {} UpperCamelCase...
45
1
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) U...
45
def A ( lowercase__ : int ) -> Optional[Any]: stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def A ( lowercase__ : Union[str, Any] , lowercase__ : Dict , lowercase__ : str ) -> List[str]: if i >= h: return # If first element is smaller than the last the...
45
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = {"configuration_speech_encoder_decoder": ["SpeechEncoderDecoderConfig"]} try: if not is_torch_available(): raise OptionalDep...
45
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCamelCase = "." # Internal TensorFlow ops that can be s...
45
1
from __future__ import annotations def A ( lowercase__ : list[int] ) -> int: if not nums: return 0 UpperCamelCase__ :Tuple = nums[0] UpperCamelCase__ :int = 0 for num in nums[1:]: UpperCamelCase__ , UpperCamelCase__ :Optional[Any] = ( max_excluding...
45
from __future__ import annotations def A ( lowercase__ : str , lowercase__ : list[str] | None = None , lowercase__ : dict[str, float] | None = None , lowercase__ : bool = False , ) -> tuple[int, float, str]: UpperCamelCase__ :Dict = cipher_alphabet or [chr(lowercase__ ) for ...
45
1
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def A ( lowercase__ : dict ) -> tuple: return (data["data"], d...
45
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( lowercase ): """simple docstring""" def __init__( self :Union[str, Any] , *lo...
45
1
from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCAmelCase_ ( lowercase...
45
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import D...
45
1
from __future__ import annotations import bisect def A ( lowercase__ : list[int] , lowercase__ : int , lowercase__ : int = 0 , lowercase__ : int = -1 ) -> int: if hi < 0: UpperCamelCase__ :List[Any] = len(lowercase__ ) while lo < hi: UpperCamelCase__ :List[str] =...
45
import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class lowerCAmelCase_ ( lowercase ): """simple docstring""" de...
45
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_te...
45
def A ( lowercase__ : int ) -> bool: if num < 0: return False UpperCamelCase__ :int = num UpperCamelCase__ :int = 0 while num > 0: UpperCamelCase__ :Optional[int] = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main__":...
45
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta impo...
45
from __future__ import annotations def A ( lowercase__ : list[int] ) -> bool: return len(set(lowercase__ ) ) == len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
45
1
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import D...
45
from __future__ import annotations class lowerCAmelCase_ : """simple docstring""" def __init__( self :List[Any] , lowerCamelCase__ :int = 0 ): UpperCamelCase__ :List[str] = key def __a ( self :Optional[Any] , lowerCamelCase__...
45
1
import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, requ...
45
import random def A ( lowercase__ : Dict , lowercase__ : str , lowercase__ : Optional[Any] ) -> int: UpperCamelCase__ :List[Any] = a[left_index] UpperCamelCase__ :Dict = left_index + 1 for j in range(left_index + 1 , lowercase__ ): if a[j] < pivot: UpperCamelC...
45
1
import unittest from transformers.testing_utils import require_bsa from transformers.utils import is_bsa_available from ...test_feature_extraction_common import FeatureExtractionSavingTestMixin if is_bsa_available(): from transformers import MarkupLMFeatureExtractor class lowerCAmelCase_ ( ...
45
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "shi-labs/dinat-mini-in1k-224": "https:/...
45
1
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, a...
45
def A ( lowercase__ : int , lowercase__ : int ) -> int: return int(input_a == input_a == 0 ) def A ( ) -> None: print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(f"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) p...
45
1
def A ( lowercase__ : int , lowercase__ : int ) -> int: UpperCamelCase__ :List[str] = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): UpperCamelCase__ :Any = n - k # Calculate C(n,k) for i in range(lowercase__ ): result *= n - i result...
45
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): impo...
45
1
import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class ...
45
import math def A ( lowercase__ : Tuple , lowercase__ : Union[str, Any] ) -> Optional[Any]: if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(lowercase__ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0:...
45
1
from PIL import Image def A ( lowercase__ : Image , lowercase__ : float ) -> Image: def brightness(lowercase__ : int ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError("""level must be between -255.0 (black) and 255.0 (white)""" ) return ...
45
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_te...
45
1
import json import sys def A ( lowercase__ : Optional[Any] , lowercase__ : Any ) -> Tuple: with open(lowercase__ , encoding="""utf-8""" ) as f: UpperCamelCase__ :Union[str, Any] = json.load(lowercase__ ) UpperCamelCase__ :Optional[Any] = ["""<details>""", """<summa...
45
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCamelCase = False class lowerCAmelCase_ ( unittest.TestCase ...
45
1
import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) lo...
45
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property ...
45
1
import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( lowercase ): """simple docstring""" def __init__( self :Any , *lowerCamelCase__ ...
45
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import I...
45
1
import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... im...
45
from __future__ import annotations def A ( lowercase__ : int ) -> list[int]: UpperCamelCase__ :Union[str, Any] = [True] * limit UpperCamelCase__ :int = False UpperCamelCase__ :Optional[Any] = False UpperCamelCase__ :str = True for i in range(3 , int...
45
1
UpperCamelCase = { "A": ".-", "B": "-...", "C": "-.-.", "D": "-..", "E": ".", "F": "..-.", "G": "--.", "H": "....", "I": "..", "J": ".---", "K": "-.-", "L": ".-..", "M": "--", "N": "-.", "O": "---", "P": ".--.", "Q": "--.-", "R": ".-.", "S": "...", "T": "-", "U": "..-", "V": "...-", "W"...
45
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTe...
45
1
from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class lowerCAmelCase_ ( lowercase ): ...
45
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def A ( lowercase__ : dict ) -> tuple: return (data["data"], d...
45
1
import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class lowerCAmelCase_ : """simple docstring""" def __init__( self :str , lowerCamelCase__ :Dict ): ...
45
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def A ( lowercase__ : Optional[int] ) -> Optional[Any]: UpperCamelCase__ :Union[str, Any] = {} UpperCamelCase...
45
1
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from .....
45
def A ( lowercase__ : int ) -> Optional[Any]: stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def A ( lowercase__ : Union[str, Any] , lowercase__ : Dict , lowercase__ : str ) -> List[str]: if i >= h: return # If first element is smaller than the last the...
45
1
import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def A ( lowercase__ : Optional[Any] , lowercase__ : str=7 ) -> List[str]: UpperCamelCase__ :List[Any] = None if token is not None: UpperCamelCase__ :Optional[int]...
45
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCamelCase = "." # Internal TensorFlow ops that can be s...
45
1
from random import shuffle import tensorflow as tf from numpy import array def A ( lowercase__ : str , lowercase__ : Optional[int] ) -> str: UpperCamelCase__ :Any = int(lowercase__ ) assert noofclusters < len(lowercase__ ) # Find out the dimensionality UpperCamelCase__ :str ...
45
from __future__ import annotations def A ( lowercase__ : str , lowercase__ : list[str] | None = None , lowercase__ : dict[str, float] | None = None , lowercase__ : bool = False , ) -> tuple[int, float, str]: UpperCamelCase__ :Dict = cipher_alphabet or [chr(lowercase__ ) for ...
45
1
from __future__ import annotations def A ( lowercase__ : list[int] ) -> bool: return len(set(lowercase__ ) ) == len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
45
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( lowercase ): """simple docstring""" def __init__( self :Union[str, Any] , *lo...
45
1
import argparse import torch from huggingface_hub import hf_hub_download from transformers import AutoTokenizer, RobertaPreLayerNormConfig, RobertaPreLayerNormForMaskedLM from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def A ...
45
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import D...
45
1
from __future__ import annotations UpperCamelCase = 8.988E9 # units = N * m^s * C^-2 def A ( lowercase__ : float , lowercase__ : float , lowercase__ : float , lowercase__ : float ) -> dict[str, float]: UpperCamelCase__ :Optional[int] = abs(chargea * chargea ) i...
45
import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class lowerCAmelCase_ ( lowercase ): """simple docstring""" de...
45
1
def A ( lowercase__ : int ) -> bool: return str(lowercase__ ) == str(lowercase__ )[::-1] def A ( lowercase__ : int ) -> int: return int(lowercase__ ) + int(str(lowercase__ )[::-1] ) def A ( lowercase__ : int = 1_0000 ) -> int: UpperCamelCase__ :int = ...
45
def A ( lowercase__ : int ) -> bool: if num < 0: return False UpperCamelCase__ :int = num UpperCamelCase__ :int = 0 while num > 0: UpperCamelCase__ :Optional[int] = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main__":...
45
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available UpperCamelCase = { "configuration_audio_spectrogram_transformer": [ "AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "ASTConfig...
45
from __future__ import annotations def A ( lowercase__ : list[int] ) -> bool: return len(set(lowercase__ ) ) == len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
45
1
def A ( lowercase__ : list[int] , lowercase__ : list[int] ) -> None: UpperCamelCase__ :Union[str, Any] = len(lowercase__ ) print("""The following activities are selected:""" ) # The first activity is always selected UpperCamelCase__ :int = 0 print(lowercase__ , end=""...
45
from __future__ import annotations class lowerCAmelCase_ : """simple docstring""" def __init__( self :List[Any] , lowerCamelCase__ :int = 0 ): UpperCamelCase__ :List[str] = key def __a ( self :Optional[Any] , lowerCamelCase__...
45
1
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase = {"configuration_mmbt": ["MMBTConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAv...
45
import random def A ( lowercase__ : Dict , lowercase__ : str , lowercase__ : Optional[Any] ) -> int: UpperCamelCase__ :List[Any] = a[left_index] UpperCamelCase__ :Dict = left_index + 1 for j in range(left_index + 1 , lowercase__ ): if a[j] < pivot: UpperCamelC...
45
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase = {} try: if not is_sentencepiece_available(): raise OptionalDepend...
45
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "shi-labs/dinat-mini-in1k-224": "https:/...
45
1
import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import loggi...
45
def A ( lowercase__ : int , lowercase__ : int ) -> int: return int(input_a == input_a == 0 ) def A ( ) -> None: print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(f"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) p...
45
1
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def A ( lowercase__ : str = "" ) -> dict[str, float]: UpperCamelCase__ :Union[str, Any] = url or """https://www.imdb.com/chart/top/?ref_=nv_mv_250""" UpperCamelCase__ :List[str] = ...
45
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): impo...
45
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCamelCase = False class lowerCAmelCase_ ( unittest.TestCase ...
45
import math def A ( lowercase__ : Tuple , lowercase__ : Union[str, Any] ) -> Optional[Any]: if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(lowercase__ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0:...
45
1
def A ( lowercase__ : int ) -> bool: if num < 0: return False UpperCamelCase__ :int = num UpperCamelCase__ :int = 0 while num > 0: UpperCamelCase__ :Optional[int] = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main__":...
45
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_te...
45
1
# Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar UpperCamelCase = TypeVar("T") class lowerCAmelCase_ ( Generic[T] ): """simple do...
45
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCamelCase = False class lowerCAmelCase_ ( unittest.TestCase ...
45
1
def A ( lowercase__ : int = 50 ) -> int: UpperCamelCase__ :Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row...
45
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property ...
45
1
from __future__ import annotations def A ( lowercase__ : list[float] ) -> float: UpperCamelCase__ :Optional[int] = 0.00 UpperCamelCase__ :Tuple = 0 for resistor in resistors: if resistor <= 0: UpperCamelCase__ :List[Any] = f"""Resistor at index {index} has a...
45
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import I...
45
1
UpperCamelCase = range(2, 20 + 1) UpperCamelCase = [10**k for k in range(ks[-1] + 1)] UpperCamelCase = {} def A ( lowercase__ : List[Any] , lowercase__ : Tuple , lowercase__ : str , lowercase__ : Optional[Any] ) -> Any: UpperCamelCase__ :List[Any] = ...
45
from __future__ import annotations def A ( lowercase__ : int ) -> list[int]: UpperCamelCase__ :Union[str, Any] = [True] * limit UpperCamelCase__ :int = False UpperCamelCase__ :Optional[Any] = False UpperCamelCase__ :str = True for i in range(3 , int...
45
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers...
45
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTe...
45
1
import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate.test_uti...
45
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def A ( lowercase__ : dict ) -> tuple: return (data["data"], d...
45
1
import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available ...
45
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def A ( lowercase__ : Optional[int] ) -> Optional[Any]: UpperCamelCase__ :Union[str, Any] = {} UpperCamelCase...
45
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "asapp/sew-d-tiny-100k": "https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json", # Se...
45
def A ( lowercase__ : int ) -> Optional[Any]: stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def A ( lowercase__ : Union[str, Any] , lowercase__ : Dict , lowercase__ : str ) -> List[str]: if i >= h: return # If first element is smaller than the last the...
45
1
from __future__ import annotations def A ( lowercase__ : int | float | str , lowercase__ : int | float | str ) -> list[str]: if nth_term == "": return [""] UpperCamelCase__ :Dict = int(lowercase__ ) UpperCamelCase__ :Union[str, Any] = int(lowercase__ ) UpperCamelCa...
45
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCamelCase = "." # Internal TensorFlow ops that can be s...
45
1
def A ( lowercase__ : int = 50 ) -> int: UpperCamelCase__ :Optional[int] = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_num...
45
from __future__ import annotations def A ( lowercase__ : str , lowercase__ : list[str] | None = None , lowercase__ : dict[str, float] | None = None , lowercase__ : bool = False , ) -> tuple[int, float, str]: UpperCamelCase__ :Dict = cipher_alphabet or [chr(lowercase__ ) for ...
45
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ......
45
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( lowercase ): """simple docstring""" def __init__( self :Union[str, Any] , *lo...
45
1
# flake8: noqa # Lint as: python3 from typing import Dict, List, Optional, Type from .. import config from ..utils import logging from .formatting import ( ArrowFormatter, CustomFormatter, Formatter, PandasFormatter, PythonFormatter, TensorFormatter, format_table, query_table, )...
45
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import D...
45
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { "configuration_squeezebert": [ "SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "SqueezeBertConfig", "SqueezeBertO...
45
import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class lowerCAmelCase_ ( lowercase ): """simple docstring""" de...
45
1
import random from typing import Any def A ( lowercase__ : list ) -> list[Any]: for _ in range(len(lowercase__ ) ): UpperCamelCase__ :Optional[int] = random.randint(0 , len(lowercase__ ) - 1 ) UpperCamelCase__ :Optional[int] = random.randint(0 , len(lowercase__ ) ...
45
def A ( lowercase__ : int ) -> bool: if num < 0: return False UpperCamelCase__ :int = num UpperCamelCase__ :int = 0 while num > 0: UpperCamelCase__ :Optional[int] = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main__":...
45
1
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, ...
45
from __future__ import annotations def A ( lowercase__ : list[int] ) -> bool: return len(set(lowercase__ ) ) == len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
45
1
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.s...
45
from __future__ import annotations class lowerCAmelCase_ : """simple docstring""" def __init__( self :List[Any] , lowerCamelCase__ :int = 0 ): UpperCamelCase__ :List[str] = key def __a ( self :Optional[Any] , lowerCamelCase__...
45
1
from __future__ import annotations def A ( lowercase__ : int ) -> list[int]: UpperCamelCase__ :Union[str, Any] = [True] * limit UpperCamelCase__ :int = False UpperCamelCase__ :Optional[Any] = False UpperCamelCase__ :str = True for i in range(3 , int...
45
import random def A ( lowercase__ : Dict , lowercase__ : str , lowercase__ : Optional[Any] ) -> int: UpperCamelCase__ :List[Any] = a[left_index] UpperCamelCase__ :Dict = left_index + 1 for j in range(left_index + 1 , lowercase__ ): if a[j] < pivot: UpperCamelC...
45
1
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand ...
45
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "shi-labs/dinat-mini-in1k-224": "https:/...
45
1
from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounte...
45
def A ( lowercase__ : int , lowercase__ : int ) -> int: return int(input_a == input_a == 0 ) def A ( ) -> None: print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(f"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) p...
45
1
import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def A ( lowercase__ : Optional[int] ) -> int: UpperCamelCase__ :Optional[int] = os.path.join(args.tf_model_dir , """parameters.json""" ) UpperCamelCase...
45
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): impo...
45
1
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTe...
45
import math def A ( lowercase__ : Tuple , lowercase__ : Union[str, Any] ) -> Optional[Any]: if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(lowercase__ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0:...
45
1
class lowerCAmelCase_ : # Public class to implement a graph """simple docstring""" def __init__( self :List[Any] , lowerCamelCase__ :int , lowerCamelCase__ :int , lowerCamelCase__ :list[list[bool]] ): UpperCamelCase__ :List[str] = ro...
45
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_te...
45
1
# Algorithm for the pigeonhole sorting def A ( lowercase__ : Dict ) -> Optional[int]: UpperCamelCase__ :Tuple = min(lowercase__ ) # min() finds the minimum value UpperCamelCase__ :Optional[int] = max(lowercase__ ) # max() finds the maximum value UpperCamelCase__ :st...
45
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCamelCase = False class lowerCAmelCase_ ( unittest.TestCase ...
45
1
def A ( lowercase__ : list ) -> bool: if not isinstance(lowercase__ , lowercase__ ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(lowercase__ ) == 0: raise ValueError("""Input list must be a non empty list""" ) if len(lowercase__ ) == 1: return T...
45
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property ...
45
1
import pytest import datasets # Import fixture modules as plugins UpperCamelCase = ["tests.fixtures.files", "tests.fixtures.hub", "tests.fixtures.fsspec"] def A ( lowercase__ : int , lowercase__ : Optional[Any] ) -> int: # Mark tests as "unit" by default if not marked as "integr...
45
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import I...
45
1
import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( Ef...
45
from __future__ import annotations def A ( lowercase__ : int ) -> list[int]: UpperCamelCase__ :Union[str, Any] = [True] * limit UpperCamelCase__ :int = False UpperCamelCase__ :Optional[Any] = False UpperCamelCase__ :str = True for i in range(3 , int...
45
1
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--dump_path", default=None, type=st...
45
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTe...
45
1
import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device UpperCamelCase = False class lowerCAmelCase_ ( unittest.TestCase ): "...
45
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def A ( lowercase__ : dict ) -> tuple: return (data["data"], d...
45
1
import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import r...
45
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def A ( lowercase__ : Optional[int] ) -> Optional[Any]: UpperCamelCase__ :Union[str, Any] = {} UpperCamelCase...
45
1
import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class lowerCAmelCase_ ( lowercase ): """simple do...
45
def A ( lowercase__ : int ) -> Optional[Any]: stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def A ( lowercase__ : Union[str, Any] , lowercase__ : Dict , lowercase__ : str ) -> List[str]: if i >= h: return # If first element is smaller than the last the...
45
1
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow UpperCamelCase = logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""" ) @...
45
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCamelCase = "." # Internal TensorFlow ops that can be s...
45
1
from math import factorial UpperCamelCase = {str(d): factorial(d) for d in range(10)} def A ( lowercase__ : int ) -> int: return sum(DIGIT_FACTORIAL[d] for d in str(lowercase__ ) ) def A ( ) -> int: UpperCamelCase__ :Dict = 7 * factorial(9 ) + 1 return su...
45
from __future__ import annotations def A ( lowercase__ : str , lowercase__ : list[str] | None = None , lowercase__ : dict[str, float] | None = None , lowercase__ : bool = False , ) -> tuple[int, float, str]: UpperCamelCase__ :Dict = cipher_alphabet or [chr(lowercase__ ) for ...
45
1
def A ( lowercase__ : int , lowercase__ : int ) -> int: return int(input_a == input_a == 0 ) def A ( ) -> None: print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(f"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) p...
45
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( lowercase ): """simple docstring""" def __init__( self :Union[str, Any] , *lo...
45
1
from __future__ import annotations from collections.abc import Sequence from typing import Literal def A ( lowercase__ : str , lowercase__ : str ) -> str | Literal[False]: UpperCamelCase__ :Any = list(lowercase__ ) UpperCamelCase__ :Optional[Any] = list(lowercase__ ) ...
45
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import D...
45
1
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=lowercase ) class lowerCAmelCase_ ( lowercase ): """simple docstring""" _snake_case : ...
45
import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class lowerCAmelCase_ ( lowercase ): """simple docstring""" de...
45
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging ...
45
def A ( lowercase__ : int ) -> bool: if num < 0: return False UpperCamelCase__ :int = num UpperCamelCase__ :int = 0 while num > 0: UpperCamelCase__ :Optional[int] = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main__":...
45
1
from __future__ import annotations def A ( lowercase__ : int ) -> list[int]: UpperCamelCase__ :Tuple = 2 UpperCamelCase__ :Dict = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(lowercase__ ) if n > 1: factors.append(lowercase__ ) retu...
45
from __future__ import annotations def A ( lowercase__ : list[int] ) -> bool: return len(set(lowercase__ ) ) == len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
45
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "shi-labs/dinat-mini-in1k-224": "https:/...
45
from __future__ import annotations class lowerCAmelCase_ : """simple docstring""" def __init__( self :List[Any] , lowerCamelCase__ :int = 0 ): UpperCamelCase__ :List[str] = key def __a ( self :Optional[Any] , lowerCamelCase__...
45
1
UpperCamelCase = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] UpperCamelCase = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] UpperCamelCase = { 0: "Sunday", 1: "Monday", 2: "Tuesday", 3: "Wednesday", 4: "Thursday", 5: "Friday", 6: "Saturday", } def A ( lowerc...
45
import random def A ( lowercase__ : Dict , lowercase__ : str , lowercase__ : Optional[Any] ) -> int: UpperCamelCase__ :List[Any] = a[left_index] UpperCamelCase__ :Dict = left_index + 1 for j in range(left_index + 1 , lowercase__ ): if a[j] < pivot: UpperCamelC...
45
1
import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def A ( lowercase__ : int ) -> Optional[int]: UpperCamelCase__ :Dict = FileLock(str(tmpdir / """foo.lock""" ) ) UpperCamelCase__ :str = FileLock(str(tmpdir / """foo.lock""" ) ) ...
45
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "shi-labs/dinat-mini-in1k-224": "https:/...
45
1
from __future__ import annotations from PIL import Image # Define glider example UpperCamelCase = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0...
45
def A ( lowercase__ : int , lowercase__ : int ) -> int: return int(input_a == input_a == 0 ) def A ( ) -> None: print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(f"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) p...
45
1
from __future__ import annotations import math from collections.abc import Callable def A ( lowercase__ : Callable[[int | float], int | float] , lowercase__ : int | float , lowercase__ : int | float , lowercase__ : int = 100 , ) -> float: UpperCamelCase__ :int = x_start Upp...
45
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): impo...
45
1
from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class lowerCAmelCase_ : """simple docstring""" def __init__( self :Dict , lowerCamelCase__ :Collection[float] | None = None ): if co...
45
import math def A ( lowercase__ : Tuple , lowercase__ : Union[str, Any] ) -> Optional[Any]: if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(lowercase__ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0:...
45
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging....
45
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_te...
45
1
def A ( lowercase__ : Union[str, Any] ) -> List[Any]: UpperCamelCase__ :Dict = [] UpperCamelCase__ :List[str] = [] UpperCamelCase__ :Optional[Any] = { """^""": 3, """*""": 2, """/""": 2, """%""": 2, """+""": 1, """-""": 1, } # Priority...
45
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCamelCase = False class lowerCAmelCase_ ( unittest.TestCase ...
45
1