python_code
stringlengths
0
992k
repo_name
stringlengths
8
46
file_path
stringlengths
5
162
from .tin_lr_hook import TINLrUpdaterHook __all__ = ['TINLrUpdaterHook']
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/core/lr/__init__.py
from mmcv.runner import HOOKS, LrUpdaterHook from mmcv.runner.hooks.lr_updater import annealing_cos @HOOKS.register_module() class TINLrUpdaterHook(LrUpdaterHook): def __init__(self, min_lr, **kwargs): self.min_lr = min_lr super(TINLrUpdaterHook, self).__init__(**kwargs) def get_warmup_lr(se...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/core/lr/tin_lr_hook.py
import os.path as osp import torch from .base import BaseDataset from .registry import DATASETS @DATASETS.register_module() class AudioDataset(BaseDataset): """Audio dataset for video recognition. Extracts the audio feature on-the- fly. Annotation file can be that of the rawframe dataset, or: .. code-b...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/audio_dataset.py
import copy import os import os.path as osp import warnings import mmcv import numpy as np from ..core import average_recall_at_avg_proposals from .base import BaseDataset from .registry import DATASETS @DATASETS.register_module() class ActivityNetDataset(BaseDataset): """ActivityNet dataset for temporal action...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/activitynet_dataset.py
import copy import os.path as osp import numpy as np import torch from .base import BaseDataset from .registry import DATASETS @DATASETS.register_module() class RawframeDataset(BaseDataset): """Rawframe dataset for action recognition. The dataset loads raw frames and apply specified transforms to return a ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/rawframe_dataset.py
import copy import os.path as osp import mmcv import numpy as np from mmcv.utils import print_log from ..core import mean_average_precision from .base import BaseDataset from .registry import DATASETS @DATASETS.register_module() class HVUDataset(BaseDataset): """HVU dataset, which supports the recognition tags ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/hvu_dataset.py
from .registry import DATASETS from .video_dataset import VideoDataset @DATASETS.register_module() class ImageDataset(VideoDataset): """Image dataset for action recognition, used in the Project OmniSource. The dataset loads image list and apply specified transforms to return a dict containing the image t...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/image_dataset.py
from mmcv.utils import Registry DATASETS = Registry('dataset') PIPELINES = Registry('pipeline')
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/registry.py
import os.path as osp import torch from .base import BaseDataset from .registry import DATASETS @DATASETS.register_module() class AudioFeatureDataset(BaseDataset): """Audio feature dataset for video recognition. Reads the features extracted off-line. Annotation file can be that of the rawframe dataset, ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/audio_feature_dataset.py
from .activitynet_dataset import ActivityNetDataset from .audio_dataset import AudioDataset from .audio_feature_dataset import AudioFeatureDataset from .audio_visual_dataset import AudioVisualDataset from .ava_dataset import AVADataset from .base import BaseDataset from .builder import build_dataloader, build_dataset f...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/__init__.py
import platform import random from functools import partial import numpy as np from mmcv.parallel import collate from mmcv.runner import get_dist_info from mmcv.utils import build_from_cfg from torch.utils.data import DataLoader from .dataset_wrappers import RepeatDataset from .registry import DATASETS from .samplers...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/builder.py
import os.path as osp from .rawframe_dataset import RawframeDataset from .registry import DATASETS @DATASETS.register_module() class AudioVisualDataset(RawframeDataset): """Dataset that reads both audio and visual data, supporting both rawframes and videos. The annotation file is same as that of the rawframe...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/audio_visual_dataset.py
from .registry import DATASETS @DATASETS.register_module() class RepeatDataset: """A wrapper of repeated dataset. The length of repeated dataset will be ``times`` larger than the original dataset. This is useful when the data loading time is long but the dataset is small. Using RepeatDataset can redu...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/dataset_wrappers.py
import os.path as osp import torch from .base import BaseDataset from .registry import DATASETS @DATASETS.register_module() class VideoDataset(BaseDataset): """Video dataset for action recognition. The dataset loads raw videos and apply specified transforms to return a dict containing the frame tensors...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/video_dataset.py
import copy import os.path as osp import warnings import mmcv import numpy as np from torch.nn.modules.utils import _pair from ..core import softmax from ..localization import (eval_ap, load_localize_proposal_file, perform_regression, temporal_iou, temporal_nms) from ..utils import get_roo...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/ssn_dataset.py
import copy import os.path as osp from collections import defaultdict import mmcv import numpy as np from ..utils import get_root_logger from .base import BaseDataset from .registry import DATASETS @DATASETS.register_module() class AVADataset(BaseDataset): """AVA dataset for spatial temporal detection. Bas...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/ava_dataset.py
import copy import os.path as osp import warnings from abc import ABCMeta, abstractmethod from collections import defaultdict import mmcv import numpy as np import torch from mmcv.utils import print_log from torch.utils.data import Dataset from ..core import (mean_average_precision, mean_class_accuracy, ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/base.py
import copy import os.path as osp import random import mmcv from .base import BaseDataset from .registry import DATASETS @DATASETS.register_module() class RawVideoDataset(BaseDataset): """RawVideo dataset for action recognition, used in the Project OmniSource. The dataset loads clips of raw videos and appl...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/rawvideo_dataset.py
import io import os import os.path as osp import shutil import warnings import mmcv import numpy as np import torch from mmcv.fileio import FileClient from torch.nn.modules.utils import _pair from ...utils import get_random_string, get_shm_dir, get_thread_id from ..registry import PIPELINES @PIPELINES.register_modu...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/loading.py
from collections.abc import Sequence from mmcv.utils import build_from_cfg from ..registry import PIPELINES @PIPELINES.register_module() class Compose: """Compose a data pipeline with a sequence of transforms. Args: transforms (list[dict | callable]): Either config dicts of transforms o...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/compose.py
from .augmentations import (AudioAmplify, CenterCrop, ColorJitter, EntityBoxClip, EntityBoxCrop, EntityBoxFlip, EntityBoxPad, EntityBoxRescale, Flip, Fuse, MelSpectrogram, MultiGroupCrop, MultiScaleCrop, Norm...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/__init__.py
import random from collections.abc import Sequence import mmcv import numpy as np from torch.nn.modules.utils import _pair from ..registry import PIPELINES def _init_lazy_if_proper(results, lazy): """Initialize lazy operation properly. Make sure that a lazy operation is properly initialized, and avoid ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/augmentations.py
from collections.abc import Sequence import mmcv import numpy as np import torch from mmcv.parallel import DataContainer as DC from ..registry import PIPELINES def to_tensor(data): """Convert objects of various python types to :obj:`torch.Tensor`. Supported types are: :class:`numpy.ndarray`, :class:`torch....
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/formating.py
from .distributed_sampler import DistributedPowerSampler, DistributedSampler __all__ = ['DistributedSampler', 'DistributedPowerSampler']
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/samplers/__init__.py
import torch from torch.utils.data import DistributedSampler as _DistributedSampler class DistributedSampler(_DistributedSampler): """DistributedSampler inheriting from ``torch.utils.data.DistributedSampler``. In pytorch of lower versions, there is no ``shuffle`` argument. This child class will port ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/datasets/samplers/distributed_sampler.py
import ctypes import random import string def get_random_string(length=15): """Get random string with letters and digits. Args: length (int): Length of random string. Default: 15. """ return ''.join( random.choice(string.ascii_letters + string.digits) for _ in range(length)) ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/utils/misc.py
from mmcv.utils import collect_env as collect_basic_env from mmcv.utils import get_git_hash import mmaction def collect_env(): env_info = collect_basic_env() env_info['MMAction2'] = ( mmaction.__version__ + '+' + get_git_hash(digits=7)) return env_info if __name__ == '__main__': for name, v...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/utils/collect_env.py
from .collect_env import collect_env from .gradcam_utils import GradCAM from .logger import get_root_logger from .misc import get_random_string, get_shm_dir, get_thread_id __all__ = [ 'get_root_logger', 'collect_env', 'get_random_string', 'get_thread_id', 'get_shm_dir', 'GradCAM' ]
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/utils/__init__.py
import logging from mmcv.utils import get_logger def get_root_logger(log_file=None, log_level=logging.INFO): """Use ``get_logger`` method in mmcv to get the root logger. The logger will be initialized if it has not been initialized. By default a StreamHandler will be added. If ``log_file`` is specified,...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/utils/logger.py
import torch import torch.nn.functional as F class GradCAM: """GradCAM class helps create visualization results. Visualization results are blended by heatmaps and input images. This class is modified from https://github.com/facebookresearch/SlowFast/blob/master/slowfast/visualization/gradcam_utils.py...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/utils/gradcam_utils.py
from mmcv.utils import Registry BACKBONES = Registry('backbone') NECKS = Registry('neck') HEADS = Registry('head') RECOGNIZERS = Registry('recognizer') LOSSES = Registry('loss') LOCALIZERS = Registry('localizer')
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/registry.py
from .backbones import (C3D, X3D, ResNet, ResNet2Plus1d, ResNet3d, ResNet3dCSN, ResNet3dSlowFast, ResNet3dSlowOnly, ResNetAudio, ResNetTIN, ResNetTSM) from .builder import (build_backbone, build_head, build_localizer, build_loss, build_model, build_n...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/__init__.py
import torch.nn as nn from mmcv.utils import build_from_cfg from .registry import BACKBONES, HEADS, LOCALIZERS, LOSSES, NECKS, RECOGNIZERS def build(cfg, registry, default_args=None): """Build a module. Args: cfg (dict, list[dict]): The config of modules, it is either a dict or a list of...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/builder.py
import torch import torch.nn as nn import torch.nn.functional as F from ..registry import LOSSES from ...core import top_k_accuracy @LOSSES.register_module() class BayesianNNLoss(nn.Module): """Bayesian NN Loss.""" def forward(self, cls_score, labels, output_dict, beta=1.0, **kwargs): """Forward func...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/bnn_loss.py
import torch import torch.nn.functional as F from ..registry import LOSSES from .base import BaseWeightedLoss def relu_evidence(y): return F.relu(y) def exp_evidence(y): return torch.exp(torch.clamp(y, -10, 10)) def softplus_evidence(y): return F.softplus(y) @LOSSES.register_module() class EvidenceLoss...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/edl_loss.py
import torch import torch.nn.functional as F import numpy as np from ..registry import LOSSES from .base import BaseWeightedLoss @LOSSES.register_module() class RebiasLoss(BaseWeightedLoss): """Rebias Loss.""" def __init__(self, lambda_g=1.0, criteria='hsic'): super().__init__() self.lambda_g...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/rebias_loss.py
from .base import BaseWeightedLoss from .binary_logistic_regression_loss import BinaryLogisticRegressionLoss from .bmn_loss import BMNLoss from .cross_entropy_loss import BCELossWithLogits, CrossEntropyLoss from .bnn_loss import BayesianNNLoss from .edl_loss import EvidenceLoss from .hvu_loss import HVULoss from .nll_l...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/__init__.py
import torch import torch.nn as nn import torch.nn.functional as F from ..registry import LOSSES from .ohem_hinge_loss import OHEMHingeLoss @LOSSES.register_module() class SSNLoss(nn.Module): @staticmethod def activity_loss(activity_score, labels, activity_indexer): """Activity Loss. It wil...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/ssn_loss.py
import torch import torch.nn.functional as F from ..registry import LOSSES from .base import BaseWeightedLoss from ...core import top_k_accuracy @LOSSES.register_module() class GCPLoss(BaseWeightedLoss): """Reciprocal Point Learning Loss.""" def __init__(self, temperature=1, weight_pl=0.1, radius_init=1): ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/gcp_loss.py
import torch import torch.nn as nn from ..registry import LOSSES def binary_logistic_regression_loss(reg_score, label, threshold=0.5, ratio_range=(1.05, 21), eps=1e-5): ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/binary_logistic_regression_loss.py
import torch.nn.functional as F from ..registry import LOSSES from .base import BaseWeightedLoss @LOSSES.register_module() class CrossEntropyLoss(BaseWeightedLoss): """Cross Entropy Loss.""" def _forward(self, cls_score, label, **kwargs): """Forward function. Args: cls_score (to...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/cross_entropy_loss.py
import torch class OHEMHingeLoss(torch.autograd.Function): """This class is the core implementation for the completeness loss in paper. It compute class-wise hinge loss and performs online hard example mining (OHEM). """ @staticmethod def forward(ctx, pred, labels, is_positive, ohem_rati...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/ohem_hinge_loss.py
import torch.nn.functional as F from ..registry import LOSSES from .base import BaseWeightedLoss @LOSSES.register_module() class NLLLoss(BaseWeightedLoss): """NLL Loss. It will calculate NLL loss given cls_score and label. """ def _forward(self, cls_score, label, **kwargs): """Forward funct...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/nll_loss.py
import torch import torch.nn.functional as F from ..registry import LOSSES from .base import BaseWeightedLoss @LOSSES.register_module() class HVULoss(BaseWeightedLoss): """Calculate the BCELoss for HVU. Args: categories (tuple[str]): Names of tag categories, tags are organized in this or...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/hvu_loss.py
from abc import ABCMeta, abstractmethod import torch.nn as nn class BaseWeightedLoss(nn.Module, metaclass=ABCMeta): """Base class for loss. All subclass should overwrite the ``_forward()`` method which returns the normal loss without loss weights. Args: loss_weight (float): Factor scalar mu...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/base.py
import torch import torch.nn.functional as F from ..registry import LOSSES from .base import BaseWeightedLoss from ...core import top_k_accuracy @LOSSES.register_module() class RPLoss(BaseWeightedLoss): """Reciprocal Point Learning Loss.""" def __init__(self, temperature=1, weight_pl=0.1, radius_init=1): ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/rpl_loss.py
import torch import torch.nn as nn import torch.nn.functional as F from ..registry import LOSSES from .binary_logistic_regression_loss import binary_logistic_regression_loss @LOSSES.register_module() class BMNLoss(nn.Module): """BMN Loss. From paper https://arxiv.org/abs/1907.09702, code https://github....
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/losses/bmn_loss.py
from .tpn import TPN __all__ = ['TPN']
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/necks/__init__.py
import numpy as np import torch import torch.nn as nn from mmcv.cnn import ConvModule, constant_init, normal_init, xavier_init from ..heads import AuxHead, RebiasHead from ..registry import NECKS class Identity(nn.Module): """Identity mapping.""" def forward(self, x): return x class DownSample(nn.M...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/necks/tpn.py
import torch import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import BaseHead @HEADS.register_module() class SlowFastHead(BaseHead): """The classification head for SlowFast. Args: num_classes (int): Number of classes to be classified. in_channels...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/slowfast_head.py
import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import BaseHead @HEADS.register_module() class X3DHead(BaseHead): """Classification head for I3D. Args: num_classes (int): Number of classes to be classified. in_channels (int): Number of channels ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/x3d_head.py
import torch import torch.nn as nn from mmcv.cnn import ConvModule, constant_init, normal_init, xavier_init from ..builder import build_loss class AuxHead(nn.Module): """Auxiliary Head. This auxiliary head is appended to receive stronger supervision, leading to enhanced semantics. Args: in_c...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/aux_head.py
import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import AvgConsensus, BaseHead @HEADS.register_module() class TSNHead(BaseHead): """Class head for TSN. Args: num_classes (int): Number of classes to be classified. in_channels (int): Number of chan...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tsn_head.py
import torch import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import BaseHead @HEADS.register_module() class SlowFastRPLHead(BaseHead): """The classification head for SlowFast. Args: num_classes (int): Number of classes to be classified. in_chann...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/slowfast_rpl_head.py
import torch import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import BaseHead @HEADS.register_module() class I3DRPLHead(BaseHead): """Classification head for I3D. Args: num_classes (int): Number of classes to be classified. in_channels (int): Num...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/i3d_rpl_head.py
import torch import torch.nn as nn from mmcv.cnn import ConvModule, constant_init, normal_init, xavier_init from ..builder import build_loss class RebiasHead(nn.Module): def __init__(self, in_channels, out_channels, loss_weight=0.5, loss_rebias=d...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/rebias_head.py
import torch import torch.nn as nn import torch.nn.functional as F import math class Gaussian(object): def __init__(self, mu, rho): super().__init__() self.mu = mu self.rho = rho self.normal = torch.distributions.Normal(0,1) @property def sigma(self): return tor...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/bnn.py
import torch import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS def parse_stage_config(stage_cfg): """Parse config of STPP for three stages. Args: stage_cfg (int | tuple[int]): Config of structured temporal pyramid pooling. Returns: tuple[tup...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/ssn_head.py
from .audio_tsn_head import AudioTSNHead from .base import BaseHead from .i3d_head import I3DHead from .i3d_bnn_head import I3DBNNHead from .i3d_rpl_head import I3DRPLHead from .slowfast_head import SlowFastHead from .slowfast_rpl_head import SlowFastRPLHead from .slowfast_bnn_head import SlowFastBNNHead from .ssn_head...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/__init__.py
import torch.nn as nn from ..registry import HEADS from .tsn_head import TSNHead from ..builder import build_loss from .bnn import BayesianPredictor, get_uncertainty @HEADS.register_module() class TPNBNNHead(TSNHead): """Class head for TPN. Args: num_classes (int): Number of classes to be classified...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tpn_bnn_head.py
import torch import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import AvgConsensus, BaseHead @HEADS.register_module() class TSMRPLHead(BaseHead): """Class head for TSM. Args: num_classes (int): Number of classes to be classified. in_channels (int)...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tsm_rpl_head.py
import torch import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import AvgConsensus, BaseHead @HEADS.register_module() class TSMHead(BaseHead): """Class head for TSM. Args: num_classes (int): Number of classes to be classified. in_channels (int): N...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tsm_head.py
import torch import torch.nn as nn from ..registry import HEADS from .base import BaseHead from ..builder import build_loss from .bnn import BayesianPredictor, get_uncertainty @HEADS.register_module() class I3DBNNHead(BaseHead): """Classification head for I3D. Args: num_classes (int): Number of clas...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/i3d_bnn_head.py
import torch.nn as nn from ..registry import HEADS from .tsn_head import TSNHead @HEADS.register_module() class TPNHead(TSNHead): """Class head for TPN. Args: num_classes (int): Number of classes to be classified. in_channels (int): Number of channels in input feature. loss_cls (dict...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tpn_head.py
# Copyright (c) OpenMMLab. All rights reserved. import torch import torch.nn as nn from mmcv.cnn import normal_init from ..builder import HEADS from .base import BaseHead import pdb @HEADS.register_module() class BaseClsHead(BaseHead): """The classification head for SlowFast. Args: num_classes (int):...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/base_cls_head.py
import torch import torch.nn as nn class Dist(nn.Module): def __init__(self, num_classes=10, num_centers=1, feat_dim=2, init='random'): super(Dist, self).__init__() self.feat_dim = feat_dim self.num_classes = num_classes self.num_centers = num_centers if init == 'random': ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/rpl_dist.py
import torch import torch.nn as nn from ..registry import HEADS from .base import BaseHead from ..builder import build_loss from .bnn import BayesianPredictor, get_uncertainty @HEADS.register_module() class SlowFastBNNHead(BaseHead): """The classification head for SlowFast. Args: num_classes (int): ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/slowfast_bnn_head.py
import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import BaseHead @HEADS.register_module() class I3DHead(BaseHead): """Classification head for I3D. Args: num_classes (int): Number of classes to be classified. in_channels (int): Number of channels ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/i3d_head.py
import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .base import BaseHead @HEADS.register_module() class AudioTSNHead(BaseHead): """Classification head for TSN on audio. Args: num_classes (int): Number of classes to be classified. in_channels (int): Numbe...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/audio_tsn_head.py
import torch import torch.nn as nn from mmcv.cnn import normal_init from ..registry import HEADS from .tsn_head import TSNHead @HEADS.register_module() class TPNRPLHead(TSNHead): """Class head for TPN. Args: num_classes (int): Number of classes to be classified. in_channels (int): Number of ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tpn_rpl_head.py
import torch import torch.nn as nn from ..registry import HEADS from .base import AvgConsensus, BaseHead from ..builder import build_loss from .bnn import BayesianPredictor, get_uncertainty @HEADS.register_module() class TSMBNNHead(BaseHead): """Class head for TSM. Args: num_classes (int): Number of...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tsm_bnn_head.py
import torch.nn as nn import torch from mmcv.cnn import ConvModule, constant_init, normal_init, xavier_init import numpy as np from ..registry import HEADS from .base import BaseHead @HEADS.register_module() class DebiasHead(BaseHead): """Debias head. Args: num_classes (int): Number of classes to be ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/debias_head.py
from abc import ABCMeta, abstractmethod import torch import torch.nn as nn from ...core import top_k_accuracy from ..builder import build_loss class AvgConsensus(nn.Module): """Average consensus module. Args: dim (int): Decide which dim consensus function to apply. Default: 1. """ ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/heads/base.py
import torch.nn as nn from mmcv.cnn import CONV_LAYERS, build_norm_layer, constant_init, kaiming_init from torch.nn.modules.utils import _triple @CONV_LAYERS.register_module() class Conv2plus1d(nn.Module): """(2+1)d Conv module for R(2+1)d backbone. https://arxiv.org/pdf/1711.11248.pdf. Args: in...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/common/conv2plus1d.py
from .conv2plus1d import Conv2plus1d from .conv_audio import ConvAudio __all__ = ['Conv2plus1d', 'ConvAudio']
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/common/__init__.py
import torch import torch.nn as nn from mmcv.cnn import CONV_LAYERS, ConvModule, constant_init, kaiming_init from torch.nn.modules.utils import _pair @CONV_LAYERS.register_module() class ConvAudio(nn.Module): """Conv2d module for AudioResNet backbone. <https://arxiv.org/abs/2001.08740>`_. Args: ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/common/conv_audio.py
from ..registry import RECOGNIZERS from .base import BaseRecognizer import pdb @RECOGNIZERS.register_module() class Recognizer3D(BaseRecognizer): """3D recognizer model framework.""" def forward_train(self, imgs, labels, **kwargs): """Defines the computation performed at every call when training.""" ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer3d.py
from ..registry import RECOGNIZERS from .base import BaseRecognizer @RECOGNIZERS.register_module() class Recognizer2D(BaseRecognizer): """2D recognizer model framework.""" def forward_train(self, imgs, labels, **kwargs): """Defines the computation performed at every call when training.""" bat...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer2d.py
from ..registry import RECOGNIZERS from .recognizer3d import Recognizer3D @RECOGNIZERS.register_module() class Recognizer3DRPL(Recognizer3D): """3D recognizer model framework.""" def forward_train(self, imgs, labels, **kwargs): """Defines the computation performed at every call when training.""" ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer3d_rpl.py
from .audio_recognizer import AudioRecognizer from .base import BaseRecognizer from .recognizer2d import Recognizer2D from .recognizer3d import Recognizer3D from .recognizer2d_bnn import Recognizer2DBNN from .recognizer3d_bnn import Recognizer3DBNN from .recognizer2d_rpl import Recognizer2DRPL from .recognizer3d_rpl im...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/__init__.py
from ..registry import RECOGNIZERS from .recognizer2d import Recognizer2D @RECOGNIZERS.register_module() class Recognizer2DRPL(Recognizer2D): """2D recognizer model framework.""" def forward_train(self, imgs, labels, **kwargs): """Defines the computation performed at every call when training.""" ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer2d_rpl.py
from ..registry import RECOGNIZERS from .recognizer2d import Recognizer2D @RECOGNIZERS.register_module() class Recognizer2DBNN(Recognizer2D): """2D recognizer model framework.""" def forward_train(self, imgs, labels, **kwargs): """Defines the computation performed at every call when training.""" ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer2d_bnn.py
from abc import ABCMeta, abstractmethod from collections import OrderedDict import torch import torch.distributed as dist import torch.nn as nn import torch.nn.functional as F from mmcv.runner import auto_fp16 from .. import builder class BaseRecognizer(nn.Module, metaclass=ABCMeta): """Base class for recognize...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/base.py
from ..registry import RECOGNIZERS from .recognizer3d import Recognizer3D @RECOGNIZERS.register_module() class Recognizer3DBNN(Recognizer3D): """3D recognizer model framework.""" def forward_train(self, imgs, labels, **kwargs): """Defines the computation performed at every call when training.""" ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer3d_bnn.py
from ..registry import RECOGNIZERS from .base import BaseRecognizer @RECOGNIZERS.register_module() class AudioRecognizer(BaseRecognizer): """Audio recognizer model framework.""" def forward(self, audios, label=None, return_loss=True): """Define the computation performed at every call.""" if r...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/audio_recognizer.py
import torch.nn as nn import torch.utils.checkpoint as cp from mmcv.cnn import ConvModule, constant_init, kaiming_init from mmcv.runner import load_checkpoint from torch.nn.modules.batchnorm import _BatchNorm from torch.nn.modules.utils import _ntuple from mmaction.models.registry import BACKBONES from mmaction.utils ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet_audio.py
import torch.nn as nn import torch.utils.checkpoint as cp from mmcv.cnn import (ConvModule, NonLocal3d, build_activation_layer, constant_init, kaiming_init) from mmcv.runner import _load_checkpoint, load_checkpoint from mmcv.utils import _BatchNorm from torch.nn.modules.utils import _ntuple, _trip...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet3d.py
from functools import partial import imp from tkinter.messagebox import NO import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from timm.models.layers import drop_path, to_2tuple, trunc_normal_ from timm.models.registry import register_model from mmcv.runner import load_checkpoint,_loa...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/ViT3D.py
import math import torch.nn as nn import torch.utils.checkpoint as cp from mmcv.cnn import (ConvModule, Swish, build_activation_layer, constant_init, kaiming_init) from mmcv.runner import load_checkpoint from mmcv.utils import _BatchNorm from ...utils import get_root_logger from ..registry impor...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/x3d.py
import torch.nn as nn from mmcv.cnn import ConvModule from mmcv.utils import _BatchNorm from ..registry import BACKBONES from .resnet3d import Bottleneck3d, ResNet3d class CSNBottleneck3d(Bottleneck3d): """Channel-Separated Bottleneck Block. This module is proposed in "Video Classification with Channel-...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet3d_csn.py
from .c3d import C3D from .resnet import ResNet from .resnet2plus1d import ResNet2Plus1d from .resnet3d import ResNet3d from .resnet3d_csn import ResNet3dCSN from .resnet3d_slowfast import ResNet3dSlowFast from .resnet3d_slowonly import ResNet3dSlowOnly from .resnet_audio import ResNetAudio from .resnet_tin import ResN...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/__init__.py
import warnings import torch import torch.nn as nn from ..registry import BACKBONES from .resnet_tsm import ResNetTSM try: from mmcv.ops import tin_shift except (ImportError, ModuleNotFoundError): warnings.warn('Please install mmcv-full to support "tin_shift"') def linear_sampler(data, offset): """Diff...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet_tin.py
import torch import torch.nn as nn from mmcv.cnn import NonLocal3d from torch.nn.modules.utils import _ntuple from ..registry import BACKBONES from .resnet import ResNet class NL3DWrapper(nn.Module): """3D Non-local wrapper for ResNet50. Wrap ResNet layers with 3D NonLocal modules. Args: block ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet_tsm.py
import torch.nn as nn from mmcv.cnn import ConvModule, constant_init, kaiming_init from mmcv.runner import _load_checkpoint, load_checkpoint from mmcv.utils import _BatchNorm from torch.utils import checkpoint as cp from ...utils import get_root_logger from ..registry import BACKBONES class BasicBlock(nn.Module): ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet.py
from ..registry import BACKBONES from .resnet3d import ResNet3d @BACKBONES.register_module() class ResNet2Plus1d(ResNet3d): """ResNet (2+1)d backbone. This model is proposed in `A Closer Look at Spatiotemporal Convolutions for Action Recognition <https://arxiv.org/abs/1711.11248>`_ """ def __ini...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet2plus1d.py
import torch import torch.nn as nn from mmcv.cnn import ConvModule, kaiming_init from mmcv.runner import _load_checkpoint, load_checkpoint from mmcv.utils import print_log from ...utils import get_root_logger from ..registry import BACKBONES from .resnet3d import ResNet3d class ResNet3dPathway(ResNet3d): """A pa...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet3d_slowfast.py
from ..registry import BACKBONES from .resnet3d_slowfast import ResNet3dPathway @BACKBONES.register_module() class ResNet3dSlowOnly(ResNet3dPathway): """SlowOnly backbone based on ResNet3dPathway. Args: *args (arguments): Arguments same as :class:`ResNet3dPathway`. conv1_kernel (Sequence[int]...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet3d_slowonly.py
import torch.nn as nn from mmcv.cnn import ConvModule, constant_init, kaiming_init, normal_init from mmcv.runner import load_checkpoint from mmcv.utils import _BatchNorm from ...utils import get_root_logger from ..registry import BACKBONES @BACKBONES.register_module() class C3D(nn.Module): """C3D backbone. ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/c3d.py
import torch import torch.nn as nn from .. import builder from ..registry import LOCALIZERS from .base import BaseLocalizer @LOCALIZERS.register_module() class SSN(BaseLocalizer): """Temporal Action Detection with Structured Segment Networks. Args: backbone (dict): Config for building backbone. ...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/localizers/ssn.py
import math import numpy as np import torch import torch.nn as nn from ...localization import temporal_iop, temporal_iou from ..builder import build_loss from ..registry import LOCALIZERS from .base import BaseLocalizer from .utils import post_processing @LOCALIZERS.register_module() class BMN(BaseLocalizer): "...
InternVideo-main
Downstream/Open-Set-Action-Recognition/mmaction/models/localizers/bmn.py