python_code stringlengths 0 992k | repo_name stringlengths 8 46 | file_path stringlengths 5 162 |
|---|---|---|
from .tin_lr_hook import TINLrUpdaterHook
__all__ = ['TINLrUpdaterHook']
| InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/core/lr/__init__.py |
from mmcv.runner import HOOKS, LrUpdaterHook
from mmcv.runner.hooks.lr_updater import annealing_cos
@HOOKS.register_module()
class TINLrUpdaterHook(LrUpdaterHook):
def __init__(self, min_lr, **kwargs):
self.min_lr = min_lr
super(TINLrUpdaterHook, self).__init__(**kwargs)
def get_warmup_lr(se... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/core/lr/tin_lr_hook.py |
import os.path as osp
import torch
from .base import BaseDataset
from .registry import DATASETS
@DATASETS.register_module()
class AudioDataset(BaseDataset):
"""Audio dataset for video recognition. Extracts the audio feature on-the-
fly. Annotation file can be that of the rawframe dataset, or:
.. code-b... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/audio_dataset.py |
import copy
import os
import os.path as osp
import warnings
import mmcv
import numpy as np
from ..core import average_recall_at_avg_proposals
from .base import BaseDataset
from .registry import DATASETS
@DATASETS.register_module()
class ActivityNetDataset(BaseDataset):
"""ActivityNet dataset for temporal action... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/activitynet_dataset.py |
import copy
import os.path as osp
import numpy as np
import torch
from .base import BaseDataset
from .registry import DATASETS
@DATASETS.register_module()
class RawframeDataset(BaseDataset):
"""Rawframe dataset for action recognition.
The dataset loads raw frames and apply specified transforms to return a
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/rawframe_dataset.py |
import copy
import os.path as osp
import mmcv
import numpy as np
from mmcv.utils import print_log
from ..core import mean_average_precision
from .base import BaseDataset
from .registry import DATASETS
@DATASETS.register_module()
class HVUDataset(BaseDataset):
"""HVU dataset, which supports the recognition tags ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/hvu_dataset.py |
from .registry import DATASETS
from .video_dataset import VideoDataset
@DATASETS.register_module()
class ImageDataset(VideoDataset):
"""Image dataset for action recognition, used in the Project OmniSource.
The dataset loads image list and apply specified transforms to return a
dict containing the image t... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/image_dataset.py |
from mmcv.utils import Registry
DATASETS = Registry('dataset')
PIPELINES = Registry('pipeline')
| InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/registry.py |
import os.path as osp
import torch
from .base import BaseDataset
from .registry import DATASETS
@DATASETS.register_module()
class AudioFeatureDataset(BaseDataset):
"""Audio feature dataset for video recognition. Reads the features
extracted off-line. Annotation file can be that of the rawframe dataset,
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/audio_feature_dataset.py |
from .activitynet_dataset import ActivityNetDataset
from .audio_dataset import AudioDataset
from .audio_feature_dataset import AudioFeatureDataset
from .audio_visual_dataset import AudioVisualDataset
from .ava_dataset import AVADataset
from .base import BaseDataset
from .builder import build_dataloader, build_dataset
f... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/__init__.py |
import platform
import random
from functools import partial
import numpy as np
from mmcv.parallel import collate
from mmcv.runner import get_dist_info
from mmcv.utils import build_from_cfg
from torch.utils.data import DataLoader
from .dataset_wrappers import RepeatDataset
from .registry import DATASETS
from .samplers... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/builder.py |
import os.path as osp
from .rawframe_dataset import RawframeDataset
from .registry import DATASETS
@DATASETS.register_module()
class AudioVisualDataset(RawframeDataset):
"""Dataset that reads both audio and visual data, supporting both rawframes
and videos. The annotation file is same as that of the rawframe... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/audio_visual_dataset.py |
from .registry import DATASETS
@DATASETS.register_module()
class RepeatDataset:
"""A wrapper of repeated dataset.
The length of repeated dataset will be ``times`` larger than the original
dataset. This is useful when the data loading time is long but the dataset
is small. Using RepeatDataset can redu... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/dataset_wrappers.py |
import os.path as osp
import torch
from .base import BaseDataset
from .registry import DATASETS
@DATASETS.register_module()
class VideoDataset(BaseDataset):
"""Video dataset for action recognition.
The dataset loads raw videos and apply specified transforms to return a
dict containing the frame tensors... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/video_dataset.py |
import copy
import os.path as osp
import warnings
import mmcv
import numpy as np
from torch.nn.modules.utils import _pair
from ..core import softmax
from ..localization import (eval_ap, load_localize_proposal_file,
perform_regression, temporal_iou, temporal_nms)
from ..utils import get_roo... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/ssn_dataset.py |
import copy
import os.path as osp
from collections import defaultdict
import mmcv
import numpy as np
from ..utils import get_root_logger
from .base import BaseDataset
from .registry import DATASETS
@DATASETS.register_module()
class AVADataset(BaseDataset):
"""AVA dataset for spatial temporal detection.
Bas... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/ava_dataset.py |
import copy
import os.path as osp
import warnings
from abc import ABCMeta, abstractmethod
from collections import defaultdict
import mmcv
import numpy as np
import torch
from mmcv.utils import print_log
from torch.utils.data import Dataset
from ..core import (mean_average_precision, mean_class_accuracy,
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/base.py |
import copy
import os.path as osp
import random
import mmcv
from .base import BaseDataset
from .registry import DATASETS
@DATASETS.register_module()
class RawVideoDataset(BaseDataset):
"""RawVideo dataset for action recognition, used in the Project OmniSource.
The dataset loads clips of raw videos and appl... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/rawvideo_dataset.py |
import io
import os
import os.path as osp
import shutil
import warnings
import mmcv
import numpy as np
import torch
from mmcv.fileio import FileClient
from torch.nn.modules.utils import _pair
from ...utils import get_random_string, get_shm_dir, get_thread_id
from ..registry import PIPELINES
@PIPELINES.register_modu... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/loading.py |
from collections.abc import Sequence
from mmcv.utils import build_from_cfg
from ..registry import PIPELINES
@PIPELINES.register_module()
class Compose:
"""Compose a data pipeline with a sequence of transforms.
Args:
transforms (list[dict | callable]):
Either config dicts of transforms o... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/compose.py |
from .augmentations import (AudioAmplify, CenterCrop, ColorJitter,
EntityBoxClip, EntityBoxCrop, EntityBoxFlip,
EntityBoxPad, EntityBoxRescale, Flip, Fuse,
MelSpectrogram, MultiGroupCrop, MultiScaleCrop,
Norm... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/__init__.py |
import random
from collections.abc import Sequence
import mmcv
import numpy as np
from torch.nn.modules.utils import _pair
from ..registry import PIPELINES
def _init_lazy_if_proper(results, lazy):
"""Initialize lazy operation properly.
Make sure that a lazy operation is properly initialized,
and avoid ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/augmentations.py |
from collections.abc import Sequence
import mmcv
import numpy as np
import torch
from mmcv.parallel import DataContainer as DC
from ..registry import PIPELINES
def to_tensor(data):
"""Convert objects of various python types to :obj:`torch.Tensor`.
Supported types are: :class:`numpy.ndarray`, :class:`torch.... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/pipelines/formating.py |
from .distributed_sampler import DistributedPowerSampler, DistributedSampler
__all__ = ['DistributedSampler', 'DistributedPowerSampler']
| InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/samplers/__init__.py |
import torch
from torch.utils.data import DistributedSampler as _DistributedSampler
class DistributedSampler(_DistributedSampler):
"""DistributedSampler inheriting from
``torch.utils.data.DistributedSampler``.
In pytorch of lower versions, there is no ``shuffle`` argument. This child
class will port ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/datasets/samplers/distributed_sampler.py |
import ctypes
import random
import string
def get_random_string(length=15):
"""Get random string with letters and digits.
Args:
length (int): Length of random string. Default: 15.
"""
return ''.join(
random.choice(string.ascii_letters + string.digits)
for _ in range(length))
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/utils/misc.py |
from mmcv.utils import collect_env as collect_basic_env
from mmcv.utils import get_git_hash
import mmaction
def collect_env():
env_info = collect_basic_env()
env_info['MMAction2'] = (
mmaction.__version__ + '+' + get_git_hash(digits=7))
return env_info
if __name__ == '__main__':
for name, v... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/utils/collect_env.py |
from .collect_env import collect_env
from .gradcam_utils import GradCAM
from .logger import get_root_logger
from .misc import get_random_string, get_shm_dir, get_thread_id
__all__ = [
'get_root_logger', 'collect_env', 'get_random_string', 'get_thread_id',
'get_shm_dir', 'GradCAM'
]
| InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/utils/__init__.py |
import logging
from mmcv.utils import get_logger
def get_root_logger(log_file=None, log_level=logging.INFO):
"""Use ``get_logger`` method in mmcv to get the root logger.
The logger will be initialized if it has not been initialized. By default a
StreamHandler will be added. If ``log_file`` is specified,... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/utils/logger.py |
import torch
import torch.nn.functional as F
class GradCAM:
"""GradCAM class helps create visualization results.
Visualization results are blended by heatmaps and input images.
This class is modified from
https://github.com/facebookresearch/SlowFast/blob/master/slowfast/visualization/gradcam_utils.py... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/utils/gradcam_utils.py |
from mmcv.utils import Registry
BACKBONES = Registry('backbone')
NECKS = Registry('neck')
HEADS = Registry('head')
RECOGNIZERS = Registry('recognizer')
LOSSES = Registry('loss')
LOCALIZERS = Registry('localizer')
| InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/registry.py |
from .backbones import (C3D, X3D, ResNet, ResNet2Plus1d, ResNet3d, ResNet3dCSN,
ResNet3dSlowFast, ResNet3dSlowOnly, ResNetAudio,
ResNetTIN, ResNetTSM)
from .builder import (build_backbone, build_head, build_localizer, build_loss,
build_model, build_n... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/__init__.py |
import torch.nn as nn
from mmcv.utils import build_from_cfg
from .registry import BACKBONES, HEADS, LOCALIZERS, LOSSES, NECKS, RECOGNIZERS
def build(cfg, registry, default_args=None):
"""Build a module.
Args:
cfg (dict, list[dict]): The config of modules, it is either a dict
or a list of... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/builder.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
from ..registry import LOSSES
from ...core import top_k_accuracy
@LOSSES.register_module()
class BayesianNNLoss(nn.Module):
"""Bayesian NN Loss."""
def forward(self, cls_score, labels, output_dict, beta=1.0, **kwargs):
"""Forward func... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/bnn_loss.py |
import torch
import torch.nn.functional as F
from ..registry import LOSSES
from .base import BaseWeightedLoss
def relu_evidence(y):
return F.relu(y)
def exp_evidence(y):
return torch.exp(torch.clamp(y, -10, 10))
def softplus_evidence(y):
return F.softplus(y)
@LOSSES.register_module()
class EvidenceLoss... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/edl_loss.py |
import torch
import torch.nn.functional as F
import numpy as np
from ..registry import LOSSES
from .base import BaseWeightedLoss
@LOSSES.register_module()
class RebiasLoss(BaseWeightedLoss):
"""Rebias Loss."""
def __init__(self, lambda_g=1.0, criteria='hsic'):
super().__init__()
self.lambda_g... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/rebias_loss.py |
from .base import BaseWeightedLoss
from .binary_logistic_regression_loss import BinaryLogisticRegressionLoss
from .bmn_loss import BMNLoss
from .cross_entropy_loss import BCELossWithLogits, CrossEntropyLoss
from .bnn_loss import BayesianNNLoss
from .edl_loss import EvidenceLoss
from .hvu_loss import HVULoss
from .nll_l... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/__init__.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
from ..registry import LOSSES
from .ohem_hinge_loss import OHEMHingeLoss
@LOSSES.register_module()
class SSNLoss(nn.Module):
@staticmethod
def activity_loss(activity_score, labels, activity_indexer):
"""Activity Loss.
It wil... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/ssn_loss.py |
import torch
import torch.nn.functional as F
from ..registry import LOSSES
from .base import BaseWeightedLoss
from ...core import top_k_accuracy
@LOSSES.register_module()
class GCPLoss(BaseWeightedLoss):
"""Reciprocal Point Learning Loss."""
def __init__(self, temperature=1, weight_pl=0.1, radius_init=1):
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/gcp_loss.py |
import torch
import torch.nn as nn
from ..registry import LOSSES
def binary_logistic_regression_loss(reg_score,
label,
threshold=0.5,
ratio_range=(1.05, 21),
eps=1e-5):
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/binary_logistic_regression_loss.py |
import torch.nn.functional as F
from ..registry import LOSSES
from .base import BaseWeightedLoss
@LOSSES.register_module()
class CrossEntropyLoss(BaseWeightedLoss):
"""Cross Entropy Loss."""
def _forward(self, cls_score, label, **kwargs):
"""Forward function.
Args:
cls_score (to... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/cross_entropy_loss.py |
import torch
class OHEMHingeLoss(torch.autograd.Function):
"""This class is the core implementation for the completeness loss in
paper.
It compute class-wise hinge loss and performs online hard example mining
(OHEM).
"""
@staticmethod
def forward(ctx, pred, labels, is_positive, ohem_rati... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/ohem_hinge_loss.py |
import torch.nn.functional as F
from ..registry import LOSSES
from .base import BaseWeightedLoss
@LOSSES.register_module()
class NLLLoss(BaseWeightedLoss):
"""NLL Loss.
It will calculate NLL loss given cls_score and label.
"""
def _forward(self, cls_score, label, **kwargs):
"""Forward funct... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/nll_loss.py |
import torch
import torch.nn.functional as F
from ..registry import LOSSES
from .base import BaseWeightedLoss
@LOSSES.register_module()
class HVULoss(BaseWeightedLoss):
"""Calculate the BCELoss for HVU.
Args:
categories (tuple[str]): Names of tag categories, tags are organized in
this or... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/hvu_loss.py |
from abc import ABCMeta, abstractmethod
import torch.nn as nn
class BaseWeightedLoss(nn.Module, metaclass=ABCMeta):
"""Base class for loss.
All subclass should overwrite the ``_forward()`` method which returns the
normal loss without loss weights.
Args:
loss_weight (float): Factor scalar mu... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/base.py |
import torch
import torch.nn.functional as F
from ..registry import LOSSES
from .base import BaseWeightedLoss
from ...core import top_k_accuracy
@LOSSES.register_module()
class RPLoss(BaseWeightedLoss):
"""Reciprocal Point Learning Loss."""
def __init__(self, temperature=1, weight_pl=0.1, radius_init=1):
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/rpl_loss.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
from ..registry import LOSSES
from .binary_logistic_regression_loss import binary_logistic_regression_loss
@LOSSES.register_module()
class BMNLoss(nn.Module):
"""BMN Loss.
From paper https://arxiv.org/abs/1907.09702,
code https://github.... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/losses/bmn_loss.py |
from .tpn import TPN
__all__ = ['TPN']
| InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/necks/__init__.py |
import numpy as np
import torch
import torch.nn as nn
from mmcv.cnn import ConvModule, constant_init, normal_init, xavier_init
from ..heads import AuxHead, RebiasHead
from ..registry import NECKS
class Identity(nn.Module):
"""Identity mapping."""
def forward(self, x):
return x
class DownSample(nn.M... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/necks/tpn.py |
import torch
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import BaseHead
@HEADS.register_module()
class SlowFastHead(BaseHead):
"""The classification head for SlowFast.
Args:
num_classes (int): Number of classes to be classified.
in_channels... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/slowfast_head.py |
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import BaseHead
@HEADS.register_module()
class X3DHead(BaseHead):
"""Classification head for I3D.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int): Number of channels ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/x3d_head.py |
import torch
import torch.nn as nn
from mmcv.cnn import ConvModule, constant_init, normal_init, xavier_init
from ..builder import build_loss
class AuxHead(nn.Module):
"""Auxiliary Head.
This auxiliary head is appended to receive stronger supervision,
leading to enhanced semantics.
Args:
in_c... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/aux_head.py |
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import AvgConsensus, BaseHead
@HEADS.register_module()
class TSNHead(BaseHead):
"""Class head for TSN.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int): Number of chan... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tsn_head.py |
import torch
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import BaseHead
@HEADS.register_module()
class SlowFastRPLHead(BaseHead):
"""The classification head for SlowFast.
Args:
num_classes (int): Number of classes to be classified.
in_chann... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/slowfast_rpl_head.py |
import torch
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import BaseHead
@HEADS.register_module()
class I3DRPLHead(BaseHead):
"""Classification head for I3D.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int): Num... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/i3d_rpl_head.py |
import torch
import torch.nn as nn
from mmcv.cnn import ConvModule, constant_init, normal_init, xavier_init
from ..builder import build_loss
class RebiasHead(nn.Module):
def __init__(self,
in_channels,
out_channels,
loss_weight=0.5,
loss_rebias=d... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/rebias_head.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
import math
class Gaussian(object):
def __init__(self, mu, rho):
super().__init__()
self.mu = mu
self.rho = rho
self.normal = torch.distributions.Normal(0,1)
@property
def sigma(self):
return tor... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/bnn.py |
import torch
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
def parse_stage_config(stage_cfg):
"""Parse config of STPP for three stages.
Args:
stage_cfg (int | tuple[int]):
Config of structured temporal pyramid pooling.
Returns:
tuple[tup... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/ssn_head.py |
from .audio_tsn_head import AudioTSNHead
from .base import BaseHead
from .i3d_head import I3DHead
from .i3d_bnn_head import I3DBNNHead
from .i3d_rpl_head import I3DRPLHead
from .slowfast_head import SlowFastHead
from .slowfast_rpl_head import SlowFastRPLHead
from .slowfast_bnn_head import SlowFastBNNHead
from .ssn_head... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/__init__.py |
import torch.nn as nn
from ..registry import HEADS
from .tsn_head import TSNHead
from ..builder import build_loss
from .bnn import BayesianPredictor, get_uncertainty
@HEADS.register_module()
class TPNBNNHead(TSNHead):
"""Class head for TPN.
Args:
num_classes (int): Number of classes to be classified... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tpn_bnn_head.py |
import torch
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import AvgConsensus, BaseHead
@HEADS.register_module()
class TSMRPLHead(BaseHead):
"""Class head for TSM.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int)... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tsm_rpl_head.py |
import torch
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import AvgConsensus, BaseHead
@HEADS.register_module()
class TSMHead(BaseHead):
"""Class head for TSM.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int): N... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tsm_head.py |
import torch
import torch.nn as nn
from ..registry import HEADS
from .base import BaseHead
from ..builder import build_loss
from .bnn import BayesianPredictor, get_uncertainty
@HEADS.register_module()
class I3DBNNHead(BaseHead):
"""Classification head for I3D.
Args:
num_classes (int): Number of clas... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/i3d_bnn_head.py |
import torch.nn as nn
from ..registry import HEADS
from .tsn_head import TSNHead
@HEADS.register_module()
class TPNHead(TSNHead):
"""Class head for TPN.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int): Number of channels in input feature.
loss_cls (dict... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tpn_head.py |
# Copyright (c) OpenMMLab. All rights reserved.
import torch
import torch.nn as nn
from mmcv.cnn import normal_init
from ..builder import HEADS
from .base import BaseHead
import pdb
@HEADS.register_module()
class BaseClsHead(BaseHead):
"""The classification head for SlowFast.
Args:
num_classes (int):... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/base_cls_head.py |
import torch
import torch.nn as nn
class Dist(nn.Module):
def __init__(self, num_classes=10, num_centers=1, feat_dim=2, init='random'):
super(Dist, self).__init__()
self.feat_dim = feat_dim
self.num_classes = num_classes
self.num_centers = num_centers
if init == 'random':
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/rpl_dist.py |
import torch
import torch.nn as nn
from ..registry import HEADS
from .base import BaseHead
from ..builder import build_loss
from .bnn import BayesianPredictor, get_uncertainty
@HEADS.register_module()
class SlowFastBNNHead(BaseHead):
"""The classification head for SlowFast.
Args:
num_classes (int): ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/slowfast_bnn_head.py |
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import BaseHead
@HEADS.register_module()
class I3DHead(BaseHead):
"""Classification head for I3D.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int): Number of channels ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/i3d_head.py |
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .base import BaseHead
@HEADS.register_module()
class AudioTSNHead(BaseHead):
"""Classification head for TSN on audio.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int): Numbe... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/audio_tsn_head.py |
import torch
import torch.nn as nn
from mmcv.cnn import normal_init
from ..registry import HEADS
from .tsn_head import TSNHead
@HEADS.register_module()
class TPNRPLHead(TSNHead):
"""Class head for TPN.
Args:
num_classes (int): Number of classes to be classified.
in_channels (int): Number of ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tpn_rpl_head.py |
import torch
import torch.nn as nn
from ..registry import HEADS
from .base import AvgConsensus, BaseHead
from ..builder import build_loss
from .bnn import BayesianPredictor, get_uncertainty
@HEADS.register_module()
class TSMBNNHead(BaseHead):
"""Class head for TSM.
Args:
num_classes (int): Number of... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/tsm_bnn_head.py |
import torch.nn as nn
import torch
from mmcv.cnn import ConvModule, constant_init, normal_init, xavier_init
import numpy as np
from ..registry import HEADS
from .base import BaseHead
@HEADS.register_module()
class DebiasHead(BaseHead):
"""Debias head.
Args:
num_classes (int): Number of classes to be ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/debias_head.py |
from abc import ABCMeta, abstractmethod
import torch
import torch.nn as nn
from ...core import top_k_accuracy
from ..builder import build_loss
class AvgConsensus(nn.Module):
"""Average consensus module.
Args:
dim (int): Decide which dim consensus function to apply.
Default: 1.
"""
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/heads/base.py |
import torch.nn as nn
from mmcv.cnn import CONV_LAYERS, build_norm_layer, constant_init, kaiming_init
from torch.nn.modules.utils import _triple
@CONV_LAYERS.register_module()
class Conv2plus1d(nn.Module):
"""(2+1)d Conv module for R(2+1)d backbone.
https://arxiv.org/pdf/1711.11248.pdf.
Args:
in... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/common/conv2plus1d.py |
from .conv2plus1d import Conv2plus1d
from .conv_audio import ConvAudio
__all__ = ['Conv2plus1d', 'ConvAudio']
| InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/common/__init__.py |
import torch
import torch.nn as nn
from mmcv.cnn import CONV_LAYERS, ConvModule, constant_init, kaiming_init
from torch.nn.modules.utils import _pair
@CONV_LAYERS.register_module()
class ConvAudio(nn.Module):
"""Conv2d module for AudioResNet backbone.
<https://arxiv.org/abs/2001.08740>`_.
Args:
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/common/conv_audio.py |
from ..registry import RECOGNIZERS
from .base import BaseRecognizer
import pdb
@RECOGNIZERS.register_module()
class Recognizer3D(BaseRecognizer):
"""3D recognizer model framework."""
def forward_train(self, imgs, labels, **kwargs):
"""Defines the computation performed at every call when training."""
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer3d.py |
from ..registry import RECOGNIZERS
from .base import BaseRecognizer
@RECOGNIZERS.register_module()
class Recognizer2D(BaseRecognizer):
"""2D recognizer model framework."""
def forward_train(self, imgs, labels, **kwargs):
"""Defines the computation performed at every call when training."""
bat... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer2d.py |
from ..registry import RECOGNIZERS
from .recognizer3d import Recognizer3D
@RECOGNIZERS.register_module()
class Recognizer3DRPL(Recognizer3D):
"""3D recognizer model framework."""
def forward_train(self, imgs, labels, **kwargs):
"""Defines the computation performed at every call when training."""
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer3d_rpl.py |
from .audio_recognizer import AudioRecognizer
from .base import BaseRecognizer
from .recognizer2d import Recognizer2D
from .recognizer3d import Recognizer3D
from .recognizer2d_bnn import Recognizer2DBNN
from .recognizer3d_bnn import Recognizer3DBNN
from .recognizer2d_rpl import Recognizer2DRPL
from .recognizer3d_rpl im... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/__init__.py |
from ..registry import RECOGNIZERS
from .recognizer2d import Recognizer2D
@RECOGNIZERS.register_module()
class Recognizer2DRPL(Recognizer2D):
"""2D recognizer model framework."""
def forward_train(self, imgs, labels, **kwargs):
"""Defines the computation performed at every call when training."""
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer2d_rpl.py |
from ..registry import RECOGNIZERS
from .recognizer2d import Recognizer2D
@RECOGNIZERS.register_module()
class Recognizer2DBNN(Recognizer2D):
"""2D recognizer model framework."""
def forward_train(self, imgs, labels, **kwargs):
"""Defines the computation performed at every call when training."""
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer2d_bnn.py |
from abc import ABCMeta, abstractmethod
from collections import OrderedDict
import torch
import torch.distributed as dist
import torch.nn as nn
import torch.nn.functional as F
from mmcv.runner import auto_fp16
from .. import builder
class BaseRecognizer(nn.Module, metaclass=ABCMeta):
"""Base class for recognize... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/base.py |
from ..registry import RECOGNIZERS
from .recognizer3d import Recognizer3D
@RECOGNIZERS.register_module()
class Recognizer3DBNN(Recognizer3D):
"""3D recognizer model framework."""
def forward_train(self, imgs, labels, **kwargs):
"""Defines the computation performed at every call when training."""
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/recognizer3d_bnn.py |
from ..registry import RECOGNIZERS
from .base import BaseRecognizer
@RECOGNIZERS.register_module()
class AudioRecognizer(BaseRecognizer):
"""Audio recognizer model framework."""
def forward(self, audios, label=None, return_loss=True):
"""Define the computation performed at every call."""
if r... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/recognizers/audio_recognizer.py |
import torch.nn as nn
import torch.utils.checkpoint as cp
from mmcv.cnn import ConvModule, constant_init, kaiming_init
from mmcv.runner import load_checkpoint
from torch.nn.modules.batchnorm import _BatchNorm
from torch.nn.modules.utils import _ntuple
from mmaction.models.registry import BACKBONES
from mmaction.utils ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet_audio.py |
import torch.nn as nn
import torch.utils.checkpoint as cp
from mmcv.cnn import (ConvModule, NonLocal3d, build_activation_layer,
constant_init, kaiming_init)
from mmcv.runner import _load_checkpoint, load_checkpoint
from mmcv.utils import _BatchNorm
from torch.nn.modules.utils import _ntuple, _trip... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet3d.py |
from functools import partial
import imp
from tkinter.messagebox import NO
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
from timm.models.layers import drop_path, to_2tuple, trunc_normal_
from timm.models.registry import register_model
from mmcv.runner import load_checkpoint,_loa... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/ViT3D.py |
import math
import torch.nn as nn
import torch.utils.checkpoint as cp
from mmcv.cnn import (ConvModule, Swish, build_activation_layer, constant_init,
kaiming_init)
from mmcv.runner import load_checkpoint
from mmcv.utils import _BatchNorm
from ...utils import get_root_logger
from ..registry impor... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/x3d.py |
import torch.nn as nn
from mmcv.cnn import ConvModule
from mmcv.utils import _BatchNorm
from ..registry import BACKBONES
from .resnet3d import Bottleneck3d, ResNet3d
class CSNBottleneck3d(Bottleneck3d):
"""Channel-Separated Bottleneck Block.
This module is proposed in
"Video Classification with Channel-... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet3d_csn.py |
from .c3d import C3D
from .resnet import ResNet
from .resnet2plus1d import ResNet2Plus1d
from .resnet3d import ResNet3d
from .resnet3d_csn import ResNet3dCSN
from .resnet3d_slowfast import ResNet3dSlowFast
from .resnet3d_slowonly import ResNet3dSlowOnly
from .resnet_audio import ResNetAudio
from .resnet_tin import ResN... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/__init__.py |
import warnings
import torch
import torch.nn as nn
from ..registry import BACKBONES
from .resnet_tsm import ResNetTSM
try:
from mmcv.ops import tin_shift
except (ImportError, ModuleNotFoundError):
warnings.warn('Please install mmcv-full to support "tin_shift"')
def linear_sampler(data, offset):
"""Diff... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet_tin.py |
import torch
import torch.nn as nn
from mmcv.cnn import NonLocal3d
from torch.nn.modules.utils import _ntuple
from ..registry import BACKBONES
from .resnet import ResNet
class NL3DWrapper(nn.Module):
"""3D Non-local wrapper for ResNet50.
Wrap ResNet layers with 3D NonLocal modules.
Args:
block ... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet_tsm.py |
import torch.nn as nn
from mmcv.cnn import ConvModule, constant_init, kaiming_init
from mmcv.runner import _load_checkpoint, load_checkpoint
from mmcv.utils import _BatchNorm
from torch.utils import checkpoint as cp
from ...utils import get_root_logger
from ..registry import BACKBONES
class BasicBlock(nn.Module):
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet.py |
from ..registry import BACKBONES
from .resnet3d import ResNet3d
@BACKBONES.register_module()
class ResNet2Plus1d(ResNet3d):
"""ResNet (2+1)d backbone.
This model is proposed in `A Closer Look at Spatiotemporal Convolutions for
Action Recognition <https://arxiv.org/abs/1711.11248>`_
"""
def __ini... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet2plus1d.py |
import torch
import torch.nn as nn
from mmcv.cnn import ConvModule, kaiming_init
from mmcv.runner import _load_checkpoint, load_checkpoint
from mmcv.utils import print_log
from ...utils import get_root_logger
from ..registry import BACKBONES
from .resnet3d import ResNet3d
class ResNet3dPathway(ResNet3d):
"""A pa... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet3d_slowfast.py |
from ..registry import BACKBONES
from .resnet3d_slowfast import ResNet3dPathway
@BACKBONES.register_module()
class ResNet3dSlowOnly(ResNet3dPathway):
"""SlowOnly backbone based on ResNet3dPathway.
Args:
*args (arguments): Arguments same as :class:`ResNet3dPathway`.
conv1_kernel (Sequence[int]... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/resnet3d_slowonly.py |
import torch.nn as nn
from mmcv.cnn import ConvModule, constant_init, kaiming_init, normal_init
from mmcv.runner import load_checkpoint
from mmcv.utils import _BatchNorm
from ...utils import get_root_logger
from ..registry import BACKBONES
@BACKBONES.register_module()
class C3D(nn.Module):
"""C3D backbone.
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/backbones/c3d.py |
import torch
import torch.nn as nn
from .. import builder
from ..registry import LOCALIZERS
from .base import BaseLocalizer
@LOCALIZERS.register_module()
class SSN(BaseLocalizer):
"""Temporal Action Detection with Structured Segment Networks.
Args:
backbone (dict): Config for building backbone.
... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/localizers/ssn.py |
import math
import numpy as np
import torch
import torch.nn as nn
from ...localization import temporal_iop, temporal_iou
from ..builder import build_loss
from ..registry import LOCALIZERS
from .base import BaseLocalizer
from .utils import post_processing
@LOCALIZERS.register_module()
class BMN(BaseLocalizer):
"... | InternVideo-main | Downstream/Open-Set-Action-Recognition/mmaction/models/localizers/bmn.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.