python_code stringlengths 0 992k | repo_name stringlengths 8 46 | file_path stringlengths 5 162 |
|---|---|---|
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
import math
import os
import torch
from mmf.common.registry import registry
from mmf.models import BaseModel
from mmf.modules.embeddings import BertVisioLinguisticEmbeddings
from mmf.utils.configuration import get_mmf_cache_dir
from mmf.utils.file_io i... | EXA-1-master | exa/models/mmf-main/mmf/models/krisp.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# MMBTModel, ModalEmbeddings is copied from [1]
# as we have internal dependency on transformers v2.3.
# These will be removed when we upgrade to package v2.5+.
# [1]: https://github.com/huggingface/transformers/blob/master/src/transformers/modeling_mmbt.py # noqa
im... | EXA-1-master | exa/models/mmf-main/mmf/models/mmbt.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import torch
from mmf.common.registry import registry
from mmf.models.pythia import Pythia
from mmf.modules.layers import ClassifierLayer
@registry.register_model("butd")
class BUTD(Pythia):
def __init__(self, config):
super().__init__(config)
@clas... | EXA-1-master | exa/models/mmf-main/mmf/models/butd.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import functools
import logging
import math
import torch
import torch.nn.functional as F
from mmf.common.registry import registry
from mmf.models.base_model import BaseModel
from mmf.modules.layers import ClassifierLayer
from mmf.utils.build import build_image_encoder... | EXA-1-master | exa/models/mmf-main/mmf/models/m4c.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import copy
import omegaconf
import torch
from mmf.common.registry import registry
from mmf.models.base_model import BaseModel
from mmf.modules.embeddings import (
ImageFeatureEmbedding,
MultiHeadImageFeatureEmbedding,
PreExtractedEmbedding,
TextEmbedd... | EXA-1-master | exa/models/mmf-main/mmf/models/pythia.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Mostly copy-pasted from
# https://github.com/facebookresearch/detr/blob/master/util/misc.py
from typing import List
import torch
from torch import Tensor
class NestedTensor:
"""
A data class to hold images of different sizes in a batch.
It contains `... | EXA-1-master | exa/models/mmf-main/mmf/models/unit/misc.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Mostly copy-pasted from
# https://github.com/facebookresearch/detr/blob/master/models/matcher.py
"""
Modules to compute the matching cost and solve the corresponding LSAP.
"""
from typing import Dict, List
import torch
from mmf.utils.box_ops import box_cxcywh_to_xy... | EXA-1-master | exa/models/mmf-main/mmf/models/unit/matcher.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Mostly copy-pasted from
# https://github.com/facebookresearch/detr/blob/master/models/backbone.py
"""
Backbone modules.
"""
import math
from collections import OrderedDict
import torch
import torch.nn.functional as F
import torchvision
from mmf.models.unit.misc imp... | EXA-1-master | exa/models/mmf-main/mmf/models/unit/backbone.py |
# Copyright (c) Facebook, Inc. and its affiliates.
__all__ = ["UniT"]
from .unit import UniT
| EXA-1-master | exa/models/mmf-main/mmf/models/unit/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Mostly copy-pasted from
# https://github.com/facebookresearch/detr/blob/master/models/transformer.py
import copy
from typing import Optional
import torch
import torch.nn.functional as F
from torch import nn, Tensor
class Transformer(nn.Module):
def __init__(s... | EXA-1-master | exa/models/mmf-main/mmf/models/unit/transformer.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
from copy import deepcopy
from typing import Dict
import torch
from mmf.common.registry import registry
from mmf.models import BaseModel
from mmf.models.unit.unit_base_model import (
AttributeHead,
build_detection_loss,
MLP,
UniTBaseMod... | EXA-1-master | exa/models/mmf-main/mmf/models/unit/unit.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Mostly copy-pasted from
# https://github.com/facebookresearch/detr/blob/master/models/detr.py
from typing import Optional
import torch
import torch.nn.functional as F
from mmf.models.unit.backbone import build_unit_convnet_backbone
from mmf.models.unit.matcher impo... | EXA-1-master | exa/models/mmf-main/mmf/models/unit/unit_base_model.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import mmf.models.transformers.backends # noqa
from mmf.models.transformers.base import ( # noqa
BaseTransformer,
BaseTransformerBackend,
BaseTransformerBackendConfig,
BaseTransformerHead,
BaseTransformerInput,
BaseTransformerModalityConfig,
... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
from abc import ABC, abstractmethod
from dataclasses import asdict, dataclass, field
from typing import Any, Dict, List, NamedTuple, Optional, Tuple, Type
from mmf.common.registry import registry
from mmf.models import BaseModel
from mmf.modules.encode... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/base.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from mmf.utils.env import import_files
import_files(__file__, "mmf.models.transformers.backends")
| EXA-1-master | exa/models/mmf-main/mmf/models/transformers/backends/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from copy import deepcopy
from typing import Any, Dict, List, Type
import torch
from mmf.common.registry import registry
from mmf.models.transformers.base import BaseTransformer, BaseTransformerBackend
from mmf.modules.hf_layers import BertModelJit, replace_with_jit
... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/backends/huggingface.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from dataclasses import dataclass, field
from typing import Dict, List, Optional, Type
import torch
from mmf.common.registry import registry
from mmf.models.transformers.base import BaseTransformerHead
from mmf.modules.losses import RefinerContrastiveLoss, RefinerMSL... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/refiner.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from dataclasses import dataclass
from typing import Dict, List, Optional
import torch
from mmf.common.registry import registry
from mmf.models.transformers.base import BaseTransformerHead
try:
from transformers3.modeling_bert import BertOnlyNSPHead, BertPooler
... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/itm.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from dataclasses import dataclass
from typing import Dict, Optional
import torch
from mmf.common.registry import registry
from mmf.models.transformers.base import BaseTransformerHead
from mmf.models.transformers.heads.mlp import MLP
LABEL_KEY = "three_way_constrasti... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/contrastive.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
from dataclasses import dataclass
from typing import Dict, List, Optional
import torch
import torch.nn as nn
from mmf.common.registry import registry
from mmf.models.transformers.base import BaseTransformerHead
from mmf.modules.losses import LogitBinary... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/refnet_classifier.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import warnings
from dataclasses import dataclass
from typing import Dict, List, Optional
import torch
from mmf.common.registry import registry
from mmf.models.transformers.base import BaseTransformerHead
try:
from transformers3.modeling_bert import BertOnlyMLMH... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/mlm.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from mmf.utils.env import import_files
import_files(__file__, "mmf.models.transformers.heads")
| EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Initial version was taken from https://github.com/ChenRocks/UNITER/
# and adapted for MMF.
from typing import Dict
import torch
import torch.nn.functional as F
from mmf.common.registry import registry
from mmf.models.transformers.heads.utils import compute_masked_... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/mrfr.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import copy
from dataclasses import dataclass
from typing import Dict, List, Optional
import torch
from mmf.common.registry import registry
from mmf.models.transformers.base import BaseTransformerHead
from mmf.modules import layers
from omegaconf import OmegaConf, op... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/mlp.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import collections
import collections.abc
from typing import Dict, List, Optional, Union
from mmf.common.registry import registry
from torch import nn, Tensor
def build_heads_dict(head_configs: Union[Dict, List], tasks: List, losses: Dict):
"""
HeadsDict st... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/utils.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Initial version was taken from https://github.com/ChenRocks/UNITER/
# and adapted for MMF.
from typing import Dict
import torch
import torch.nn.functional as F
from mmf.common.registry import registry
from mmf.models.transformers.heads.utils import compute_masked_... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/mrc.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Initial version was taken from https://github.com/ChenRocks/UNITER/
# and adapted for MMF.
from typing import Dict
from mmf.common.registry import registry
from mmf.modules.ot import optimal_transport_dist
from torch import nn, Tensor
@registry.register_transfor... | EXA-1-master | exa/models/mmf-main/mmf/models/transformers/heads/wra.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import mmf.models.albef.vit # noqa
| EXA-1-master | exa/models/mmf-main/mmf/models/albef/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Initial version was taken from https://github.com/rwightman/pytorch-image-models
# which was cleaned up and adapted for MMF.
import collections.abc
import math
import warnings
from dataclasses import dataclass
from functools import partial
from itertools import rep... | EXA-1-master | exa/models/mmf-main/mmf/models/albef/vit.py |
# Copyright (c) Facebook, Inc. and its affiliates.
| EXA-1-master | exa/models/mmf-main/mmf/models/interfaces/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import os
import tempfile
from pathlib import Path
from typing import Type, Union
import torch
import torchvision.datasets.folder as tv_helpers
from mmf.common.sample import Sample, SampleList
from mmf.models.base_model import BaseModel
from mmf.utils.build import bu... | EXA-1-master | exa/models/mmf-main/mmf/models/interfaces/mmbt.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
import warnings
import omegaconf
import torch
from mmf.common.registry import registry
from mmf.datasets.multi_datamodule import MultiDataModule
from mmf.modules.metrics import Metrics
from mmf.trainers.base_trainer import BaseTrainer
from mmf.trainers.... | EXA-1-master | exa/models/mmf-main/mmf/trainers/mmf_trainer.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from abc import ABC, abstractmethod
from mmf.common.registry import registry
from mmf.utils.logger import log_class_usage
from omegaconf import DictConfig
@registry.register_trainer("base")
class BaseTrainer(ABC):
def __init__(self, config: DictConfig):
... | EXA-1-master | exa/models/mmf-main/mmf/trainers/base_trainer.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
import math
import os
from typing import Any, Dict, List, Optional
import omegaconf
from mmf.common.registry import registry
from mmf.datasets.lightning_multi_datamodule import LightningMultiDataModule
from mmf.modules.metrics import Metrics
from mmf.tr... | EXA-1-master | exa/models/mmf-main/mmf/trainers/lightning_trainer.py |
# Copyright (c) Facebook, Inc. and its affiliates.
__all__ = ["BaseTrainer"]
from .base_trainer import BaseTrainer
| EXA-1-master | exa/models/mmf-main/mmf/trainers/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
from abc import ABC
from typing import Any, Dict, Tuple, Type
import torch
import tqdm
from mmf.common.meter import Meter
from mmf.common.report import Report
from mmf.common.sample import to_device
from mmf.utils.distributed import gather_tensor, is_m... | EXA-1-master | exa/models/mmf-main/mmf/trainers/core/evaluation_loop.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
import warnings
from abc import ABC
import torch
from mmf.common.registry import registry
from mmf.utils.distributed import (
broadcast_xla_master_model_param,
get_world_size,
is_xla,
)
from omegaconf import open_dict
logger = logging.get... | EXA-1-master | exa/models/mmf-main/mmf/trainers/core/device.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from abc import ABC
from typing import List
from mmf.trainers.callbacks.base import Callback
class TrainerCallbackHookMixin(ABC):
callbacks: List[Callback] = []
def teardown(self, **kwargs) -> None:
"""Called to teardown the callback at end of trai... | EXA-1-master | exa/models/mmf-main/mmf/trainers/core/callback_hook.py |
# Copyright (c) Facebook, Inc. and its affiliates.
| EXA-1-master | exa/models/mmf-main/mmf/trainers/core/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import gc
import logging
from abc import ABC
from typing import Any, Dict
import torch
from mmf.common.meter import Meter
from mmf.common.registry import registry
from mmf.common.report import Report
from mmf.common.sample import to_device
from mmf.utils.distributed ... | EXA-1-master | exa/models/mmf-main/mmf/trainers/core/training_loop.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
from abc import ABC
from typing import Type
from mmf.utils.timer import Timer
logger = logging.getLogger(__name__)
class TrainerProfilingMixin(ABC):
profiler: Type[Timer] = Timer()
def profile(self, text: str) -> None:
if self.trai... | EXA-1-master | exa/models/mmf-main/mmf/trainers/core/profiling.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from mmf.trainers.callbacks.base import Callback
from mmf.utils.build import build_scheduler
class LRSchedulerCallback(Callback):
"""Callback which executes a LR scheduler. It is executed after every
batch iteration.
"""
def __init__(self, config, t... | EXA-1-master | exa/models/mmf-main/mmf/trainers/callbacks/lr_scheduler.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
from mmf.trainers.callbacks.base import Callback
from mmf.utils.checkpoint import Checkpoint, consolidate_optim_state_dict
logger = logging.getLogger(__name__)
class CheckpointCallback(Callback):
"""Callback for executing different checkpoint re... | EXA-1-master | exa/models/mmf-main/mmf/trainers/callbacks/checkpoint.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
import torch
from mmf.trainers.callbacks.base import Callback
from mmf.utils.configuration import get_mmf_env
from mmf.utils.logger import (
calculate_time_left,
setup_output_folder,
summarize_report,
TensorboardLogger,
WandbLogger,
... | EXA-1-master | exa/models/mmf-main/mmf/trainers/callbacks/logistics.py |
# Copyright (c) Facebook, Inc. and its affiliates.
| EXA-1-master | exa/models/mmf-main/mmf/trainers/callbacks/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from mmf.trainers.callbacks.base import Callback
from mmf.utils.checkpoint import consolidate_optim_state_dict
from mmf.utils.distributed import broadcast_scalar
from mmf.utils.early_stopping import EarlyStopping
class EarlyStoppingCallback(Callback):
"""Callbac... | EXA-1-master | exa/models/mmf-main/mmf/trainers/callbacks/early_stopping.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from typing import Type
from mmf.trainers.base_trainer import BaseTrainer
from omegaconf import DictConfig
class Callback:
"""
Base class for callbacks that can be registered with type :class:`BaseTrainer`
Attr:
config(omegaconf.DictConfig): Co... | EXA-1-master | exa/models/mmf-main/mmf/trainers/callbacks/base.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
from typing import Any, Dict, List, Optional
import torch
from mmf.common.registry import registry
from mmf.common.sample import SampleList
from mmf.utils.timer import Timer
from pytorch_lightning import LightningModule, Trainer
from pytorch_lightning.c... | EXA-1-master | exa/models/mmf-main/mmf/trainers/lightning_core/loop_callback_with_torchmetrics.py |
# Copyright (c) Facebook, Inc. and its affiliates.
| EXA-1-master | exa/models/mmf-main/mmf/trainers/lightning_core/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import collections
import logging
from typing import Dict, List
import torch
from mmf.common.registry import registry
from mmf.common.sample import SampleList
logger = logging.getLogger(__name__)
class LightningTorchMetrics:
"""
A class used in LightningTr... | EXA-1-master | exa/models/mmf-main/mmf/trainers/lightning_core/torchmetric.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import logging
from typing import Any, Dict, List
import torch
from mmf.common.meter import Meter
from mmf.common.registry import registry
from mmf.common.report import Report
from mmf.common.sample import SampleList
from mmf.utils.logger import calculate_time_left, ... | EXA-1-master | exa/models/mmf-main/mmf/trainers/lightning_core/loop_callback.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import csv
import json
import logging
import os
import warnings
from dataclasses import dataclass, field
from typing import List
import pytorch_lightning as pl
from mmf.common.registry import registry
from mmf.common.sample import convert_batch_to_sample_list
from mmf... | EXA-1-master | exa/models/mmf-main/mmf/common/test_reporter.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# Inspired from maskrcnn benchmark
from collections import defaultdict, deque
import torch
from mmf.common.registry import registry
from mmf.utils.distributed import reduce_dict
from mmf.utils.general import scalarize_dict_values
class SmoothedValue:
"""Track a ... | EXA-1-master | exa/models/mmf-main/mmf/common/meter.py |
# Copyright (c) Facebook, Inc. and its affiliates.
"""
Registry is central source of truth in MMF. Inspired from Redux's
concept of global store, Registry maintains mappings of various information
to unique keys. Special functions in registry can be used as decorators to
register different kind of classes.
Import the ... | EXA-1-master | exa/models/mmf-main/mmf/common/registry.py |
# Copyright (c) Facebook, Inc. and its affiliates.
imdb_version = 1
FASTTEXT_WIKI_URL = (
"https://dl.fbaipublicfiles.com/pythia/pretrained_models/fasttext/wiki.en.bin"
)
CLEVR_DOWNLOAD_URL = "https://dl.fbaipublicfiles.com/clevr/CLEVR_v1.0.zip"
VISUAL_GENOME_CONSTS = {
"imdb_url": "https://dl.fbaipublicfiles... | EXA-1-master | exa/models/mmf-main/mmf/common/constants.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from dataclasses import dataclass
from typing import Any, Dict, List
@dataclass
class PerSetAttributeType:
train: List[str]
val: List[str]
test: List[str]
@dataclass
class ProcessorConfigType:
type: str
params: Dict[str, Any]
@dataclass
class ... | EXA-1-master | exa/models/mmf-main/mmf/common/typings.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from .meter import Meter
from .registry import registry
from .sample import Sample, SampleList
__all__ = ["Sample", "SampleList", "Meter", "registry"]
| EXA-1-master | exa/models/mmf-main/mmf/common/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from mmf.common.sample import convert_batch_to_sample_list
class BatchCollator:
def __init__(self, dataset_name, dataset_type):
self._dataset_name = dataset_name
self._dataset_type = dataset_type
def __call__(self, batch):
sample_list... | EXA-1-master | exa/models/mmf-main/mmf/common/batch_collator.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import warnings
from mmf.common.sample import SampleList
from mmf.datasets.multi_dataset_loader import MultiDatasetLoader
from mmf.utils.build import build_multiple_datamodules, build_test_reporter
class DatasetLoader:
def __init__(self, config):
# TODO:... | EXA-1-master | exa/models/mmf-main/mmf/common/dataset_loader.py |
# Copyright (c) Facebook, Inc. and its affiliates.
"""
``Sample`` and ``SampleList`` are data structures for arbitrary data returned from a
dataset. To work with MMF, minimum requirement for datasets is to return
an object of ``Sample`` class and for models to accept an object of type `SampleList`
as an argument.
``Sa... | EXA-1-master | exa/models/mmf-main/mmf/common/sample.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import collections
import collections.abc
import copy
import warnings
from collections import OrderedDict
from typing import Any, Callable, Dict, List, Optional, Union
import torch
from mmf.common.sample import detach_tensor, SampleList
class Report(OrderedDict):
... | EXA-1-master | exa/models/mmf-main/mmf/common/report.py |
# Copyright (c) Facebook, Inc. and its affiliates.
"""
The metrics module contains implementations of various metrics used commonly to
understand how well our models are performing. For e.g. accuracy, vqa_accuracy,
r@1 etc.
For implementing your own metric, you need to follow these steps:
1. Create your own metric cl... | EXA-1-master | exa/models/mmf-main/mmf/modules/metrics.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import math
from typing import Optional, Tuple, Type
import torch
from mmf.modules.layers import GatedTanh, ModalCombineLayer, TransformLayer
from torch import nn
class AttentionLayer(nn.Module):
def __init__(self, image_dim, question_dim, **kwargs):
su... | EXA-1-master | exa/models/mmf-main/mmf/modules/attention.py |
# Copyright (c) Facebook, Inc. and its affiliates.
"""
This module contains implementations for various pooling methods from
transformer encoder layers
.. code::
from mmf.common.registry import registry
from torch import nn
@registry.register_pooler("custom")
class CustomPool(nn.Module):
...
"""
f... | EXA-1-master | exa/models/mmf-main/mmf/modules/poolers.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import torch
from mmf.common.registry import registry
from torch import nn
from torch.nn.utils.weight_norm import weight_norm
class VisDialDiscriminator(nn.Module):
def __init__(self, config, embedding):
super().__init__()
self.config = config
... | EXA-1-master | exa/models/mmf-main/mmf/modules/decoders.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import math
from typing import List, Optional, Tuple
import torch
from mmf.utils.patch import restore_saved_modules, safecopy_modules
from torch import nn, Tensor
try:
from transformers3.modeling_bert import (
BertAttention,
BertEmbeddings,
... | EXA-1-master | exa/models/mmf-main/mmf/modules/hf_layers.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import mmf.modules.losses # noqa
import mmf.modules.metrics # noqa
import mmf.modules.optimizers # noqa
import mmf.modules.schedulers # noqa
| EXA-1-master | exa/models/mmf-main/mmf/modules/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import importlib
import logging
import os
import pickle
import re
from collections import OrderedDict
from copy import deepcopy
from dataclasses import asdict, dataclass
from enum import Enum
from typing import Any
import torch
import torchvision
from mmf.common.regis... | EXA-1-master | exa/models/mmf-main/mmf/modules/encoders.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import math
from typing import Callable
import torch
from mmf.common.registry import registry
try:
from transformers3.optimization import AdamW
except ImportError:
from transformers.optimization import AdamW
registry.register_optimizer("adam_w")(AdamW)
@r... | EXA-1-master | exa/models/mmf-main/mmf/modules/optimizers.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from collections import namedtuple
from dataclasses import asdict, dataclass
import torch
from mmf.utils.general import retry_n
from omegaconf import OmegaConf
from packaging import version
from torch import nn
try:
from transformers3 import __version__ as trans... | EXA-1-master | exa/models/mmf-main/mmf/modules/vit.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# TODO: Update kwargs with defaults
import os
import pickle
from copy import deepcopy
from functools import lru_cache
from typing import Optional, Tuple
import numpy as np
import torch
from mmf.modules.attention import AttentionLayer, SelfAttention, SelfGuidedAttenti... | EXA-1-master | exa/models/mmf-main/mmf/modules/embeddings.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from collections import OrderedDict
from typing import Optional, Tuple, Type
import torch
import torch.nn as nn
from torchvision.models.resnet import Bottleneck, conv1x1, conv3x3
from torchvision.ops.misc import FrozenBatchNorm2d
class ChannelPool(nn.Module):
"... | EXA-1-master | exa/models/mmf-main/mmf/modules/bottleneck.py |
# Copyright (c) Facebook, Inc. and its affiliates.
"""
The fusions module contains various Fusion techniques, some based on BLOCK:
Bilinear Superdiagonal Fusion for VQA and VRD. For e.g. LinearSum, ConcatMLP
etc taken from https://github.com/Cadene/block.bootstrap.pytorch#fusions.
For implementing your own fusion tec... | EXA-1-master | exa/models/mmf-main/mmf/modules/fusions.py |
# Copyright (c) Facebook, Inc. and its affiliates.
"""
Initial version was taken from https://github.com/ChenRocks/UNITER/
Licensed under the MIT license.
Wasserstein Distance (Optimal Transport)
"""
import torch
from torch import Tensor
from torch.nn import functional as F
def cost_matrix_cosine(x: Tensor, y: Ten... | EXA-1-master | exa/models/mmf-main/mmf/modules/ot.py |
# Copyright (c) Facebook, Inc. and its affiliates.
"""
Losses module contains implementations for various losses used generally
in vision and language space. One can register custom losses to be detected by
MMF using the following example.
.. code::
from mmf.common.registry import registry
from torch import nn
... | EXA-1-master | exa/models/mmf-main/mmf/modules/losses.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from typing import Optional
import torch
from mmf.common.registry import registry
from mmf.modules.decoders import LanguageDecoder
from torch import nn
from torch.nn.utils.weight_norm import weight_norm
class ConvNet(nn.Module):
def __init__(
self,
... | EXA-1-master | exa/models/mmf-main/mmf/modules/layers.py |
# Copyright (c) Facebook, Inc. and its affiliates.
from bisect import bisect_right
from mmf.common.registry import registry
from torch.optim.lr_scheduler import LambdaLR
try:
from transformers3.optimization import (
get_cosine_schedule_with_warmup,
get_linear_schedule_with_warmup,
)
except Im... | EXA-1-master | exa/models/mmf-main/mmf/modules/schedulers.py |
#!/usr/bin/env python3 -u
# Copyright (c) Facebook, Inc. and its affiliates.
import argparse
import logging
import random
import typing
import torch
from mmf.common.registry import registry
from mmf.utils.build import build_config, build_trainer
from mmf.utils.configuration import Configuration
from mmf.utils.distribu... | EXA-1-master | exa/models/mmf-main/mmf_cli/run.py |
#!/usr/bin/env python3 -u
# Copyright (c) Facebook, Inc. and its affiliates.
import sys
from mmf_cli.run import run
def predict(opts=None):
if opts is None:
sys.argv.extend(["evaluation.predict=true"])
else:
opts.extend(["evaluation.predict=true"])
run(predict=True)
if __name__ == "__... | EXA-1-master | exa/models/mmf-main/mmf_cli/predict.py |
# Copyright (c) Facebook, Inc. and its affiliates.
| EXA-1-master | exa/models/mmf-main/mmf_cli/__init__.py |
# Copyright (c) Facebook, Inc. and its affiliates.
"""
Entrypoint script used by TorchX to start the training run in each process
"""
from mmf_cli.fb_run import fb_scheduler_run
if __name__ == "__main__":
fb_scheduler_run()
| EXA-1-master | exa/models/mmf-main/mmf_cli/torchx_entryscript.py |
#!/usr/bin/env python3 -u
# Copyright (c) Facebook, Inc. and its affiliates.
import argparse
import logging
import typing
from mmf.utils.configuration import _merge_with_dotlist
from mmf.utils.flags import flags
from mmf.utils.inference import Inference
from mmf.utils.logger import setup_logger
from omegaconf import O... | EXA-1-master | exa/models/mmf-main/mmf_cli/interactive.py |
# Copyright (c) Facebook, Inc. and its affiliates.
import argparse
import hashlib
import os
import subprocess
import warnings
import zipfile
from mmf.utils.configuration import Configuration
from mmf.utils.download import copy, decompress, move
from mmf.utils.file_io import PathManager
class HMConverter:
IMAGE_... | EXA-1-master | exa/models/mmf-main/mmf_cli/hm_convert.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
from pathlib import Path
import re
from typing import List, Tuple
from setuptools import setup, find_packages
NAME = "... | EXA-1-master | exa/models/dinov2/setup.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
import torch
import torch.nn as nn
dependencies = ["torch"]
_DINOV2_BASE_URL = "https://dl.fbaipublicfiles.com/dinov2... | EXA-1-master | exa/models/dinov2/hubconf.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
__version__ = "0.0.1"
| EXA-1-master | exa/models/dinov2/dinov2/__init__.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
# References:
# https://github.com/facebookresearch/dino/blob/master/vision_transformer.py
# https://github.com/rwigh... | EXA-1-master | exa/models/dinov2/dinov2/layers/attention.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
import torch
import torch.nn as nn
from torch.nn.init import trunc_normal_
from torch.nn.utils import weight_norm
class... | EXA-1-master | exa/models/dinov2/dinov2/layers/dino_head.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
from typing import Callable, Optional
from torch import Tensor, nn
import torch.nn.functional as F
class SwiGLUFFN(nn.... | EXA-1-master | exa/models/dinov2/dinov2/layers/swiglu_ffn.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
from .dino_head import DINOHead
from .mlp import Mlp
from .patch_embed import PatchEmbed
from .swiglu_ffn import SwiGLUFF... | EXA-1-master | exa/models/dinov2/dinov2/layers/__init__.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
# Modified from: https://github.com/huggingface/pytorch-image-models/blob/main/timm/models/vision_transformer.py#L103-L11... | EXA-1-master | exa/models/dinov2/dinov2/layers/layer_scale.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
# References:
# https://github.com/facebookresearch/dino/blob/master/vision_transformer.py
# https://github.com/rwigh... | EXA-1-master | exa/models/dinov2/dinov2/layers/mlp.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
# References:
# https://github.com/facebookresearch/dino/blob/master/vision_transformer.py
# https://github.com/rwigh... | EXA-1-master | exa/models/dinov2/dinov2/layers/patch_embed.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
# References:
# https://github.com/facebookresearch/dino/blob/master/vision_transformer.py
# https://github.com/rwigh... | EXA-1-master | exa/models/dinov2/dinov2/layers/block.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
# References:
# https://github.com/facebookresearch/dino/blob/master/vision_transformer.py
# https://github.com/rwigh... | EXA-1-master | exa/models/dinov2/dinov2/layers/drop_path.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
import torch
import torch.distributed as dist
import torch.nn.functional as F
from torch import nn
import logging
logg... | EXA-1-master | exa/models/dinov2/dinov2/loss/ibot_patch_loss.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
import logging
import torch
import torch.nn as nn
import torch.nn.functional as F
# import torch.distributed as dist
... | EXA-1-master | exa/models/dinov2/dinov2/loss/koleo_loss.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
import torch
import torch.distributed as dist
import torch.nn.functional as F
from torch import nn
class DINOLoss(nn.Mo... | EXA-1-master | exa/models/dinov2/dinov2/loss/dino_clstoken_loss.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
from .dino_clstoken_loss import DINOLoss
from .ibot_patch_loss import iBOTPatchLoss
from .koleo_loss import KoLeoLoss
| EXA-1-master | exa/models/dinov2/dinov2/loss/__init__.py |
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
import os
import random
import re
import socket
from typing import Dict, List
import torch
import torch.distributed as d... | EXA-1-master | exa/models/dinov2/dinov2/distributed/__init__.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.