python_code
stringlengths
0
992k
repo_name
stringlengths
8
46
file_path
stringlengths
5
162
# Copyright (c) Facebook, Inc. and its affiliates. import logging import math import os import torch from mmf.common.registry import registry from mmf.models import BaseModel from mmf.modules.embeddings import BertVisioLinguisticEmbeddings from mmf.utils.configuration import get_mmf_cache_dir from mmf.utils.file_io i...
EXA-1-master
exa/models/mmf-main/mmf/models/krisp.py
# Copyright (c) Facebook, Inc. and its affiliates. # MMBTModel, ModalEmbeddings is copied from [1] # as we have internal dependency on transformers v2.3. # These will be removed when we upgrade to package v2.5+. # [1]: https://github.com/huggingface/transformers/blob/master/src/transformers/modeling_mmbt.py # noqa im...
EXA-1-master
exa/models/mmf-main/mmf/models/mmbt.py
# Copyright (c) Facebook, Inc. and its affiliates. import torch from mmf.common.registry import registry from mmf.models.pythia import Pythia from mmf.modules.layers import ClassifierLayer @registry.register_model("butd") class BUTD(Pythia): def __init__(self, config): super().__init__(config) @clas...
EXA-1-master
exa/models/mmf-main/mmf/models/butd.py
# Copyright (c) Facebook, Inc. and its affiliates. import functools import logging import math import torch import torch.nn.functional as F from mmf.common.registry import registry from mmf.models.base_model import BaseModel from mmf.modules.layers import ClassifierLayer from mmf.utils.build import build_image_encoder...
EXA-1-master
exa/models/mmf-main/mmf/models/m4c.py
# Copyright (c) Facebook, Inc. and its affiliates. import copy import omegaconf import torch from mmf.common.registry import registry from mmf.models.base_model import BaseModel from mmf.modules.embeddings import ( ImageFeatureEmbedding, MultiHeadImageFeatureEmbedding, PreExtractedEmbedding, TextEmbedd...
EXA-1-master
exa/models/mmf-main/mmf/models/pythia.py
# Copyright (c) Facebook, Inc. and its affiliates. # Mostly copy-pasted from # https://github.com/facebookresearch/detr/blob/master/util/misc.py from typing import List import torch from torch import Tensor class NestedTensor: """ A data class to hold images of different sizes in a batch. It contains `...
EXA-1-master
exa/models/mmf-main/mmf/models/unit/misc.py
# Copyright (c) Facebook, Inc. and its affiliates. # Mostly copy-pasted from # https://github.com/facebookresearch/detr/blob/master/models/matcher.py """ Modules to compute the matching cost and solve the corresponding LSAP. """ from typing import Dict, List import torch from mmf.utils.box_ops import box_cxcywh_to_xy...
EXA-1-master
exa/models/mmf-main/mmf/models/unit/matcher.py
# Copyright (c) Facebook, Inc. and its affiliates. # Mostly copy-pasted from # https://github.com/facebookresearch/detr/blob/master/models/backbone.py """ Backbone modules. """ import math from collections import OrderedDict import torch import torch.nn.functional as F import torchvision from mmf.models.unit.misc imp...
EXA-1-master
exa/models/mmf-main/mmf/models/unit/backbone.py
# Copyright (c) Facebook, Inc. and its affiliates. __all__ = ["UniT"] from .unit import UniT
EXA-1-master
exa/models/mmf-main/mmf/models/unit/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # Mostly copy-pasted from # https://github.com/facebookresearch/detr/blob/master/models/transformer.py import copy from typing import Optional import torch import torch.nn.functional as F from torch import nn, Tensor class Transformer(nn.Module): def __init__(s...
EXA-1-master
exa/models/mmf-main/mmf/models/unit/transformer.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging from copy import deepcopy from typing import Dict import torch from mmf.common.registry import registry from mmf.models import BaseModel from mmf.models.unit.unit_base_model import ( AttributeHead, build_detection_loss, MLP, UniTBaseMod...
EXA-1-master
exa/models/mmf-main/mmf/models/unit/unit.py
# Copyright (c) Facebook, Inc. and its affiliates. # Mostly copy-pasted from # https://github.com/facebookresearch/detr/blob/master/models/detr.py from typing import Optional import torch import torch.nn.functional as F from mmf.models.unit.backbone import build_unit_convnet_backbone from mmf.models.unit.matcher impo...
EXA-1-master
exa/models/mmf-main/mmf/models/unit/unit_base_model.py
# Copyright (c) Facebook, Inc. and its affiliates. import mmf.models.transformers.backends # noqa from mmf.models.transformers.base import ( # noqa BaseTransformer, BaseTransformerBackend, BaseTransformerBackendConfig, BaseTransformerHead, BaseTransformerInput, BaseTransformerModalityConfig, ...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging from abc import ABC, abstractmethod from dataclasses import asdict, dataclass, field from typing import Any, Dict, List, NamedTuple, Optional, Tuple, Type from mmf.common.registry import registry from mmf.models import BaseModel from mmf.modules.encode...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/base.py
# Copyright (c) Facebook, Inc. and its affiliates. from mmf.utils.env import import_files import_files(__file__, "mmf.models.transformers.backends")
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/backends/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. from copy import deepcopy from typing import Any, Dict, List, Type import torch from mmf.common.registry import registry from mmf.models.transformers.base import BaseTransformer, BaseTransformerBackend from mmf.modules.hf_layers import BertModelJit, replace_with_jit ...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/backends/huggingface.py
# Copyright (c) Facebook, Inc. and its affiliates. from dataclasses import dataclass, field from typing import Dict, List, Optional, Type import torch from mmf.common.registry import registry from mmf.models.transformers.base import BaseTransformerHead from mmf.modules.losses import RefinerContrastiveLoss, RefinerMSL...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/refiner.py
# Copyright (c) Facebook, Inc. and its affiliates. from dataclasses import dataclass from typing import Dict, List, Optional import torch from mmf.common.registry import registry from mmf.models.transformers.base import BaseTransformerHead try: from transformers3.modeling_bert import BertOnlyNSPHead, BertPooler ...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/itm.py
# Copyright (c) Facebook, Inc. and its affiliates. from dataclasses import dataclass from typing import Dict, Optional import torch from mmf.common.registry import registry from mmf.models.transformers.base import BaseTransformerHead from mmf.models.transformers.heads.mlp import MLP LABEL_KEY = "three_way_constrasti...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/contrastive.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging from dataclasses import dataclass from typing import Dict, List, Optional import torch import torch.nn as nn from mmf.common.registry import registry from mmf.models.transformers.base import BaseTransformerHead from mmf.modules.losses import LogitBinary...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/refnet_classifier.py
# Copyright (c) Facebook, Inc. and its affiliates. import warnings from dataclasses import dataclass from typing import Dict, List, Optional import torch from mmf.common.registry import registry from mmf.models.transformers.base import BaseTransformerHead try: from transformers3.modeling_bert import BertOnlyMLMH...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/mlm.py
# Copyright (c) Facebook, Inc. and its affiliates. from mmf.utils.env import import_files import_files(__file__, "mmf.models.transformers.heads")
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # Initial version was taken from https://github.com/ChenRocks/UNITER/ # and adapted for MMF. from typing import Dict import torch import torch.nn.functional as F from mmf.common.registry import registry from mmf.models.transformers.heads.utils import compute_masked_...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/mrfr.py
# Copyright (c) Facebook, Inc. and its affiliates. import copy from dataclasses import dataclass from typing import Dict, List, Optional import torch from mmf.common.registry import registry from mmf.models.transformers.base import BaseTransformerHead from mmf.modules import layers from omegaconf import OmegaConf, op...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/mlp.py
# Copyright (c) Facebook, Inc. and its affiliates. import collections import collections.abc from typing import Dict, List, Optional, Union from mmf.common.registry import registry from torch import nn, Tensor def build_heads_dict(head_configs: Union[Dict, List], tasks: List, losses: Dict): """ HeadsDict st...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/utils.py
# Copyright (c) Facebook, Inc. and its affiliates. # Initial version was taken from https://github.com/ChenRocks/UNITER/ # and adapted for MMF. from typing import Dict import torch import torch.nn.functional as F from mmf.common.registry import registry from mmf.models.transformers.heads.utils import compute_masked_...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/mrc.py
# Copyright (c) Facebook, Inc. and its affiliates. # Initial version was taken from https://github.com/ChenRocks/UNITER/ # and adapted for MMF. from typing import Dict from mmf.common.registry import registry from mmf.modules.ot import optimal_transport_dist from torch import nn, Tensor @registry.register_transfor...
EXA-1-master
exa/models/mmf-main/mmf/models/transformers/heads/wra.py
# Copyright (c) Facebook, Inc. and its affiliates. import mmf.models.albef.vit # noqa
EXA-1-master
exa/models/mmf-main/mmf/models/albef/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # Initial version was taken from https://github.com/rwightman/pytorch-image-models # which was cleaned up and adapted for MMF. import collections.abc import math import warnings from dataclasses import dataclass from functools import partial from itertools import rep...
EXA-1-master
exa/models/mmf-main/mmf/models/albef/vit.py
# Copyright (c) Facebook, Inc. and its affiliates.
EXA-1-master
exa/models/mmf-main/mmf/models/interfaces/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. import os import tempfile from pathlib import Path from typing import Type, Union import torch import torchvision.datasets.folder as tv_helpers from mmf.common.sample import Sample, SampleList from mmf.models.base_model import BaseModel from mmf.utils.build import bu...
EXA-1-master
exa/models/mmf-main/mmf/models/interfaces/mmbt.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging import warnings import omegaconf import torch from mmf.common.registry import registry from mmf.datasets.multi_datamodule import MultiDataModule from mmf.modules.metrics import Metrics from mmf.trainers.base_trainer import BaseTrainer from mmf.trainers....
EXA-1-master
exa/models/mmf-main/mmf/trainers/mmf_trainer.py
# Copyright (c) Facebook, Inc. and its affiliates. from abc import ABC, abstractmethod from mmf.common.registry import registry from mmf.utils.logger import log_class_usage from omegaconf import DictConfig @registry.register_trainer("base") class BaseTrainer(ABC): def __init__(self, config: DictConfig): ...
EXA-1-master
exa/models/mmf-main/mmf/trainers/base_trainer.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging import math import os from typing import Any, Dict, List, Optional import omegaconf from mmf.common.registry import registry from mmf.datasets.lightning_multi_datamodule import LightningMultiDataModule from mmf.modules.metrics import Metrics from mmf.tr...
EXA-1-master
exa/models/mmf-main/mmf/trainers/lightning_trainer.py
# Copyright (c) Facebook, Inc. and its affiliates. __all__ = ["BaseTrainer"] from .base_trainer import BaseTrainer
EXA-1-master
exa/models/mmf-main/mmf/trainers/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging from abc import ABC from typing import Any, Dict, Tuple, Type import torch import tqdm from mmf.common.meter import Meter from mmf.common.report import Report from mmf.common.sample import to_device from mmf.utils.distributed import gather_tensor, is_m...
EXA-1-master
exa/models/mmf-main/mmf/trainers/core/evaluation_loop.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging import warnings from abc import ABC import torch from mmf.common.registry import registry from mmf.utils.distributed import ( broadcast_xla_master_model_param, get_world_size, is_xla, ) from omegaconf import open_dict logger = logging.get...
EXA-1-master
exa/models/mmf-main/mmf/trainers/core/device.py
# Copyright (c) Facebook, Inc. and its affiliates. from abc import ABC from typing import List from mmf.trainers.callbacks.base import Callback class TrainerCallbackHookMixin(ABC): callbacks: List[Callback] = [] def teardown(self, **kwargs) -> None: """Called to teardown the callback at end of trai...
EXA-1-master
exa/models/mmf-main/mmf/trainers/core/callback_hook.py
# Copyright (c) Facebook, Inc. and its affiliates.
EXA-1-master
exa/models/mmf-main/mmf/trainers/core/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. import gc import logging from abc import ABC from typing import Any, Dict import torch from mmf.common.meter import Meter from mmf.common.registry import registry from mmf.common.report import Report from mmf.common.sample import to_device from mmf.utils.distributed ...
EXA-1-master
exa/models/mmf-main/mmf/trainers/core/training_loop.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging from abc import ABC from typing import Type from mmf.utils.timer import Timer logger = logging.getLogger(__name__) class TrainerProfilingMixin(ABC): profiler: Type[Timer] = Timer() def profile(self, text: str) -> None: if self.trai...
EXA-1-master
exa/models/mmf-main/mmf/trainers/core/profiling.py
# Copyright (c) Facebook, Inc. and its affiliates. from mmf.trainers.callbacks.base import Callback from mmf.utils.build import build_scheduler class LRSchedulerCallback(Callback): """Callback which executes a LR scheduler. It is executed after every batch iteration. """ def __init__(self, config, t...
EXA-1-master
exa/models/mmf-main/mmf/trainers/callbacks/lr_scheduler.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging from mmf.trainers.callbacks.base import Callback from mmf.utils.checkpoint import Checkpoint, consolidate_optim_state_dict logger = logging.getLogger(__name__) class CheckpointCallback(Callback): """Callback for executing different checkpoint re...
EXA-1-master
exa/models/mmf-main/mmf/trainers/callbacks/checkpoint.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging import torch from mmf.trainers.callbacks.base import Callback from mmf.utils.configuration import get_mmf_env from mmf.utils.logger import ( calculate_time_left, setup_output_folder, summarize_report, TensorboardLogger, WandbLogger, ...
EXA-1-master
exa/models/mmf-main/mmf/trainers/callbacks/logistics.py
# Copyright (c) Facebook, Inc. and its affiliates.
EXA-1-master
exa/models/mmf-main/mmf/trainers/callbacks/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. from mmf.trainers.callbacks.base import Callback from mmf.utils.checkpoint import consolidate_optim_state_dict from mmf.utils.distributed import broadcast_scalar from mmf.utils.early_stopping import EarlyStopping class EarlyStoppingCallback(Callback): """Callbac...
EXA-1-master
exa/models/mmf-main/mmf/trainers/callbacks/early_stopping.py
# Copyright (c) Facebook, Inc. and its affiliates. from typing import Type from mmf.trainers.base_trainer import BaseTrainer from omegaconf import DictConfig class Callback: """ Base class for callbacks that can be registered with type :class:`BaseTrainer` Attr: config(omegaconf.DictConfig): Co...
EXA-1-master
exa/models/mmf-main/mmf/trainers/callbacks/base.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging from typing import Any, Dict, List, Optional import torch from mmf.common.registry import registry from mmf.common.sample import SampleList from mmf.utils.timer import Timer from pytorch_lightning import LightningModule, Trainer from pytorch_lightning.c...
EXA-1-master
exa/models/mmf-main/mmf/trainers/lightning_core/loop_callback_with_torchmetrics.py
# Copyright (c) Facebook, Inc. and its affiliates.
EXA-1-master
exa/models/mmf-main/mmf/trainers/lightning_core/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. import collections import logging from typing import Dict, List import torch from mmf.common.registry import registry from mmf.common.sample import SampleList logger = logging.getLogger(__name__) class LightningTorchMetrics: """ A class used in LightningTr...
EXA-1-master
exa/models/mmf-main/mmf/trainers/lightning_core/torchmetric.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging from typing import Any, Dict, List import torch from mmf.common.meter import Meter from mmf.common.registry import registry from mmf.common.report import Report from mmf.common.sample import SampleList from mmf.utils.logger import calculate_time_left, ...
EXA-1-master
exa/models/mmf-main/mmf/trainers/lightning_core/loop_callback.py
# Copyright (c) Facebook, Inc. and its affiliates. import csv import json import logging import os import warnings from dataclasses import dataclass, field from typing import List import pytorch_lightning as pl from mmf.common.registry import registry from mmf.common.sample import convert_batch_to_sample_list from mmf...
EXA-1-master
exa/models/mmf-main/mmf/common/test_reporter.py
# Copyright (c) Facebook, Inc. and its affiliates. # Inspired from maskrcnn benchmark from collections import defaultdict, deque import torch from mmf.common.registry import registry from mmf.utils.distributed import reduce_dict from mmf.utils.general import scalarize_dict_values class SmoothedValue: """Track a ...
EXA-1-master
exa/models/mmf-main/mmf/common/meter.py
# Copyright (c) Facebook, Inc. and its affiliates. """ Registry is central source of truth in MMF. Inspired from Redux's concept of global store, Registry maintains mappings of various information to unique keys. Special functions in registry can be used as decorators to register different kind of classes. Import the ...
EXA-1-master
exa/models/mmf-main/mmf/common/registry.py
# Copyright (c) Facebook, Inc. and its affiliates. imdb_version = 1 FASTTEXT_WIKI_URL = ( "https://dl.fbaipublicfiles.com/pythia/pretrained_models/fasttext/wiki.en.bin" ) CLEVR_DOWNLOAD_URL = "https://dl.fbaipublicfiles.com/clevr/CLEVR_v1.0.zip" VISUAL_GENOME_CONSTS = { "imdb_url": "https://dl.fbaipublicfiles...
EXA-1-master
exa/models/mmf-main/mmf/common/constants.py
# Copyright (c) Facebook, Inc. and its affiliates. from dataclasses import dataclass from typing import Any, Dict, List @dataclass class PerSetAttributeType: train: List[str] val: List[str] test: List[str] @dataclass class ProcessorConfigType: type: str params: Dict[str, Any] @dataclass class ...
EXA-1-master
exa/models/mmf-main/mmf/common/typings.py
# Copyright (c) Facebook, Inc. and its affiliates. from .meter import Meter from .registry import registry from .sample import Sample, SampleList __all__ = ["Sample", "SampleList", "Meter", "registry"]
EXA-1-master
exa/models/mmf-main/mmf/common/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. from mmf.common.sample import convert_batch_to_sample_list class BatchCollator: def __init__(self, dataset_name, dataset_type): self._dataset_name = dataset_name self._dataset_type = dataset_type def __call__(self, batch): sample_list...
EXA-1-master
exa/models/mmf-main/mmf/common/batch_collator.py
# Copyright (c) Facebook, Inc. and its affiliates. import warnings from mmf.common.sample import SampleList from mmf.datasets.multi_dataset_loader import MultiDatasetLoader from mmf.utils.build import build_multiple_datamodules, build_test_reporter class DatasetLoader: def __init__(self, config): # TODO:...
EXA-1-master
exa/models/mmf-main/mmf/common/dataset_loader.py
# Copyright (c) Facebook, Inc. and its affiliates. """ ``Sample`` and ``SampleList`` are data structures for arbitrary data returned from a dataset. To work with MMF, minimum requirement for datasets is to return an object of ``Sample`` class and for models to accept an object of type `SampleList` as an argument. ``Sa...
EXA-1-master
exa/models/mmf-main/mmf/common/sample.py
# Copyright (c) Facebook, Inc. and its affiliates. import collections import collections.abc import copy import warnings from collections import OrderedDict from typing import Any, Callable, Dict, List, Optional, Union import torch from mmf.common.sample import detach_tensor, SampleList class Report(OrderedDict): ...
EXA-1-master
exa/models/mmf-main/mmf/common/report.py
# Copyright (c) Facebook, Inc. and its affiliates. """ The metrics module contains implementations of various metrics used commonly to understand how well our models are performing. For e.g. accuracy, vqa_accuracy, r@1 etc. For implementing your own metric, you need to follow these steps: 1. Create your own metric cl...
EXA-1-master
exa/models/mmf-main/mmf/modules/metrics.py
# Copyright (c) Facebook, Inc. and its affiliates. import math from typing import Optional, Tuple, Type import torch from mmf.modules.layers import GatedTanh, ModalCombineLayer, TransformLayer from torch import nn class AttentionLayer(nn.Module): def __init__(self, image_dim, question_dim, **kwargs): su...
EXA-1-master
exa/models/mmf-main/mmf/modules/attention.py
# Copyright (c) Facebook, Inc. and its affiliates. """ This module contains implementations for various pooling methods from transformer encoder layers .. code:: from mmf.common.registry import registry from torch import nn @registry.register_pooler("custom") class CustomPool(nn.Module): ... """ f...
EXA-1-master
exa/models/mmf-main/mmf/modules/poolers.py
# Copyright (c) Facebook, Inc. and its affiliates. import torch from mmf.common.registry import registry from torch import nn from torch.nn.utils.weight_norm import weight_norm class VisDialDiscriminator(nn.Module): def __init__(self, config, embedding): super().__init__() self.config = config ...
EXA-1-master
exa/models/mmf-main/mmf/modules/decoders.py
# Copyright (c) Facebook, Inc. and its affiliates. import math from typing import List, Optional, Tuple import torch from mmf.utils.patch import restore_saved_modules, safecopy_modules from torch import nn, Tensor try: from transformers3.modeling_bert import ( BertAttention, BertEmbeddings, ...
EXA-1-master
exa/models/mmf-main/mmf/modules/hf_layers.py
# Copyright (c) Facebook, Inc. and its affiliates. import mmf.modules.losses # noqa import mmf.modules.metrics # noqa import mmf.modules.optimizers # noqa import mmf.modules.schedulers # noqa
EXA-1-master
exa/models/mmf-main/mmf/modules/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. import importlib import logging import os import pickle import re from collections import OrderedDict from copy import deepcopy from dataclasses import asdict, dataclass from enum import Enum from typing import Any import torch import torchvision from mmf.common.regis...
EXA-1-master
exa/models/mmf-main/mmf/modules/encoders.py
# Copyright (c) Facebook, Inc. and its affiliates. import math from typing import Callable import torch from mmf.common.registry import registry try: from transformers3.optimization import AdamW except ImportError: from transformers.optimization import AdamW registry.register_optimizer("adam_w")(AdamW) @r...
EXA-1-master
exa/models/mmf-main/mmf/modules/optimizers.py
# Copyright (c) Facebook, Inc. and its affiliates. from collections import namedtuple from dataclasses import asdict, dataclass import torch from mmf.utils.general import retry_n from omegaconf import OmegaConf from packaging import version from torch import nn try: from transformers3 import __version__ as trans...
EXA-1-master
exa/models/mmf-main/mmf/modules/vit.py
# Copyright (c) Facebook, Inc. and its affiliates. # TODO: Update kwargs with defaults import os import pickle from copy import deepcopy from functools import lru_cache from typing import Optional, Tuple import numpy as np import torch from mmf.modules.attention import AttentionLayer, SelfAttention, SelfGuidedAttenti...
EXA-1-master
exa/models/mmf-main/mmf/modules/embeddings.py
# Copyright (c) Facebook, Inc. and its affiliates. from collections import OrderedDict from typing import Optional, Tuple, Type import torch import torch.nn as nn from torchvision.models.resnet import Bottleneck, conv1x1, conv3x3 from torchvision.ops.misc import FrozenBatchNorm2d class ChannelPool(nn.Module): "...
EXA-1-master
exa/models/mmf-main/mmf/modules/bottleneck.py
# Copyright (c) Facebook, Inc. and its affiliates. """ The fusions module contains various Fusion techniques, some based on BLOCK: Bilinear Superdiagonal Fusion for VQA and VRD. For e.g. LinearSum, ConcatMLP etc taken from https://github.com/Cadene/block.bootstrap.pytorch#fusions. For implementing your own fusion tec...
EXA-1-master
exa/models/mmf-main/mmf/modules/fusions.py
# Copyright (c) Facebook, Inc. and its affiliates. """ Initial version was taken from https://github.com/ChenRocks/UNITER/ Licensed under the MIT license. Wasserstein Distance (Optimal Transport) """ import torch from torch import Tensor from torch.nn import functional as F def cost_matrix_cosine(x: Tensor, y: Ten...
EXA-1-master
exa/models/mmf-main/mmf/modules/ot.py
# Copyright (c) Facebook, Inc. and its affiliates. """ Losses module contains implementations for various losses used generally in vision and language space. One can register custom losses to be detected by MMF using the following example. .. code:: from mmf.common.registry import registry from torch import nn ...
EXA-1-master
exa/models/mmf-main/mmf/modules/losses.py
# Copyright (c) Facebook, Inc. and its affiliates. from typing import Optional import torch from mmf.common.registry import registry from mmf.modules.decoders import LanguageDecoder from torch import nn from torch.nn.utils.weight_norm import weight_norm class ConvNet(nn.Module): def __init__( self, ...
EXA-1-master
exa/models/mmf-main/mmf/modules/layers.py
# Copyright (c) Facebook, Inc. and its affiliates. from bisect import bisect_right from mmf.common.registry import registry from torch.optim.lr_scheduler import LambdaLR try: from transformers3.optimization import ( get_cosine_schedule_with_warmup, get_linear_schedule_with_warmup, ) except Im...
EXA-1-master
exa/models/mmf-main/mmf/modules/schedulers.py
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. import argparse import logging import random import typing import torch from mmf.common.registry import registry from mmf.utils.build import build_config, build_trainer from mmf.utils.configuration import Configuration from mmf.utils.distribu...
EXA-1-master
exa/models/mmf-main/mmf_cli/run.py
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. import sys from mmf_cli.run import run def predict(opts=None): if opts is None: sys.argv.extend(["evaluation.predict=true"]) else: opts.extend(["evaluation.predict=true"]) run(predict=True) if __name__ == "__...
EXA-1-master
exa/models/mmf-main/mmf_cli/predict.py
# Copyright (c) Facebook, Inc. and its affiliates.
EXA-1-master
exa/models/mmf-main/mmf_cli/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. """ Entrypoint script used by TorchX to start the training run in each process """ from mmf_cli.fb_run import fb_scheduler_run if __name__ == "__main__": fb_scheduler_run()
EXA-1-master
exa/models/mmf-main/mmf_cli/torchx_entryscript.py
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. import argparse import logging import typing from mmf.utils.configuration import _merge_with_dotlist from mmf.utils.flags import flags from mmf.utils.inference import Inference from mmf.utils.logger import setup_logger from omegaconf import O...
EXA-1-master
exa/models/mmf-main/mmf_cli/interactive.py
# Copyright (c) Facebook, Inc. and its affiliates. import argparse import hashlib import os import subprocess import warnings import zipfile from mmf.utils.configuration import Configuration from mmf.utils.download import copy, decompress, move from mmf.utils.file_io import PathManager class HMConverter: IMAGE_...
EXA-1-master
exa/models/mmf-main/mmf_cli/hm_convert.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. from pathlib import Path import re from typing import List, Tuple from setuptools import setup, find_packages NAME = "...
EXA-1-master
exa/models/dinov2/setup.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import torch import torch.nn as nn dependencies = ["torch"] _DINOV2_BASE_URL = "https://dl.fbaipublicfiles.com/dinov2...
EXA-1-master
exa/models/dinov2/hubconf.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. __version__ = "0.0.1"
EXA-1-master
exa/models/dinov2/dinov2/__init__.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. # References: # https://github.com/facebookresearch/dino/blob/master/vision_transformer.py # https://github.com/rwigh...
EXA-1-master
exa/models/dinov2/dinov2/layers/attention.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import torch import torch.nn as nn from torch.nn.init import trunc_normal_ from torch.nn.utils import weight_norm class...
EXA-1-master
exa/models/dinov2/dinov2/layers/dino_head.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. from typing import Callable, Optional from torch import Tensor, nn import torch.nn.functional as F class SwiGLUFFN(nn....
EXA-1-master
exa/models/dinov2/dinov2/layers/swiglu_ffn.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. from .dino_head import DINOHead from .mlp import Mlp from .patch_embed import PatchEmbed from .swiglu_ffn import SwiGLUFF...
EXA-1-master
exa/models/dinov2/dinov2/layers/__init__.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. # Modified from: https://github.com/huggingface/pytorch-image-models/blob/main/timm/models/vision_transformer.py#L103-L11...
EXA-1-master
exa/models/dinov2/dinov2/layers/layer_scale.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. # References: # https://github.com/facebookresearch/dino/blob/master/vision_transformer.py # https://github.com/rwigh...
EXA-1-master
exa/models/dinov2/dinov2/layers/mlp.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. # References: # https://github.com/facebookresearch/dino/blob/master/vision_transformer.py # https://github.com/rwigh...
EXA-1-master
exa/models/dinov2/dinov2/layers/patch_embed.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. # References: # https://github.com/facebookresearch/dino/blob/master/vision_transformer.py # https://github.com/rwigh...
EXA-1-master
exa/models/dinov2/dinov2/layers/block.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. # References: # https://github.com/facebookresearch/dino/blob/master/vision_transformer.py # https://github.com/rwigh...
EXA-1-master
exa/models/dinov2/dinov2/layers/drop_path.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import torch import torch.distributed as dist import torch.nn.functional as F from torch import nn import logging logg...
EXA-1-master
exa/models/dinov2/dinov2/loss/ibot_patch_loss.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import logging import torch import torch.nn as nn import torch.nn.functional as F # import torch.distributed as dist ...
EXA-1-master
exa/models/dinov2/dinov2/loss/koleo_loss.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import torch import torch.distributed as dist import torch.nn.functional as F from torch import nn class DINOLoss(nn.Mo...
EXA-1-master
exa/models/dinov2/dinov2/loss/dino_clstoken_loss.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. from .dino_clstoken_loss import DINOLoss from .ibot_patch_loss import iBOTPatchLoss from .koleo_loss import KoLeoLoss
EXA-1-master
exa/models/dinov2/dinov2/loss/__init__.py
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import os import random import re import socket from typing import Dict, List import torch import torch.distributed as d...
EXA-1-master
exa/models/dinov2/dinov2/distributed/__init__.py