python_code stringlengths 0 992k | repo_name stringlengths 8 46 | file_path stringlengths 5 162 |
|---|---|---|
import subprocess
import matplotlib
import os
matplotlib.use('Agg')
import librosa
import librosa.filters
import numpy as np
from scipy import signal
from scipy.io import wavfile
def save_wav(wav, path, sr, norm=False):
if norm:
wav = wav / np.abs(wav).max()
wav *= 32767
# proposed by @dsmiller
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/utils/audio.py |
from numpy import array, zeros, full, argmin, inf, ndim
from scipy.spatial.distance import cdist
from math import isinf
def dtw(x, y, dist, warp=1, w=inf, s=1.0):
"""
Computes Dynamic Time Warping (DTW) of two sequences.
:param array x: N1*M array
:param array y: N2*M array
:param func dist: dist... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/utils/dtw.py |
import os
import traceback
from multiprocessing import Queue, Process
def chunked_worker(worker_id, map_func, args, results_queue=None, init_ctx_func=None):
ctx = init_ctx_func(worker_id) if init_ctx_func is not None else None
for job_idx, arg in args:
try:
if ctx is not None:
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/utils/multiprocess_utils.py |
import os
from data_gen.tts.base_preprocess import BasePreprocessor
import glob
class LibrittsPreAlign(BasePreprocessor):
def meta_data(self):
wav_fns = sorted(glob.glob(f'{self.raw_data_dir}/*/*/*.wav'))
for wav_fn in wav_fns:
item_name = os.path.basename(wav_fn)[:-4]
txt... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/configs/tts/libritts/pre_align.py |
import os
from data_gen.tts.base_preprocess import BasePreprocessor
import glob
import re
class EmoPreAlign(BasePreprocessor):
def meta_data(self):
spks = ['0012', '0011', '0013', '0014', '0015', '0016', '0017', '0018', '0019', '0020']
pattern = re.compile('[\t\n ]+')
for spk in spks:
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/configs/tts/emotion/pre_align.py |
import torch
from inference.svs.base_svs_infer import BaseSVSInfer
from utils import load_ckpt
from utils.hparams import hparams
from modulesmodules.diff.shallow_diffusion_tts import GaussianDiffusion
from tasks.svs.diffsinger_task import DIFF_DECODERS
class DiffSingerCascadeInfer(BaseSVSInfer):
def build_model(se... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/inference/svs/ds_cascade.py |
import os
import torch
import numpy as np
from modules.hifigan.hifigan import HifiGanGenerator
from vocoders.hifigan import HifiGAN
from inference.svs.opencpop.map import cpop_pinyin2ph_func
from utils import load_ckpt
from utils.hparams import set_hparams, hparams
from utils.text_encoder import TokenTextEncoder
from... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/inference/svs/base_svs_infer.py |
import torch
# from inference.tts.fs import FastSpeechInfer
# from modules.tts.fs2_orig import FastSpeech2Orig
from inference.svs.base_svs_infer import BaseSVSInfer
from utils import load_ckpt
from utils.hparams import hparams
from modules.diff.shallow_diffusion_tts import GaussianDiffusion
from tasks.svs.diffsinger_ta... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/inference/svs/ds_e2e.py |
def cpop_pinyin2ph_func():
# In the README file of opencpop dataset, they defined a "pinyin to phoneme mapping table"
pinyin2phs = {'AP': 'AP', 'SP': 'SP'}
with open('NeuralSeq/inference/svs/opencpop/cpop_pinyin2ph.txt') as rf:
for line in rf.readlines():
elements = [x.strip() for x in l... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/inference/svs/opencpop/map.py |
import torch
from inference.tts.base_tts_infer import BaseTTSInfer
from utils.ckpt_utils import load_ckpt
from modules.portaspeech.portaspeech import PortaSpeech
class TTSInference(BaseTTSInfer):
def __init__(self, hparams, device=None):
super().__init__(hparams, device)
print("Initializing TTS mod... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/inference/tts/PortaSpeech.py |
import torch
import os
import importlib
from inference.tts.base_tts_infer import BaseTTSInfer
from utils.ckpt_utils import load_ckpt, get_last_checkpoint
from modules.GenerSpeech.model.generspeech import GenerSpeech
from data_gen.tts.emotion import inference as EmotionEncoder
from data_gen.tts.emotion.inference import ... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/inference/tts/GenerSpeech.py |
from tasks.tts.dataset_utils import FastSpeechWordDataset
from tasks.tts.tts_utils import load_data_preprocessor
from vocoders.hifigan import HifiGanGenerator
import os
import librosa
import soundfile as sf
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
from string import punctuation
import torch
from utils... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/inference/tts/base_tts_infer.py |
EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/__init__.py | |
# -*- coding: utf-8 -*-
# Copyright 2019 Tomoki Hayashi
# MIT License (https://opensource.org/licenses/MIT)
"""STFT-based Loss modules."""
import librosa
import torch
from modules.parallel_wavegan.losses import LogSTFTMagnitudeLoss, SpectralConvergengeLoss, stft
class STFTLoss(torch.nn.Module):
"""STFT loss m... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/stft_loss.py |
EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/__init__.py | |
# -*- coding: utf-8 -*-
# Copyright 2019 Tomoki Hayashi
# MIT License (https://opensource.org/licenses/MIT)
"""STFT-based Loss modules."""
import torch
import torch.nn.functional as F
def stft(x, fft_size, hop_size, win_length, window):
"""Perform STFT and convert to magnitude spectrogram.
Args:
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/losses/stft_loss.py |
from .stft_loss import * # NOQA
| EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/losses/__init__.py |
# -*- coding: utf-8 -*-
# Copyright 2020 MINH ANH (@dathudeptrai)
# MIT License (https://opensource.org/licenses/MIT)
"""Tensorflow Layer modules complatible with pytorch."""
import tensorflow as tf
class TFReflectionPad1d(tf.keras.layers.Layer):
"""Tensorflow ReflectionPad1d module."""
def __init__(self... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/layers/tf_layers.py |
# -*- coding: utf-8 -*-
# Copyright 2020 Tomoki Hayashi
# MIT License (https://opensource.org/licenses/MIT)
"""Causal convolusion layer modules."""
import torch
class CausalConv1d(torch.nn.Module):
"""CausalConv1d module with customized initialization."""
def __init__(self, in_channels, out_channels, ke... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/layers/causal_conv.py |
from .causal_conv import * # NOQA
from .pqmf import * # NOQA
from .residual_block import * # NOQA
from modules.parallel_wavegan.layers.residual_stack import * # NOQA
from .upsample import * # NOQA
| EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/layers/__init__.py |
# -*- coding: utf-8 -*-
"""Upsampling module.
This code is modified from https://github.com/r9y9/wavenet_vocoder.
"""
import numpy as np
import torch
import torch.nn.functional as F
from . import Conv1d
class Stretch2d(torch.nn.Module):
"""Stretch2d module."""
def __init__(self, x_scale, y_scale, mode="... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/layers/upsample.py |
# -*- coding: utf-8 -*-
"""Residual block module in WaveNet.
This code is modified from https://github.com/r9y9/wavenet_vocoder.
"""
import math
import torch
import torch.nn.functional as F
class Conv1d(torch.nn.Conv1d):
"""Conv1d module with customized initialization."""
def __init__(self, *args, **kwa... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/layers/residual_block.py |
# -*- coding: utf-8 -*-
# Copyright 2020 Tomoki Hayashi
# MIT License (https://opensource.org/licenses/MIT)
"""Pseudo QMF modules."""
import numpy as np
import torch
import torch.nn.functional as F
from scipy.signal import kaiser
def design_prototype_filter(taps=62, cutoff_ratio=0.15, beta=9.0):
"""Design pr... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/layers/pqmf.py |
# -*- coding: utf-8 -*-
# Copyright 2020 Tomoki Hayashi
# MIT License (https://opensource.org/licenses/MIT)
"""Residual stack module in MelGAN."""
import torch
from . import CausalConv1d
class ResidualStack(torch.nn.Module):
"""Residual stack module introduced in MelGAN."""
def __init__(self,
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/layers/residual_stack.py |
from .utils import * # NOQA
| EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/utils/__init__.py |
# -*- coding: utf-8 -*-
# Copyright 2019 Tomoki Hayashi
# MIT License (https://opensource.org/licenses/MIT)
"""Utility functions."""
import fnmatch
import logging
import os
import sys
import h5py
import numpy as np
def find_files(root_dir, query="*.wav", include_root_dir=True):
"""Find files recursively.
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/utils/utils.py |
from .melgan import * # NOQA
from .parallel_wavegan import * # NOQA
| EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/models/__init__.py |
# -*- coding: utf-8 -*-
# Copyright 2020 Tomoki Hayashi
# MIT License (https://opensource.org/licenses/MIT)
"""MelGAN Modules."""
import logging
import numpy as np
import torch
from modules.parallel_wavegan.layers import CausalConv1d
from modules.parallel_wavegan.layers import CausalConvTranspose1d
from modules.p... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/models/melgan.py |
# -*- coding: utf-8 -*-
# Copyright 2019 Tomoki Hayashi
# MIT License (https://opensource.org/licenses/MIT)
"""Parallel WaveGAN Modules."""
import logging
import math
import torch
from torch import nn
from modules.parallel_wavegan.layers import Conv1d
from modules.parallel_wavegan.layers import Conv1d1x1
from mod... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/models/parallel_wavegan.py |
import torch
import numpy as np
import sys
import torch.nn.functional as torch_nn_func
class SineGen(torch.nn.Module):
""" Definition of sine generator
SineGen(samp_rate, harmonic_num = 0,
sine_amp = 0.1, noise_std = 0.003,
voiced_threshold = 0,
flag_for_pulse=False)
s... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/models/source.py |
from torch.optim import * # NOQA
from .radam import * # NOQA
| EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/optimizers/__init__.py |
# -*- coding: utf-8 -*-
"""RAdam optimizer.
This code is drived from https://github.com/LiyuanLucasLiu/RAdam.
"""
import math
import torch
from torch.optim.optimizer import Optimizer
class RAdam(Optimizer):
"""Rectified Adam optimizer."""
def __init__(self, params, lr=1e-3, betas=(0.9, 0.999), eps=1e-8, ... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/parallel_wavegan/optimizers/radam.py |
from modules.commons.common_layers import *
from modules.commons.common_layers import Embedding
from modules.fastspeech.tts_modules import FastspeechDecoder, DurationPredictor, LengthRegulator, PitchPredictor, \
EnergyPredictor, FastspeechEncoder
from utils.cwt import cwt2f0
from utils.hparams import hparams
from u... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/diffsinger_midi/fs2.py |
import math
import torch
from torch import nn
from torch.nn import Parameter
import torch.onnx.operators
import torch.nn.functional as F
import utils
class Reshape(nn.Module):
def __init__(self, *args):
super(Reshape, self).__init__()
self.shape = args
def forward(self, x):
return x.v... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/common_layers.py |
# '''
# https://github.com/One-sixth/ms_ssim_pytorch/blob/master/ssim.py
# '''
#
# import torch
# import torch.jit
# import torch.nn.functional as F
#
#
# @torch.jit.script
# def create_window(window_size: int, sigma: float, channel: int):
# '''
# Create 1-D gauss kernel
# :param window_size: the size of ga... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/ssim.py |
import torch
from torch import nn
def fused_add_tanh_sigmoid_multiply(input_a, input_b, n_channels):
n_channels_int = n_channels[0]
in_act = input_a + input_b
t_act = torch.tanh(in_act[:, :n_channels_int, :])
s_act = torch.sigmoid(in_act[:, n_channels_int:, :])
acts = t_act * s_act
return acts... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/wavenet.py |
import math
import torch
from torch import nn
from torch.nn import functional as F
from utils.hparams import hparams
from modules.commons.common_layers import Embedding
from utils.tts_utils import group_hidden_by_segs, expand_word2ph
import transformers
def convert_pad_shape(pad_shape):
l = pad_shape[::-1]
pa... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/rel_transformer.py |
import math
import torch
from torch import nn
from torch.nn import Parameter, Linear
from modules.commons.common_layers import LayerNorm, Embedding
from utils.tts_utils import get_incremental_state, set_incremental_state, softmax, make_positions
import torch.nn.functional as F
DEFAULT_MAX_SOURCE_POSITIONS = 2000
DEFAU... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/transformer.py |
import torch
import torch.nn.functional as F
def build_word_mask(x2word, y2word):
return (x2word[:, :, None] == y2word[:, None, :]).long()
def mel2ph_to_mel2word(mel2ph, ph2word):
mel2word = (ph2word - 1).gather(1, (mel2ph - 1).clamp(min=0)) + 1
mel2word = mel2word * (mel2ph > 0).long()
return mel2w... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/align_ops.py |
import math
import torch
import torch.nn as nn
import torch.nn.functional as F
from modules.commons.common_layers import Embedding
from modules.fastspeech.tts_modules import LayerNorm
class LambdaLayer(nn.Module):
def __init__(self, lambd):
super(LambdaLayer, self).__init__()
self.lambd = lambd
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/conv.py |
import math
import torch
class PositionalEncoding(torch.nn.Module):
"""Positional encoding.
Args:
d_model (int): Embedding dimension.
dropout_rate (float): Dropout rate.
max_len (int): Maximum input length.
reverse (bool): Whether to reverse the input position.
"""
def... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/espnet_positional_embedding.py |
import scipy
from torch.nn import functional as F
import torch
from torch import nn
import numpy as np
from modules.commons.wavenet import WN
from modules.glow import utils
class ActNorm(nn.Module):
def __init__(self, channels, ddi=False, **kwargs):
super().__init__()
self.channels = channels
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/normalizing_flow/glow_modules.py |
import torch
from torch import nn
from modules.commons.conv import ConditionalConvBlocks
from modules.commons.wavenet import WN
class FlipLayer(nn.Module):
def forward(self, x, nonpadding, cond=None, reverse=False):
x = torch.flip(x, [1])
return x
class CouplingLayer(nn.Module):
def __init__... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/normalizing_flow/res_flow.py |
import torch
def squeeze(x, x_mask=None, n_sqz=2):
b, c, t = x.size()
t = (t // n_sqz) * n_sqz
x = x[:, :, :t]
x_sqz = x.view(b, c, t // n_sqz, n_sqz)
x_sqz = x_sqz.permute(0, 3, 1, 2).contiguous().view(b, c * n_sqz, t // n_sqz)
if x_mask is not None:
x_mask = x_mask[:, :, n_sqz - 1:... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/commons/normalizing_flow/utils.py |
import numpy as np
import torch
import torch.utils.data
from librosa.filters import mel as librosa_mel_fn
from scipy.io.wavfile import read
MAX_WAV_VALUE = 32768.0
def load_wav(full_path):
sampling_rate, data = read(full_path)
return data, sampling_rate
def dynamic_range_compression(x, C=1, clip_val=1e-5):... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/hifigan/mel_utils.py |
import torch
import torch.nn.functional as F
import torch.nn as nn
from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d
from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm
from modules.parallel_wavegan.layers import UpsampleNetwork, ConvInUpsampleNetwork
from modules.parallel_wavegan.m... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/hifigan/hifigan.py |
import math
import random
from functools import partial
from inspect import isfunction
from pathlib import Path
import numpy as np
import torch
import torch.nn.functional as F
from torch import nn
from tqdm import tqdm
from einops import rearrange
from modules.fastspeech.fs2 import FastSpeech2
from modules.diffsinger_... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/diff/diffusion.py |
import math
import torch
import torch.nn as nn
import torch.nn.functional as F
from math import sqrt
from .diffusion import Mish
from utils.hparams import hparams
Linear = nn.Linear
ConvTranspose2d = nn.ConvTranspose2d
class AttrDict(dict):
def __init__(self, *args, **kwargs):
super(AttrDict, self).__... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/diff/net.py |
import math
import random
from collections import deque
from functools import partial
from inspect import isfunction
from pathlib import Path
import numpy as np
import torch
import torch.nn.functional as F
from torch import nn
from tqdm import tqdm
from einops import rearrange
from modules.fastspeech.fs2 import FastSp... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/diff/shallow_diffusion_tts.py |
from modules.fastspeech.tts_modules import FastspeechDecoder
# from modules.fastspeech.fast_tacotron import DecoderRNN
# from modules.fastspeech.speedy_speech.speedy_speech import ConvBlocks
# from modules.fastspeech.conformer.conformer import ConformerDecoder
import torch
from torch.nn import functional as F
import to... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/diff/candidate_decoder.py |
import matplotlib
matplotlib.use('Agg')
from data_gen.tts.data_gen_utils import get_pitch
from modules.fastspeech.tts_modules import mel2ph_to_dur
import matplotlib.pyplot as plt
from utils import audio
from utils.pitch_utils import norm_interp_f0, denorm_f0, f0_to_coarse
from vocoders.base_vocoder import get_vocoder_c... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/GenerSpeech/task/generspeech.py |
import matplotlib
matplotlib.use('Agg')
from tasks.base_task import data_loader
from tasks.tts.fs2 import FastSpeech2Task
from tasks.tts.dataset_utils import FastSpeechDataset, BaseTTSDataset
import glob
import importlib
from utils.pitch_utils import norm_interp_f0, denorm_f0, f0_to_coarse
from inference.base_tts_infer... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/GenerSpeech/task/dataset.py |
import scipy
from torch.nn import functional as F
import torch
from torch import nn
import numpy as np
from modules.commons.common_layers import Permute
from modules.fastspeech.tts_modules import FFTBlocks
from modules.GenerSpeech.model.wavenet import fused_add_tanh_sigmoid_multiply, WN
class LayerNorm(nn.Module):
... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/GenerSpeech/model/glow_modules.py |
from modules.commons.common_layers import *
# @torch.jit.script
def fused_add_tanh_sigmoid_multiply(input_a, input_b, n_channels):
n_channels_int = n_channels[0]
in_act = input_a + input_b
t_act = torch.tanh(in_act[:, :n_channels_int, :])
s_act = torch.sigmoid(in_act[:, n_channels_int:, :])
acts =... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/GenerSpeech/model/wavenet.py |
from torch import nn
import copy
import torch
from utils.hparams import hparams
from modules.GenerSpeech.model.wavenet import WN
import math
from modules.fastspeech.tts_modules import LayerNorm
import torch.nn.functional as F
from utils.tts_utils import group_hidden_by_segs, sequence_mask
from scipy.cluster.vq import... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/GenerSpeech/model/prosody_util.py |
import torch
from modules.GenerSpeech.model.glow_modules import Glow
from modules.fastspeech.tts_modules import PitchPredictor
import random
from modules.GenerSpeech.model.prosody_util import ProsodyAligner, LocalStyleAdaptor
from utils.pitch_utils import f0_to_coarse, denorm_f0
from modules.commons.common_layers impor... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/GenerSpeech/model/generspeech.py |
from modules.commons.common_layers import *
import random
class MixStyle(nn.Module):
"""MixStyle.
Reference:
Zhou et al. Domain Generalization with MixStyle. ICLR 2021.
"""
def __init__(self, p=0.5, alpha=0.1, eps=1e-6, hidden_size=256):
"""
Args:
p (float): probabilit... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/GenerSpeech/model/mixstyle.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
import dgl
from dgl.nn.pytorch import GatedGraphConv
def sequence_mask(lengths, maxlen, dtype=torch.bool):
if maxlen is None:
maxlen = lengths.max()
mask = ~(torch.ones((len(lengths), maxlen)).to(lengths.device).cumsum(dim=1).t() > len... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/syntaspeech/syntactic_graph_encoder.py |
from copy import deepcopy
import torch
import dgl
import stanza
import networkx as nx
class Sentence2GraphParser:
def __init__(self, language='zh', use_gpu=False, download=False):
self.language = language
if download:
self.stanza_parser = stanza.Pipeline(lang=language, use_gpu=use_gpu)... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/syntaspeech/syntactic_graph_buider.py |
import numpy as np
import torch
import torch.nn as nn
class SingleWindowDisc(nn.Module):
def __init__(self, time_length, freq_length=80, kernel=(3, 3), c_in=1, hidden_size=128):
super().__init__()
padding = (kernel[0] // 2, kernel[1] // 2)
self.model = nn.ModuleList([
nn.Sequen... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/syntaspeech/multi_window_disc.py |
import math
import torch
from torch import nn
from torch.nn import Linear
from utils.hparams import hparams
from modules.commons.conv import ConvBlocks, ConditionalConvBlocks
from modules.commons.common_layers import Embedding
from modules.commons.rel_transformer import RelTransformerEncoder
from modules.commons.transf... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/syntaspeech/syntaspeech.py |
from utils.hparams import hparams
from modules.commons.common_layers import *
from modules.commons.common_layers import Embedding
from modules.fastspeech.tts_modules import FastspeechDecoder, DurationPredictor, LengthRegulator, PitchPredictor, \
EnergyPredictor, FastspeechEncoder
from utils.cwt import cwt2f0
from u... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/fastspeech/fs2.py |
import logging
import math
import torch
import torch.nn as nn
from torch.nn import functional as F
from modules.commons.espnet_positional_embedding import RelPositionalEncoding
from modules.commons.common_layers import SinusoidalPositionalEmbedding, Linear, EncSALayer, DecSALayer, BatchNorm1dTBC
from utils.hparams im... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/fastspeech/tts_modules.py |
from modules.commons.common_layers import *
from utils.hparams import hparams
from modules.fastspeech.tts_modules import PitchPredictor
from utils.pitch_utils import denorm_f0
class Prenet(nn.Module):
def __init__(self, in_dim=80, out_dim=256, kernel=5, n_layers=3, strides=None):
super(Prenet, self).__ini... | EXA-1-master | exa/models/AudioGPT/NeuralSeq/modules/fastspeech/pe.py |
import numpy as np
import torch
import torch.nn.functional as F
from torch import nn
from transformers import AutoModel
from .audio import get_audio_encoder
class Projection(nn.Module):
def __init__(self, d_in: int, d_out: int, p: float=0.5) -> None:
super().__init__()
self.linear1 = nn.Linear(d_in... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/wav_evaluation/models/clap.py |
from . import clap
from . import audio
from . import utils | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/wav_evaluation/models/__init__.py |
import argparse
import yaml
import sys
def read_config_as_args(config_path,args=None,is_config_str=False):
return_dict = {}
if config_path is not None:
if is_config_str:
yml_config = yaml.load(config_path, Loader=yaml.FullLoader)
else:
with open(config_path, "r") as f:
... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/wav_evaluation/models/utils.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
from torchlibrosa.stft import Spectrogram, LogmelFilterBank
def get_audio_encoder(name: str):
if name == "Cnn14":
return Cnn14
else:
raise Exception('The audio encoder name {} is incorrect or not supported'.format(name))
class... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/wav_evaluation/models/audio.py |
import random
import torchaudio
from torch._six import string_classes
import collections
import re
import torch.nn.functional as F
import numpy as np
from transformers import AutoTokenizer
from wav_evaluation.models.utils import read_config_as_args
from wav_evaluation.models.clap import CLAP
import math
import torchau... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/wav_evaluation/models/CLAPWrapper.py |
import os
import torch
import torch.nn.functional as F
import torch.nn as nn
from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d
from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm
from pathlib import Path
import yaml
import numpy as np
from argparse import Namespace
LRELU_SLOPE = 0.1
... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/vocoder/hifigan/modules.py |
import numpy as np
class LambdaWarmUpCosineScheduler:
"""
note: use with a base_lr of 1.0
"""
def __init__(self, warm_up_steps, lr_min, lr_max, lr_start, max_decay_steps, verbosity_interval=0):
self.lr_warm_up_steps = warm_up_steps
self.lr_start = lr_start
self.lr_min = lr_min
... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/lr_scheduler.py |
import importlib
import torch
import numpy as np
from tqdm import tqdm
from inspect import isfunction
from PIL import Image, ImageDraw, ImageFont
import hashlib
import requests
import os
URL_MAP = {
'vggishish_lpaps': 'https://a3s.fi/swift/v1/AUTH_a235c0f452d648828f745589cde1219a/specvqgan_public/vggishish16.pt',... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/util.py |
"""
与autoencoder.py的区别在于,autoencoder.py计算loss时只有一个discriminator,而此处又多了个multiwindowDiscriminator,所以优化器
优化的参数改为:
opt_disc = torch.optim.Adam(list(self.loss.discriminator.parameters()) + list(self.loss.discriminator_multi.parameters()),
lr=lr, betas=(0.5, 0.9))
"""
import os
import torch
impor... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/autoencoder_multi.py |
import os
import torch
import pytorch_lightning as pl
import torch.nn.functional as F
from contextlib import contextmanager
from packaging import version
import numpy as np
from ldm.modules.diffusionmodules.model import Encoder, Decoder
from ldm.modules.distributions.distributions import DiagonalGaussianDistribution
fr... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/autoencoder.py |
"""SAMPLING ONLY."""
import torch
import numpy as np
from tqdm import tqdm
from functools import partial
from ldm.modules.diffusionmodules.util import make_ddim_sampling_parameters, make_ddim_timesteps, noise_like, \
extract_into_tensor
class DDIMSampler(object):
def __init__(self, model, schedule="linear",... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/diffusion/ddim.py |
import os
import torch
import pytorch_lightning as pl
from omegaconf import OmegaConf
from torch.nn import functional as F
from torch.optim import AdamW
from torch.optim.lr_scheduler import LambdaLR
from copy import deepcopy
from einops import rearrange
from glob import glob
from natsort import natsorted
from ldm.modu... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/diffusion/classifier.py |
"""
wild mixture of
https://github.com/lucidrains/denoising-diffusion-pytorch/blob/7706bdfc6f527f58d33f84b7b522e61e6e3164b3/denoising_diffusion_pytorch/denoising_diffusion_pytorch.py
https://github.com/openai/improved-diffusion/blob/e94489283bb876ac1477d5dd7709bbbd2d9902ce/improved_diffusion/gaussian_diffusion.py
https... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/diffusion/ddpm_audio.py |
EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/diffusion/__init__.py | |
"""
wild mixture of
https://github.com/lucidrains/denoising-diffusion-pytorch/blob/7706bdfc6f527f58d33f84b7b522e61e6e3164b3/denoising_diffusion_pytorch/denoising_diffusion_pytorch.py
https://github.com/openai/improved-diffusion/blob/e94489283bb876ac1477d5dd7709bbbd2d9902ce/improved_diffusion/gaussian_diffusion.py
https... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/diffusion/ddpm_audio_inpaint.py |
"""SAMPLING ONLY."""
import torch
import numpy as np
from tqdm import tqdm
from functools import partial
from ldm.modules.diffusionmodules.util import make_ddim_sampling_parameters, make_ddim_timesteps, noise_like
class PLMSSampler(object):
def __init__(self, model, schedule="linear", **kwargs):
super()... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/diffusion/plms.py |
"""
wild mixture of
https://github.com/lucidrains/denoising-diffusion-pytorch/blob/7706bdfc6f527f58d33f84b7b522e61e6e3164b3/denoising_diffusion_pytorch/denoising_diffusion_pytorch.py
https://github.com/openai/improved-diffusion/blob/e94489283bb876ac1477d5dd7709bbbd2d9902ce/improved_diffusion/gaussian_diffusion.py
https... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/models/diffusion/ddpm.py |
from inspect import isfunction
import math
import torch
import torch.nn.functional as F
from torch import nn, einsum
from einops import rearrange, repeat
from ldm.modules.diffusionmodules.util import checkpoint
def exists(val):
return val is not None
def uniq(arr):
return{el: True for el in arr}.keys()
d... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/attention.py |
"""shout-out to https://github.com/lucidrains/x-transformers/tree/main/x_transformers"""
import torch
from torch import nn, einsum
import torch.nn.functional as F
from functools import partial
from inspect import isfunction
from collections import namedtuple
from einops import rearrange, repeat, reduce
# constants
DE... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/x_transformer.py |
import torch
from torch import nn
class LitEma(nn.Module):
def __init__(self, model, decay=0.9999, use_num_upates=True):
super().__init__()
if decay < 0.0 or decay > 1.0:
raise ValueError('Decay must be between 0 and 1')
self.m_name2s_name = {}
self.register_buffer('de... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/ema.py |
import functools
import torch.nn as nn
class ActNorm(nn.Module):
def __init__(self, num_features, logdet=False, affine=True,
allow_reverse_init=False):
assert affine
super().__init__()
self.logdet = logdet
self.loc = nn.Parameter(torch.zeros(1, num_features, 1, 1))... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/discriminator/model.py |
import numpy as np
import torch
import torch.nn as nn
class Discriminator2DFactory(nn.Module):
def __init__(self, time_length, freq_length=80, kernel=(3, 3), c_in=1, hidden_size=128,
norm_type='bn', reduction='sum'):
super(Discriminator2DFactory, self).__init__()
padding = (kernel... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/discriminator/multi_window_disc.py |
"""
Based on https://github.com/CompVis/taming-transformers/blob/52720829/taming/modules/losses/lpips.py
Adapted for spectrograms by Vladimir Iashin (v-iashin)
"""
from collections import namedtuple
import numpy as np
import torch
import torch.nn as nn
import sys
sys.path.insert(0, '.') # nopep8
from ldm.mod... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/lpaps.py |
from ldm.modules.losses_audio.vqperceptual import DummyLoss
# relative imports pain
import os
import sys
path = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'vggishish')
sys.path.append(path)
| EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/__init__.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
import sys
from ldm.util import exists
sys.path.insert(0, '.') # nopep8
from ldm.modules.discriminator.model import (NLayerDiscriminator, NLayerDiscriminator1dFeats,
NLayerDiscriminator1dSpecs,
... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vqperceptual.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
import sys
sys.path.insert(0, '.') # nopep8
from ldm.modules.losses_audio.vqperceptual import *
class LPAPSWithDiscriminator(nn.Module):
def __init__(self, disc_start, logvar_init=0.0, kl_weight=1.0, pixelloss_weight=1.0,
disc_n... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/contperceptual.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
import sys
sys.path.insert(0, '.') # nopep8
from ldm.modules.losses_audio.vqperceptual import *
from ldm.modules.discriminator.multi_window_disc import Discriminator
class LPAPSWithDiscriminator(nn.Module):# 相比于contperceptual.py添加了MultiWindowDiscrimi... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/contperceptual_dis.py |
import logging
import numpy as np
import scipy
import torch
from sklearn.metrics import average_precision_score, roc_auc_score
logger = logging.getLogger(f'main.{__name__}')
def metrics(targets, outputs, topk=(1, 5)):
"""
Adapted from https://github.com/hche11/VGGSound/blob/master/utils.py
Calculate sta... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/metrics.py |
import logging
import os
from pathlib import Path
import albumentations
import numpy as np
import torch
from tqdm import tqdm
logger = logging.getLogger(f'main.{__name__}')
class StandardNormalizeAudio(object):
'''
Frequency-wise normalization
'''
def __init__(self, specs_dir, train_ids_path='./... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/transforms.py |
import os
from torch.utils.data import DataLoader
import torchvision
from tqdm import tqdm
from dataset import VGGSound
import torch
import torch.nn as nn
from metrics import metrics
from omegaconf import OmegaConf
from model import VGGishish
from transforms import Crop, StandardNormalizeAudio, ToTensor
if __name__ =... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/predict.py |
import logging
import os
import time
from shutil import copytree, ignore_patterns
import torch
from omegaconf import OmegaConf
from torch.utils.tensorboard import SummaryWriter, summary
class LoggerWithTBoard(SummaryWriter):
def __init__(self, cfg):
# current time stamp and experiment log directory
... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/logger.py |
import torch
import torch.nn as nn
class VGGishish(nn.Module):
def __init__(self, conv_layers, use_bn, num_classes):
'''
Mostly from
https://pytorch.org/vision/0.8/_modules/torchvision/models/vgg.html
'''
super().__init__()
layers = []
in_channels = 1
... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/model.py |
import collections
import csv
import logging
import os
import random
from glob import glob
from pathlib import Path
import numpy as np
import torch
import torchvision
logger = logging.getLogger(f'main.{__name__}')
class VGGSound(torch.utils.data.Dataset):
def __init__(self, split, specs_dir, transforms=None, s... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/dataset.py |
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
class WeightedCrossEntropy(nn.CrossEntropyLoss):
def __init__(self, weights, **pytorch_ce_loss_args) -> None:
super().__init__(reduction='none', **pytorch_ce_loss_args)
self.weights = weights
def __... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/loss.py |
from loss import WeightedCrossEntropy
import random
import numpy as np
import torch
import torchvision
from omegaconf import OmegaConf
from torch.utils.data.dataloader import DataLoader
from tqdm import tqdm
from dataset import VGGSound
from transforms import Crop, StandardNormalizeAudio, ToTensor
from logger import ... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/train_vggishish.py |
import random
import numpy as np
import torch
import torchvision
from omegaconf import OmegaConf
from torch.utils.data.dataloader import DataLoader
from torchvision.models.inception import BasicConv2d, Inception3
from tqdm import tqdm
from dataset import VGGSound
from logger import LoggerWithTBoard
from loss import W... | EXA-1-master | exa/models/AudioGPT/text_to_audio/MakeAnAudio/ldm/modules/losses_audio/vggishish/train_melception.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.