If you find this project useful, please consider giving a star βοΈ~
π₯ Authors
*Equal contribution β Corresponding author
π° News
- [2026-04] Technical report released on arXiv.
- [2026-04] Project page is now live.
- [2026-04] Inference code and pretrained models are released.
- [2026-04] Online demo is available on Project Inference Interface.
- [Coming Soon] Skill will be released.
π Updates
- Release technical report on arXiv.
- Launch project page.
- Release inference code and pretrained models.
- Launch online inference demo (available on project page).
- Release skill.
πΊ Intro Video
https://github.com/user-attachments/assets/d63e9837-a568-4521-9009-58b4105214a9
For more results of our model, visit Project Page. For comparison with other methods, visit Demo Page.
π§ Overview
ControlFoley is a unified and controllable multimodal video-to-audio (V2A) generation framework that enables precise control over generated audio using video, text, and reference audio.
Unlike existing methods that rely on a single modality or struggle under conflicting inputs, ControlFoley is designed to handle complex multimodal interactions and maintain strong controllability even when modalities are inconsistent.
π¨ Tease Figure
Left: Overview of the ControlFoley framework with three multimodal conditioning modes for controllable video-synchronized audio generation. Right: Performance radar chart of Video-to-Audio models.
π Capabilities
ControlFoley supports a wide range of applications:
π¬ Text-Video-to-Audio Generation (TV2A)
Video-content-adaptive dubbing and synchronized sound effect generation under text guidance.π Text-Controlled Video-to-Audio (TC-V2A)
Audio generation under videoβtext conflicts, with semantics consistent with text prompts and temporally synchronized with video contents.π§ Audio-Controlled Video-to-Audio (AC-V2A)
Audio generation conditioned on reference audio, with timbre consistent with the reference audio and temporally synchronized with video contents.π Text-to-Audio Generation (T2A)
Generate audio directly from text prompts as an additional capability of the unified framework.
π§ Key Innovations
Joint Visual Encoding for Robust Multimodal Control: Combines CLIP and CAV-MAE-ST representations to capture both vision-language and audio-visual correlations, improving robustness under modality conflict.
Timbre-Focused Reference Audio Control: Extracts global timbre representations while suppressing temporal cues, enabling precise acoustic style control without affecting synchronization.
Modality-Robust Training with Unified Alignment: Introduces all-modality dropout and a unified REPA objective to improve robustness across diverse modality combinations.
VGGSound-TVC Benchmark: A new benchmark for evaluating textual controllability under visual-text semantic conflicts.
π§ͺ VGGSound-TVC Benchmark
We propose VGGSound-TVC to evaluate text controllability under varying levels of visual-text conflict. In this dataset, textual descriptions of videos are reconstructed in accordance with the rules described below.
- L0 β No conflict, where the textual description is consistent with the video content.
- L1_subject β A mild semantic conflict introduced at the subject level, where the action description remains unchanged while the sounding subject is replaced.
- L1_action β A mild semantic conflict introduced at the action level, where the subject remains unchanged while the action description is modified.
- L2 β A moderate semantic conflict in which the textual description belongs to a different semantic category while still maintaining a similar temporal structure or acoustic rhythm.
- L3 β Strong conflict, where the textual description is randomly substituted.
This enables systematic analysis of modality dominance and controllability under increasing inconsistency. Example samples from VGGSound-TVC are as follows.
π Performance
ControlFoley achieves strong performance across multiple V2A tasks, demonstrating both high generation quality and robust controllability.
π¬ TV2A
ControlFoley achieves state-of-the-art performance across multiple benchmarks, including VGGSound-Test, Kling-Audio-Eval, and MovieGen-Audio-Bench.
- Highest CLAP scores (better semantic alignment)
- Lowest DeSync (better temporal synchronization)
- Best overall IS (better audio quality)βUp to 27% relative improvement (22.08 vs. 17.36 on VGGSound).
π TC-V2A
ControlFoley demonstrates strong textual controllability under increasing visual-text conflict.
- Maintains high CLAP (text alignment) across conflict levels
- Effectively reduces IB under conflict (less reliance on visual bias)
- Achieves better balance between controllability and generation quality
π§ AC-V2A
ControlFoley achieves the best performance across all evaluation metrics on the Greatest Hits dataset.
- Better timbre similarity (Resemblyzer)
- Better synchronization (DeSync)
- Higher audio quality (IS)
Notably, it outperforms CondFoleyGen, a specialized in-domain baseline, demonstrating strong generalization ability.
ControlFoley also demonstrates competitive or superior performance compared to strong proprietary systems such as Kling-Foley, highlighting its effectiveness as an open and controllable solution.
π Quick Start
π Prerequisites
- Python 3.10+
- PyTorch 2.5.1+
- CUDA 11.8+
- FFmpeg (conda install -c conda-forge ffmpeg)
π§± Installation
# Clone the repository
git clone https://github.com/xiaomi-research/controlfoley
cd controlfoley
# Create conda environment
conda create -n controlfoley python=3.10.16
conda activate controlfoley
# Install dependencies
pip install -r requirements.txt
# Download pretrained weights
pip install huggingface-hub==0.26.2
huggingface-cli download YJX-Xiaomi/ControlFoley --resume-download --local-dir model_weights --local-dir-use-symlinks False
Or you can download the weights from here and put them in the model_weights folder.
π¨ Inference
python demo.py [OPTIONS]
Options:
--video TEXT Path to the input video file. (default: None)
--audio TEXT Path to the input reference audio file. (default: None)
--prompt TEXT Textual prompt for audio generation. (default: None)
--negative_prompt TEXT Negative textual prompt for audio generation. (default: None)
--duration FLOAT Duration of the generated audio in seconds. (default: 8.0)
--output TEXT Output directory for generated audio files. (default: ./output)
π Supported Tasks
| Task | video | audio | prompt |
|---|---|---|---|
| TV2A | required | None | required |
| TC-V2A | required | None | required |
| AC-V2A | required | required | optional |
| V2A | required | None | None |
| T2A | None | None | required |
π Usage Examples
- TV2A
python demo.py --video "assets/001.mp4" --prompt "the skateboard wheels scraping and grinding on the ground." --duration 8.0 --output "./output"
- TC-V2A
python demo.py --video "assets/002.mp4" --prompt "man whistling." --duration 8.0 --output "./output"
- AC-V2A
python demo.py --video "assets/003.mp4" --audio "assets/003.wav" --duration 8.0 --output "./output"
- V2A
python demo.py --video "assets/004.mp4" --duration 8.0 --output "./output"
- T2A
python demo.py --prompt "A bird sings melodically in a forest." --duration 8.0 --output "./output"
π Citation
If you find this repository useful, please consider citing our paper:
@misc{yang2026controlfoleyunifiedcontrollablevideotoaudio,
title={ControlFoley: Unified and Controllable Video-to-Audio Generation with Cross-Modal Conflict Handling},
author={Jianxuan Yang and Xinyue Guo and Zhi Cheng and Kai Wang and Lipan Zhang and Jinjie Hu and Qiang Ji and Yihua Cao and Yihao Meng and Zhaoyue Cui and Mengmei Liu and Meng Meng and Jian Luan},
year={2026},
eprint={2604.15086},
archivePrefix={arXiv},
primaryClass={cs.MM},
url={https://arxiv.org/abs/2604.15086},
}
π License
This repository is licensed under the Apache License 2.0 and the model weights are licensed under the CC BY-NC 4.0.
π Acknowledgments
This project uses the following datasets:
VGGSound, Kling-Audio-Eval, The Greatest Hits (CC BY 4.0),
and MovieGen-Audio-Bench (CC BY-NC 4.0).
All resources are used for academic and non-commercial demonstration purposes only.
This project is inspired by the following works:
stable-audio-tools, MMAudio, Make-An-Audio 2, Synchformer, and audiocraft.
Thanks for their contributions.
π Contact
If you have any questions or suggestions, please feel free to contact us at yangjianxuan@xiaomi.com.
2026 ControlFoley Project. All Rights Reserved.