Update license, add pipeline tag and improve model card
Browse filesHi! This is Niels from the community science team at Hugging Face.
This PR improves the model card for SCAIL by:
- Adding the `image-to-video` pipeline tag for better discoverability.
- Updating the license to `apache-2.0` to match the official GitHub repository.
- Adding descriptive tags and linking author Hugging Face profiles.
- Linking the paper to the Hugging Face papers page in the markdown section.
README.md
CHANGED
|
@@ -1,12 +1,18 @@
|
|
| 1 |
---
|
| 2 |
-
license:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
| 4 |
# SCAIL: Towards Studio-Grade Character Animation via In-Context Learning of 3D-Consistent Pose Representations
|
| 5 |
|
| 6 |
|
| 7 |
<div align="center">
|
| 8 |
-
<a href="https://
|
| 9 |
-
<img src="https://img.shields.io/badge/π%
|
| 10 |
</a>
|
| 11 |
<a href="https://teal024.github.io/SCAIL/">
|
| 12 |
<img src="https://img.shields.io/badge/π%20Project%20Page-green">
|
|
@@ -17,14 +23,20 @@ license: mit
|
|
| 17 |
</div>
|
| 18 |
|
| 19 |
|
| 20 |
-
|
| 21 |
|
| 22 |
-
|
| 23 |
-
|
|
|
|
|
|
|
| 24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
|
| 26 |
## π Note
|
| 27 |
-
This repository
|
| 28 |
|
| 29 |
## π Citation
|
| 30 |
|
|
@@ -33,8 +45,11 @@ If you find this work useful in your research, please cite:
|
|
| 33 |
```bibtex
|
| 34 |
@article{yan2025scail,
|
| 35 |
title={SCAIL: Towards Studio-Grade Character Animation via In-Context Learning of 3D-Consistent Pose Representations},
|
| 36 |
-
author={Yan, Wenhao and Ye, Sheng and Yang, Zhuoyi and Teng, Jiayan and Dong, ZhenHui and Wen, Kairui and Gu, Xiaotao and Liu, Yong-Jin and Tang, Jie},
|
| 37 |
journal={arXiv preprint arXiv:2512.05905},
|
| 38 |
year={2025}
|
| 39 |
}
|
| 40 |
```
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
pipeline_tag: image-to-video
|
| 4 |
+
tags:
|
| 5 |
+
- character-animation
|
| 6 |
+
- 3d-pose
|
| 7 |
+
- motion-transfer
|
| 8 |
---
|
| 9 |
+
|
| 10 |
# SCAIL: Towards Studio-Grade Character Animation via In-Context Learning of 3D-Consistent Pose Representations
|
| 11 |
|
| 12 |
|
| 13 |
<div align="center">
|
| 14 |
+
<a href="https://huggingface.co/papers/2512.05905">
|
| 15 |
+
<img src="https://img.shields.io/badge/π%20Paper-2512.05905-red">
|
| 16 |
</a>
|
| 17 |
<a href="https://teal024.github.io/SCAIL/">
|
| 18 |
<img src="https://img.shields.io/badge/π%20Project%20Page-green">
|
|
|
|
| 23 |
</div>
|
| 24 |
|
| 25 |
|
| 26 |
+
**SCAIL** (Studio-grade Character Animation via In-context Learning) is a framework designed to achieve high-fidelity character animation that meets studio standards. It addresses challenges in preserving structural fidelity and temporal consistency during motion transfer, especially in complex scenarios involving large motions and multi-character interactions.
|
| 27 |
|
| 28 |
+
Key features include:
|
| 29 |
+
- **3D-Consistent Pose Representations**: Provides a robust and flexible motion signal while preventing identity leakage.
|
| 30 |
+
- **Full-Context Pose Injection**: Enables effective spatio-temporal reasoning over entire motion sequences within a diffusion-transformer.
|
| 31 |
+
- **Studio-Grade Quality**: Trained on a curated data pipeline to ensure diversity and quality.
|
| 32 |
|
| 33 |
+
This repository contains the model weights for the **SCAIL-Preview (14B)** model.
|
| 34 |
+
|
| 35 |
+
## π Project Page
|
| 36 |
+
Check the model architecture design, video demos, and comparisons against other baselines at the [official project page](https://teal024.github.io/SCAIL/).
|
| 37 |
|
| 38 |
## π Note
|
| 39 |
+
This repository contains the model weights for the SCAIL model. For model inference, environment setup, and the roadmap, please refer to the [official repository](https://github.com/teal024/SCAIL-Official). For pose extraction tools, refer to [SCAIL-Pose](https://github.com/teal024/SCAIL-Pose).
|
| 40 |
|
| 41 |
## π Citation
|
| 42 |
|
|
|
|
| 45 |
```bibtex
|
| 46 |
@article{yan2025scail,
|
| 47 |
title={SCAIL: Towards Studio-Grade Character Animation via In-Context Learning of 3D-Consistent Pose Representations},
|
| 48 |
+
author={Yan, Wenhao and Ye, Sheng English and Yang, Zhuoyi and Teng, Jiayan and Dong, ZhenHui and Wen, Kairui and Gu, Xiaotao and Liu, Yong-Jin and Tang, Jie},
|
| 49 |
journal={arXiv preprint arXiv:2512.05905},
|
| 50 |
year={2025}
|
| 51 |
}
|
| 52 |
```
|
| 53 |
+
|
| 54 |
+
## π₯ Authors
|
| 55 |
+
[Wenhao Yan](https://huggingface.co/wenhaoyan77), Sheng Ye, Zhuoyi Yang, [Jiayan Teng](https://huggingface.co/tengjiayan), ZhenHui Dong, [Kairui Wen](https://huggingface.co/SKearbvaanl), [Xiaotao Gu](https://huggingface.co/xgeric), Yong-Jin Liu, [Jie Tang](https://huggingface.co/jerytang).
|