tawheed-tariq commited on
Commit
2d66b28
·
verified ·
1 Parent(s): 02961c4

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -0
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ pipeline_tag: video-text-to-text
4
+ tags:
5
+ - multi-object-tracking
6
+ - video-understanding
7
+ - vision-language-model
8
+ - spatiotemporal-reasoning
9
+ ---
10
+
11
+ # QTrack: Query-Driven Reasoning for Multi-modal MOT
12
+
13
+ [**QTrack**](https://huggingface.co/papers/2603.13759) is an end-to-end vision-language model designed for query-driven multi-object tracking (MOT). Unlike traditional MOT which tracks all objects in a scene, QTrack selectively localizes and tracks specific targets based on natural language instructions while maintaining temporal coherence and identity consistency.
14
+
15
+ - **Paper:** [QTrack: Query-Driven Reasoning for Multi-modal MOT](https://huggingface.co/papers/2603.13759)
16
+ - **Project Page:** [https://gaash-lab.github.io/QTrack/](https://gaash-lab.github.io/QTrack/)
17
+ - **Repository:** [https://github.com/gaash-lab/QTrack](https://github.com/gaash-lab/QTrack)
18
+
19
+ ## Description
20
+
21
+ Multi-object tracking has traditionally focused on estimating trajectories of all objects. QTrack introduces a **query-driven tracking paradigm** that formulates tracking as a spatiotemporal reasoning problem conditioned on natural language queries.
22
+
23
+ ### Key Contributions
24
+ - **RMOT26 Benchmark**: A large-scale benchmark with grounded queries and sequence-level splits to enable robust evaluation of generalization.
25
+ - **QTrack Model**: An end-to-end vision-language model that integrates multimodal reasoning with tracking-oriented localization.
26
+ - **Temporal Perception-Aware Policy Optimization (TPA-PO)**: A structured reward strategy to encourage motion-aware reasoning.
27
+
28
+ ## Benchmark Results
29
+
30
+ QTrack achieves state-of-the-art performance on the [RMOT26](https://huggingface.co/datasets/GAASH-Lab/RMOT26) benchmark.
31
+
32
+ | Model | Params | MCP↑ | MOTP↑ | CLE (px)↓ | NDE↓ |
33
+ |:-----:|:------:|:----:|:-----:|:---------:|:----:|
34
+ | GPT-5.2 | - | 0.25 | 0.61 | 94.2 | 0.55 |
35
+ | Qwen3-VL-Instruct | 8B | 0.25 | 0.64 | 96.0 | 0.97 |
36
+ | Gemma 3 | 27B | 0.24 | 0.56 | 58.4 | 0.88 |
37
+ | InternVL | 8B | 0.21 | 0.66 | 117.44 | 0.64 |
38
+ | **QTrack (Ours)** | **3B** | **0.30** | **0.75** | **44.61** | **0.39** |
39
+
40
+ ## Installation
41
+
42
+ To set up the environment and use the model, please follow the instructions in the [official repository](https://github.com/gaash-lab/QTrack):
43
+
44
+ ```bash
45
+ # Create conda environment
46
+ conda create -n qtrack python=3.12
47
+ conda activate qtrack
48
+
49
+ # Install QTrack and dependencies
50
+ git clone https://github.com/gaash-lab/QTrack.git
51
+ cd QTrack
52
+ pip install -r requirements.txt
53
+ pip install -e .
54
+ ```
55
+
56
+ ## Citation
57
+
58
+ If you find QTrack useful for your research, please cite:
59
+
60
+ ```bibtex
61
+ @article{ashraf2026qtrack,
62
+ title={QTrack: Query-Driven Reasoning for Multi-modal MOT},
63
+ author={Ashraf, Tajamul and Tariq, Tavaheed and Yadav, Sonia and Ul Riyaz, Abrar and Tak, Wasif and Abdar, Moloud and Bashir, Janibul},
64
+ journal={arXiv preprint arXiv:2603.13759},
65
+ year={2026}
66
+ }
67
+ ```