Text Generation
Transformers
Safetensors
English
qwen2
conversational
text-generation-inference

SWE-Next: Scalable Real-World Software Engineering Tasks for Agents

Paper Project Page Code SFT Trajs Dataset Model 7B Model 14B

SWE-Next-14B

SWE-Next-14B is a repository-level software engineering agent fine-tuned from Qwen/Qwen2.5-Coder-14B-Instruct on the released SWE-Next SFT Trajectories. The model is trained with full-parameter supervised fine-tuning on execution-grounded trajectories collected from real merged pull requests and validated repository environments.

Introduction

SWE-Next introduces reusable repo-quarter profiles, which reuse the same environment across nearby commits in time while keeping each task run separate and reproducible. Using only 30 hours and 639GB of environment storage, SWE-Next processes 3,971 seed repositories and 102,582 candidate commit pairs mined from real merged PRs to construct a dataset of 2,308 self-verifying instances. SWE-Next improves downstream pass@1 on SWE-Bench Verified and SWE-Bench Lite with fewer or comparable training trajectories, making large-scale executable data collection far more practical and accessible for research.

SWE-Next teaser

Model Overview

This model is trained on 3,693 selected SFT trajectories derived from the SWE-Next collection. The training data emphasizes clean repository-level repair traces and recovery-style debugging trajectories rather than isolated code-completion examples.

Training recipe summary:

  • Base model: Qwen/Qwen2.5-Coder-14B-Instruct
  • Finetuning: full-parameter SFT
  • Context length: 32,768
  • Learning rate: 1e-5
  • Scheduler: cosine
  • Dataset: TIGER-Lab/SWE-Next-SFT-Trajectories

Usage

For full usage details, please refer to the official SWE-Next GitHub repository. The repository provides the complete setup and evaluation workflow for released models, including:

  • environment and dependency installation,
  • dataset and trajectory downloads,
  • training configurations for the 7B and 14B models,
  • vLLM serving commands and repository-level evaluation scripts.

In particular, the GitHub repo contains the exact commands used to serve SWE-Next-14B and evaluate it on SWE-Bench-style tasks under the SWE-Next execution interface.

Relationship to the SWE-Next Release

This repo contains the released 14B model checkpoint. Related artifacts are available separately:

  • Base task dataset: TIGER-Lab/SWE-Next
  • SFT trajectories: TIGER-Lab/SWE-Next-SFT-Trajectories
  • Companion model: TIGER-Lab/SWE-Next-7B
  • Project code: github.com/TIGER-AI-Lab/SWE-Next

Citation

@misc{liang2026swenextscalablerealworldsoftware,
      title={SWE-Next: Scalable Real-World Software Engineering Tasks for Agents},
      author={Jiarong Liang and Zhiheng Lyu and Zijie Liu and Xiangchao Chen and Ping Nie and Kai Zou and Wenhu Chen},
      year={2026},
      eprint={2603.20691},
      archivePrefix={arXiv},
      primaryClass={cs.SE},
      url={https://arxiv.org/abs/2603.20691},
}
Downloads last month
41
Safetensors
Model size
841k params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for TIGER-Lab/SWE-Next-14B

Base model

Qwen/Qwen2.5-14B
Finetuned
(71)
this model
Quantizations
1 model

Datasets used to train TIGER-Lab/SWE-Next-14B

Collection including TIGER-Lab/SWE-Next-14B

Paper for TIGER-Lab/SWE-Next-14B