SCFM: Shortcutting Pre-trained Flow Matching Diffusion Models is Almost Free Lunch

Xu Cai∗†, Yang Wu, Qianli Chen, Haoran Wu, Lichuan Xiang§, Hongkai Wen§

NeurIPS 2025

🔗 Resources

arXiv arXiv | Code | CivitAI CivitAI

🚀 TL;DR

We introduce SCFM — a highly efficient post-training distillation method that converts any pre-trained flow matching diffusion model (e.g., Flux, SD3) into a 3–8 step sampler in <1 A100 day.

💡 Key Contributions

📊 Main Results (Flux.1-Dev → 3 Steps)

Method Steps Latency (A100, s) |ΔFID| ↓ FID ↓ CLIP ↑
Flux-HyperSD 3 1.33 1.52 9.65 31.95
Flux-TDD 3 1.33 4.46 8.26 31.38
Flux-SCFM (Ours) 3 1.33 1.01 6.34 33.10
Flux-Schnell (Official) 3 1.33 6.58 7.06 33.06

SCFM achieves the best FID and CLIP scores among 3-step distilled models — and does so without adversarial distillation (ADD/LADD).

🖼️ Visual Comparison

Visual Comparison

📦 Get Started

🔗 Resources

📬 Contact

Questions? Reach out to the corresponding author: caitreex@gmail.com