1 link tagged with all of: benchmarks + data-efficiency + reasoning + distillation + models
Links
This article outlines Distribution-Aligned Sequence Distillation, a new pipeline for improving reasoning tasks like math and code generation using minimal training data. It introduces models such as DASD-4B-Thinking and DASD-30B-A3B-Thinking-Preview, which outperform larger models in various benchmarks. The methodology includes temperature-scheduled learning and mixed-policy distillation for better performance.
reasoning ✓
distillation ✓
models ✓
benchmarks ✓
data-efficiency ✓