Click any tag below to further narrow down your results
Links
This article outlines Distribution-Aligned Sequence Distillation, a new pipeline for improving reasoning tasks like math and code generation using minimal training data. It introduces models such as DASD-4B-Thinking and DASD-30B-A3B-Thinking-Preview, which outperform larger models in various benchmarks. The methodology includes temperature-scheduled learning and mixed-policy distillation for better performance.