MedReason is a comprehensive medical reasoning dataset that enhances large language models (LLMs) by utilizing a structured medical knowledge graph to create detailed reasoning paths from clinical question-answer pairs. The dataset includes 32,682 QA pairs with step-by-step explanations, and the MedReason-8B model, fine-tuned on this data, achieves state-of-the-art performance in medical reasoning tasks. The project is open-sourced, providing access to models, data, and deployment codes for further research and applications.
REverse-Engineered Reasoning (REER) introduces a novel approach to instilling deep reasoning in language models by working backwards from known solutions to discover the underlying reasoning process. This method addresses the limitations of traditional reinforcement learning and instruction distillation, resulting in the creation of a large dataset, DeepWriting-20K, and a model, DeepWriter-8B, that outperforms existing models in open-ended tasks. The research emphasizes the importance of structured reasoning and iterative refinement in generating high-quality outputs.