3 links tagged with all of: deep-learning + language-models
Click any tag below to further narrow down your results
Links
This article discusses the significance of the Chain Rule of Probability and the Chain Rule of Calculus in machine learning advancements. It explains how these rules help compute complex probabilities in language models by breaking them down into smaller events, like predicting tokens based on previous ones. The author also highlights notable achievements in deep learning and diversity efforts within the AI community.
The article describes the implementation of the DeepSeek R1-zero style training for large language models (LLMs) using a single or multiple GPUs, with a focus on simplicity and efficiency. It highlights the capabilities of the nanoAhaMoment project, which includes full parameter tuning, multi-GPU support, and a full evaluation suite, while maintaining competitive performance with minimal complexity. The repository offers interactive Jupyter notebooks and scripts for training, complete with installation instructions and dependency management.
REverse-Engineered Reasoning (REER) introduces a novel approach to instilling deep reasoning in language models by working backwards from known solutions to discover the underlying reasoning process. This method addresses the limitations of traditional reinforcement learning and instruction distillation, resulting in the creation of a large dataset, DeepWriting-20K, and a model, DeepWriter-8B, that outperforms existing models in open-ended tasks. The research emphasizes the importance of structured reasoning and iterative refinement in generating high-quality outputs.