Click any tag below to further narrow down your results
Links
This article breaks down the core concepts behind LLMs—from next-token prediction training to tokens, vectors and attention layers—to show how they generate text. It also covers context windows, parameters and why model scale affects performance.
Anthropic has published a constitution for its AI model, Claude, detailing the values and behaviors it should embody. This document serves as a guiding framework for Claude's training and decision-making processes, focusing on safety, ethics, and helpfulness.
The Smol Training Playbook on Hugging Face provides a comprehensive guide for efficiently training machine learning models using the Hugging Face ecosystem. It emphasizes best practices and methodologies for optimizing training processes, making it accessible for both beginners and experienced practitioners. The playbook also includes practical examples and resources to enhance the learning experience.