OLMo 2 is a family of fully-open language models designed for accessibility and reproducibility in AI research. The largest model, OLMo 2 32B, surpasses GPT-3.5-Turbo and GPT-4o mini on various academic benchmarks, while the smaller models (7B, 13B, and 1B) are competitive with other open-weight models. Ai2 emphasizes the importance of open training data and code to advance collective scientific research.
OLMoTrace is a new feature in the Ai2 Playground that allows users to trace the outputs of language models back to their extensive training data, enhancing transparency and trust. It enables researchers and the public to inspect how specific word sequences were generated, facilitating fact-checking and understanding model capabilities. The tool showcases Ai2's commitment to an open ecosystem by making training data accessible for scientific research and public insight into AI systems.