Helium 1 is a newly released language model with 2 billion parameters, optimized for multilingual performance and designed for efficient on-device deployment. It leverages a high-quality training dataset created through a comprehensive data processing pipeline and aims to democratize access to AI technologies across European languages. The model architecture is based on transformers, and the project includes tools for reproducing the training dataset and specialized model development.
helium-1 ✓
language-model ✓
+ multilingual
ai-democratization ✓
data-processing ✓