OLMo 2 1B is the smallest model in the OLMo 2 family, featuring a transformer-style architecture with 4 trillion training tokens. It supports multiple models and fine-tuning options, and is designed for language modeling applications. The model and its associated resources are available on GitHub under an Apache 2.0 license.
NVIDIA's Nemotron-H-8B-Base-8K is a large language model designed for text completion, featuring a hybrid architecture and a context length of 8K. It supports multiple languages and offers customization tools through the NeMo Framework for enhanced performance in research and development. The model is intended for use on NVIDIA GPU-accelerated systems and is part of the Nemotron-H collection, governed by specific licensing terms.