Click any tag below to further narrow down your results
Links
The author details their process of building a domain-specific LLM using a 1 billion parameter Llama 3-style model on 8 H100 GPUs. They cover infrastructure setup, memory management, token budget, and optimization techniques like torch.compile to improve training efficiency.