Click any tag below to further narrow down your results
Links
The SGLang RL team developed an end-to-end INT4 Quantization-Aware Training (QAT) pipeline that enhances training efficiency and model stability. By using fake quantization during training and real quantization at inference, they achieved significant performance improvements for large models on a single GPU. The article details the technical steps taken and results of their approach.
DeepSeek introduced a paper detailing its innovative training method called Manifold-Constrained Hyper-Connections. This approach aims to enhance scalability and reduce energy use in AI development, addressing challenges tied to limited access to Nvidia chips in China.
Flow-GRPO-Fast is a newly introduced accelerated variant of the Flow-GRPO model that enhances training efficiency by reducing the number of denoising steps required per trajectory. Recent updates include support for various models and reward mechanisms, as well as improvements in training parameters to optimize performance on tasks such as image editing and generation. The article outlines detailed instructions for setup, training, and model implementation across multiple environments.
TRL has introduced co-located vLLM to improve the efficiency of training large language models by allowing both training and inference to run on the same GPUs, eliminating idle time and reducing hardware costs. This integration enhances throughput, simplifies deployment, and makes the system more robust for online learning setups like GRPO. The new approach is supported by a series of performance experiments demonstrating significant speedups compared to traditional server setups.