Alibaba Cloud has developed a new pooling system called Aegaeon that significantly reduces the number of Nvidia GPUs required for large language model inference by 82%, allowing 213 GPUs to perform like 1,192. This innovative approach virtualizes GPU access at the token level, enhancing overall output and efficiency during periods of fluctuating demand. The findings, which were published in a peer-reviewed paper, highlight the potential for cloud providers to maximize GPU utilization in constrained markets like China.