Click any tag below to further narrow down your results
Links
This article discusses new methods for enhancing the efficiency of large language models through sparsity. It examines various strategies like relufication and error budget thresholding to achieve significant speedups in on-device inference while maintaining accuracy. The authors are developing a unified framework in PyTorch to streamline these techniques.