research-papers
BiLLM: Pushing the Limit of Post-Training Quantization for LLMs
Original Paper: https://arxiv.org/abs/2402.04291 By: Wei Huang, Yangdong Liu, Haotong Qin, Ying Li, Shiming Zhang, Xianglong Liu, Michele Magno, Xiaojuan Qi Abstract: Pretrained large language models (LLMs) exhibit exceptional general language processing capabilities but come with significant demands on memory and computational resources. As a powerful