Search Results for author: Wenhua Cheng

Found 3 papers, 3 papers with code

TEQ: Trainable Equivalent Transformation for Quantization of LLMs

1 code implementation17 Oct 2023 Wenhua Cheng, Yiyang Cai, Kaokao Lv, Haihao Shen

As large language models (LLMs) become more prevalent, there is a growing need for new and improved quantization methods that can meet the computationalast layer demands of these modern architectures while maintaining the accuracy.

Quantization

Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs

2 code implementations11 Sep 2023 Wenhua Cheng, Weiwei Zhang, Haihao Shen, Yiyang Cai, Xin He, Kaokao Lv, Yi Liu

Large Language Models (LLMs) have demonstrated exceptional proficiency in language-related tasks, but their deployment poses significant challenges due to substantial memory and storage requirements.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.