学术论文

张朋乐

个人简介

张朋乐,致理书院信息与计算科学专业2021级本科生,曾获清华大学综合优秀奖学金,致理书院学业优秀奖学金,研究兴趣为:高效机器学习和自然语言处理。


文献著作信息

Xiao, C., Luo, Y., Zhang, W., Zhang, P., Han, X.*, Lin, Y.*, Zhang, Z., Xie, R., Liu, Z., Sun, M., & Zhou, J. (2023, December). Variator: Accelerating Pre-trained Models with Plug-and-Play Compression Modules. Findings of the Association for Computational Linguistics: EMNLP 2023, 9947-9959. doi:10.18653/v1/2023.findings-emnlp.666


论文摘要

Large language models (LLMs) have achieved remarkable results on NLP tasks but at the expense of huge parameter sizes and the consequent computational costs. In this paper, we propose Variator, a parameter-efficient acceleration method that enhances computational efficiency through plug-and-play compression plugins. Compression plugins are designed to reduce the sequence length via compressing multiple hidden vectors into one and trained with original LLMs frozen. Different from traditional model acceleration methods, which compress LLMs to smaller sizes, Variator offers two distinct advantages: (1) In real-world applications, the plug-and-play nature of our compression plugins enables dynamic selection of different compression plugins with varying acceleration ratios based on the current workload. (2) The compression plugin comprises a few compact neural network layers with minimal parameters, significantly saving storage and memory overhead, particularly in scenarios with a growing number of tasks. We validate the effectiveness of Variator on seven datasets. Experimental results show that Variator can save 53% computational costs using only 0.9% additional parameters with a performance drop of less than 2%. Moreover, when the model scales to billions of parameters, Variator matches the strong performance of uncompressed LLMs. Our code and checkpoints will be released to facilitate future work.