Memory-Tuning: A Unified Parameter-Efficient Tuning Method for Pre-Trained Language Models
Published in IEEE Transactions on Audio, Speech and Language Processing, 2024
We propose memory-tuning, a novel parameter-efficient method that unifies task-specific knowledge learning for both multi-head attention and feed-forward networks in Transformers, theoretically linking it to prefix tuning while outperforming full fine-tuning on eight benchmarks across sentence- and token-level tasks.
Recommended citation: Qi, W., Liu, R., Zuo, Y., Li, F., Chen, Y., & Wu, J. (2024). Memory-Tuning: A Unified Parameter-Efficient Tuning Method for Pre-trained Language Models. IEEE/ACM Transactions on Audio, Speech, and Language Processing. https://ieeexplore.ieee.org/abstract/document/10769026