Towards an On-Device Agent for Text Rewriting

Yun Zhu, Yinxiao Liu, Felix Stahlberg, Shankar Kumar, Yu-hui Chen, Liangchen Luo, Lei Shu, Renjie Liu, Jindong Chen, Lei Meng.
In Findings of the Association for Computational Linguistics: NAACL 2024, Mexico City, Mexico. 2024.

Abstract

Large Language Models (LLMs) have demonstrated impressive capabilities for text rewriting. However creating a smaller yet potent language model for text rewriting presents two formidable challenges: costly data collection and absence of emergent capabilities. In this paper we present solutions to address the above challenges. We propose an new instruction tuning method to develop a mobile text rewriting model that leverages LLM-generated data and heuristic reinforcement learning, eliminating the need for human data collection. Moreover, to bridge the performance gap from the constraint size, we propose a cascading approach based on the confidence levels which are distilled from the large server model’s critiques. To evaluate the text rewriting tasks for mobile scenarios, we introduce MessageRewriteEval, a human-labeled benchmark that focuses on text rewriting of messages through natural language instructions. Through empirical experiments, we demonstrate that our on-device model surpasses the current state-of-the-art LLMs in text rewriting while maintaining a significantly reduced model size using public benchmark EditEval and our new benchmark. We also demonstrate that our proposed cascading approach improves model performance further.

BibTex

@inproceedings{Zhu2024OnDeviceRewriting,
  author = {Zhu, Yun and Liu, Yinxiao and Stahlberg, Felix and Kumar, Shankar and Chen, Yu-hui and Luo, Liangchen and Shu, Lei and Liu, Renjie and Chen, Jindong and Meng, Lei},
  title = {Towards an On-Device Agent for Text Rewriting},
  booktitle = {Findings of the Association for Computational Linguistics: NAACL 2024},
  month = {June},
  year = {2024},
  address = {Mexico City, Mexico},
  pages = {2535--2552},
  publisher = {Association for Computational Linguistics}
}