Towards an On-Device Agent for Text Rewriting
Abstract
Large Language Models (LLMs) have demonstrated impressive capabilities for text rewriting. However creating a smaller yet potent language model for text rewriting presents two formidable challenges: costly data collection and absence of emergent capabilities. In this paper we present solutions to address the above challenges. We propose an new instruction tuning method to develop a mobile text rewriting model that leverages LLM-generated data and heuristic reinforcement learning, eliminating the need for human data collection. Moreover, to bridge the performance gap from the constraint size, we propose a cascading approach based on the confidence levels which are distilled from the large server model’s critiques. To evaluate the text rewriting tasks for mobile scenarios, we introduce MessageRewriteEval, a human-labeled benchmark that focuses on text rewriting of messages through natural language instructions. Through empirical experiments, we demonstrate that our on-device model surpasses the current state-of-the-art LLMs in text rewriting while maintaining a significantly reduced model size using public benchmark EditEval and our new benchmark. We also demonstrate that our proposed cascading approach improves model performance further.
BibTex
@inproceedings{Zhu2024OnDeviceRewriting,
author = {Zhu, Yun and Liu, Yinxiao and Stahlberg, Felix and Kumar, Shankar and Chen, Yu-hui and Luo, Liangchen and Shu, Lei and Liu, Renjie and Chen, Jindong and Meng, Lei},
title = {Towards an On-Device Agent for Text Rewriting},
booktitle = {Findings of the Association for Computational Linguistics: NAACL 2024},
month = {June},
year = {2024},
address = {Mexico City, Mexico},
pages = {2535--2552},
publisher = {Association for Computational Linguistics}
}