Communication is All You Need: Persuasion Dataset Construction via Multi-LLM Communication
Published in Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics, 2025
Large Language Models (LLMs) have shown proficiency in generating persuasive dialogue, yet concerns about the fluency and sophistication of their outputs persist. This paper presents a multi-LLM communication framework designed to enhance the generation of persuasive data automatically. This framework facilitates the efficient production of high-quality, diverse linguistic content with minimal human oversight. Through extensive evaluations, we demonstrate that the generated data excels in naturalness, linguistic diversity, and the strategic use of persuasion, even in complex scenarios involving social taboos. The framework also proves adept at generalizing across novel contexts. Our results highlight the framework’s potential to significantly advance research in both computational and social science domains concerning persuasive communication.
Recommended citation: Ma, Weicheng, Hefan Zhang, Ivory Yang, Shiyu Ji, Joice Chen, Farnoosh Hashemi, Shubham Mohole et al. "Communication is All You Need: Persuasion Dataset Construction via Multi-LLM Communication." arXiv preprint arXiv:2502.08896 (2025).
Download Paper