🤖 AI Summary
This study addresses the limitations of existing commonsense reasoning datasets, which suffer from insufficient scale and diversity, thereby hindering generative models from simultaneously improving response diversity and commonsense plausibility. To overcome this, the authors propose a two-stage synthetic data generation framework that constructs CommonSyn—a large-scale, high-quality dataset tailored for diverse commonsense reasoning—without relying on human annotation, substantially broadening coverage of commonsense scenarios. The approach leverages large language models for both data generation and filtering, and is adaptable to fine-tuning models of varying scales. Experimental results demonstrate that models fine-tuned on CommonSyn significantly outperform baseline systems and those trained on human-annotated data in both response diversity and commonsense quality.
📝 Abstract
Conversational agents are required to respond to their users not only with high quality (i.e. commonsense bearing) responses, but also considering multiple plausible alternative scenarios, reflecting the diversity in their responses. Despite the growing need to train diverse commonsense generators, the progress of this line of work has been significantly hindered by the lack of large-scale high-quality diverse commonsense training datasets. Due to the high annotation costs, existing Generative Commonsense Reasoning (GCR) datasets are created using a small number of human annotators, covering only a narrow set of commonsense scenarios. To address this training resource gap, we propose a two-stage method to create the first-ever synthetic dataset CommonSyn for diversified (GCR). The model fine-tuned on our synthetic data jointly increase both generation diversity and quality compared with vanilla models and the model fine-tuned on human-crafted dataset across different size Large Language Models (LLMs)