基于知识增强和潜主题学习的小语言模型低资源对话生成
DOI:
CSTR:
作者:
作者单位:

1.河南财经政法大学 计算机与信息工程学院;2.河南财经政法大学 数据科学与电子商务学院;3.河南财经政法大学 旅游管理与会展学院

作者简介:

通讯作者:

中图分类号:

TP391??????

基金项目:

国家自然科学基金项目(62072156);河南省自然科学杰出青年基金项目(252300421061);河南省科技攻关项目(242102210076)。


Knowledge Augmentation and Latent Topics Learning with Small Language Models for Low-Resource Dialogue Generation
Author:
Affiliation:

1.College of Computer and Information Engineering, Henan University of Economics and Law;2.School of Data Science and E-commerce, Henan University of Economics and Law;3.School of Tourism Management and MICE, Henan University of Economics and Law

Fund Project:

Supported by the National Natural Science Foundation of China (62072156), the Natural Science Outstanding Youth Science Foundation Project of Henan Province (252300421061), and the Key Technologies Research and Development Program of Henan Province (242102210076).

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    基于小语言模型的知识对话方法具有重要研究价值,但当前该研究还面临诸多挑战,如缺乏大语言模型的泛化和推理能力,模型对齐困难以及常常面临低资源应用场景等。针对这些问题,提出了一个新的三阶段基于知识增强和潜主题学习的小语言模型低资源对话生成模型。首先在知识对话数据集上对模型进行预训练,然后在低资源对话数据集上进行监督微调。采用基于噪声注入的对话文本降噪预训练方法,减轻模型领域适应问题。引入检索增强的知识生成模块,用于生成对话上下文相关的知识,以解决低资源对话数据集上知识缺失问题。为解决话题漂移和意图偏离问题,设计了一个基于条件变分自编码的潜主题学习方法,用于指导对话的生成。最后在低资源对话数据集上采用自批评序列训练方法对模型进行强化学习,以进一步提升模型的性能。实验结果表明,在衡量对话回复生成性能(如多样性、连贯性和准确率等)的多个指标上,提出的方法明显优于基线方法。

    Abstract:

    Knowledge-grounded dialogue methods based on small language models (SMLs) hold significant research value, yet face multiple challenges including limited generalization and reasoning capabilities compared to LLMs, model alignment difficulties, and frequent low-resource application scenarios. To address these challenges, this paper presents a novel three-stage knowledge-enhanced and latent-topic learning framework for low-resource dialogue generation with SMLs. The SMLs-based dialogue model initially is pre-trained on knowledge-grounded dialogue datasets, and then a supervised fine-tuning for the model is performed on low-resource dialogue datasets. Model domain adaptation issues are alleviated through application of a noise-injection-based denoising pre-training approach to dialog texts. A retrieval-augmented knowledge generation module is introduced to produce context-relevant knowledge, effectively mitigating knowledge scarcity in low-resource settings. To solve topic drift and intent deviation, a method via conditional variational autoencoder-based latent topic learning is proposed to guide response generation. To improve model performance, self-critical sequence training for reinforcement learning is employed for the dialogue model on low-resource datasets. Experimental results demonstrate that the proposed method significantly outperforms baseline approaches across multiple metrics evaluating response generation quality, including diversity, coherence, and accuracy.

    参考文献
    相似文献
    引证文献
引用本文
分享
相关视频

文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2025-12-08
  • 最后修改日期:2026-01-07
  • 录用日期:2026-04-08
  • 在线发布日期:
  • 出版日期:
文章二维码