首页|Towards Improving Interpretability of Language Model Generation Through a Structured Knowledge Discovery Approach

Towards Improving Interpretability of Language Model Generation Through a Structured Knowledge Discovery Approach

扫码查看
Knowledge-enhanced text generation aims to enhance the quality of generated text by utilizing internal or external knowledge sources. While language models have demonstrated impressive capabilities in generating coherent and fluent text, the lack of interpretability presents a substantial obstacle. The limited interpretability of generated text significantly impacts its practical usability, particularly in knowledge-enhanced text generation tasks that necessitate reliability and explainability. Existing methods often employ domain-specific knowledge retrievers that are tailored to specific data characteristics, limiting their generalizability to diverse data types and tasks. To overcome this limitation, we directly leverage the two-tier architecture of structured knowledge, consisting of high-level entities and lowlevel knowledge triples, to design our task-agnostic structured knowledge hunter. Specifically, we employ a local-global interaction scheme for structured knowledge representation learning and a hierarchical transformer-based pointer network as the backbone for selecting relevant knowledge triples and entities. By combining the strong generative ability of language models with the high faithfulness of the knowledge hunter, our model achieves high interpretability, enabling users to comprehend the model's output generation process. Furthermore, we empirically demonstrate the effectiveness of our model in both internal knowledge-enhanced table-to-text generation on the RotoWireFG dataset and external knowledge-enhanced dialogue response generation on the KdConv dataset. Our task-agnostic model outperforms state-of-the-art methods and corresponding language models, setting new standards on the benchmark.

Task analysisKnowledge graphsBenchmark testingAdaptation modelsStandardsPredictive modelsPlanning

Shuqi Liu、Han Wu、Guanzhi Deng、Jianshu Chen、Xiaoyang Wang、Linqi Song

展开 >

City University of Hong Kong, Hong Kong|City University of Hong Kong, Shenzhen Research Institute, Shenzhen, China

Tencent AI Lab, Seattle, US

2025

IEEE journal of selected topics in signal processing
  • 59