[Objective]The increasing maturity of large language model technology has facilitated its widespread application in downstream tasks across various vertical fields.Large language models have exhibited beneficial performance in text summarization tasks in general fields,such as news and art.However,the highly specific language style in the judicial field and the unique complexity of judicial documents in terms of structure and logic make it difficult for large language models to generate judicial document summaries.This study aims to combine prompt learning with large language models to explore their performance in summarizing judicial documents.Prompt templates containing structural information and judicial documents are used as inputs for fine-tuning large language models.As a result,large language models can generate judicial document summaries that adhere to judicial language styles and the structural and logical complexities of judicial documents.[Methods]This study proposes a judicial document summary method that combines prompt learning and the Qwen large language model.Judicial document data are used as the input for fine-tuning a large language model using supervised fine-tuning technology to enhance its applicability in the judicial field.Simultaneously,prompt templates that incorporate structural information and role instructions are designed to optimize summary generation to more accurately reflect the structural characteristics and logical relationships of documents.According to the characteristics of the pretraining data format of the large language model,the fine-tuning data were constructed in the form of question-answer pairs.[Results]The experimental results show that the proposed method improves the F1 of the baseline model by 21.44%,28.50%,and 28.97%in ROUGE-1,ROUGE-2,and ROUGE-L,respectively,and exceeds all of the comparison models.The ablation experiment demonstrated that the summary generation method using prompt learning was superior to the method without prompt learning for all indicators,and the performance of summarization generated by the large language model utilizing prompt learning was significantly enhanced.The case demonstration reveals that after prompt learning is used to enhance the perception of structural information in the judgment document by the large language model,the judgment document summary generated by this model can better capture and retain key information in the judgment document.Moreover,the language style of this model is closer to that of a real judgment document summary,which further illustrates the effectiveness of the proposed method.[Conclusions]This study integrates the structural information of a judgment document into the task of generating a judgment document summary using a large language model in the form of prompt templates.Prompt templates containing structural information are used to assist the large language model in summarization generation.Therefore,the model can focus on the key information in the judgment document and capture deeper semantic logical relationships.The results demonstrate that after fine-tuning the large language model with judicial document data and introducing structural information,the model demonstrated excellent performance and great application potential in the judicial document summary task.The proposed method can effectively enhance the capability of a large language model in the field of judicial document summaries.
关键词
裁判文书摘要/文本摘要/大语言模型/提示学习
Key words
referee's decision summary/summarization/large language model/prompt learning