首页|基于深度学习的自然语言处理鲁棒性研究综述

基于深度学习的自然语言处理鲁棒性研究综述

扫码查看
近年来,基于深度神经网络的模型在几乎所有自然语言处理任务上都取得了非常好的效果,在很多任务上甚至超越了人类.展现了极强能力的大规模语言模型也为自然语言处理模型的发展与落地提供了新的机遇和方向.然而,这些在基准测试集合上取得很好结果的模型在实际应用中的效果却经常大打折扣.近期的一些研究还发现,在测试数据上替换一个相似词语、增加一个标点符号,甚至只是修改一个字母都可能使得这些模型的预测结果发生改变,效果大幅度下降.即使是大型语言模型,也会因输入中的微小扰动而改变其预测结果.什么原因导致了这种现象的发生?深度神经网络模型真的如此脆弱吗?如何才能避免这种问题的出现?这些问题近年来受到了越来越多的关注,诸多有影响力的工作都不约而同地从不同方面讨论了自然语言处理的鲁棒性问题.在本文中,我们从自然语言处理任务的典型范式出发,从数据构建、模型表示、对抗攻防以及评估评价等四个方面对自然语言处理鲁棒性相关研究进行了总结和归纳,并对最新进展进行了介绍,最后探讨了未来的可能研究方向以及我们对自然语言处理鲁棒性问题的一些思考.
Recent Researches of Robustness in Natural Language Processing Based on Deep Neural Network
In recent years,deep neural network models have exhibited exceptional performance across a wide range of natural language processing tasks,often surpassing human performance in certain domains.The emergence of powerful large-scale language models has opened up new avenues and possibilities for the advancement and application of natural language processing models.However,the efficacy of these models,which demonstrate impressive results on standardized benchmarks,is substantially diminished when deployed in real-world scenarios.Recent investigations have also revealed that the predictions made by these models can be significantly altered through simple modifications,leading to a drastic decline in performance.Even large-scale language models are susceptible to modifying their predictions in response to minor perturbations introduced in the input data.These observations are closely tied to the concept of model robustness,which generally pertains to the ability of a model to maintain consistent output when confronted with new,independent,yet similar data.A highly robust deep learning model exhibits consistent output despite encountering minor alterations that should not significantly impact the resulting prediction.The study of model robustness in deep learning has emerged as a prominent and extensively explored area of interest in the field of natural language processing.A plethora of research endeavors has been dedicated to exploring the concept of robustness in natural language processing(NLP),although many of these studies have focused on specific tasks,failing to consider the broader context.To provide a comprehensive overview of robustness research,this review encompasses the latest advancements in deep NLP from four key perspectives:data construction,model representation,adversarial attacks and defenses,and evaluations.Existing techniques aimed at enhancing or diminishing the robustness of NLP models are also summarized.The foundation of machine learning,data construction,is initially explored,encompassing considerations such as dataset biases and dataset poisoning.Notably,dataset poisoning commonly involves backdoor attacks,wherein triggers are injected into constructed datasets.Consequently,models trained on these poisoned datasets are capable of accurate predictions on clean data,yet exhibit erroneous outputs when encountering data containing specific markers(i.e.,triggers).Subsequently,feature learning,which transforms input data into vectors,is examined with the objective of representing textual content in a task-agnostic and domain-independent manner.Various deep NLP models are introduced,alongside diverse methods for improving model robustness within the domain of model representation,including robustness encoding,knowledge incorporation,task characteristics,and causal inference.Adversarial attack and defense algorithms are presented as means to deceive models and enhance their robustness,respectively.Mainstream adversarial attack methods,such as white-box,black-box,and blind attacks,are discussed.Correspondingly,a multitude of research studies address the challenges of adversarial defense and robustness improvement,which are also included in this paper.Traditional metrics,owing to the issue of robustness,are deemed insufficient for fair and comprehensive evaluations.Consequently,a body of work proposes alternative evaluation metrics to assess the effectiveness of models,and prominent evaluation approaches for both general-purpose and specific NLP tasks are introduced.Finally,potential future research directions and considerations concerning the robustness of natural language processing are deliberated upon,including more rational data construction,more interpretable and robust model representations,imperceptible textual adversarial attacks,efficient adversarial defense techniques,evaluation methods focusing on linguistic knowledge,balancing model robustness and accuracy,and unifying robustness of different domains.

natural language processingrobustnessdeep learningpretrained language modelsadversarial attacks and defenses

桂韬、奚志恒、郑锐、刘勤、马若恬、伍婷、包容、张奇

展开 >

复旦大学现代语言学研究院 上海 200433

复旦大学计算机科学技术学院 上海 200433

自然语言处理 鲁棒性 深度学习 预训练语言模型 对抗攻防

国家自然科学基金国家自然科学基金国家自然科学基金

622060576207606961976056

2024

计算机学报
中国计算机学会 中国科学院计算技术研究所

计算机学报

CSTPCD北大核心
影响因子:3.18
ISSN:0254-4164
年,卷(期):2024.47(1)
  • 3
  • 2