首页|面向情感分类的非监督异构对比蒸馏模型

面向情感分类的非监督异构对比蒸馏模型

扫码查看
网络评论的数量呈爆炸式增长,对这些评论进行情感分析有着重要的研究价值。自从BERT模型被提出后,预训练模型成为情感分析任务上的常用方法,但存在模型参数量过大、推理速度缓慢的缺点。在此之前情感分析的方法是一些简单的神经网络模型,训练速度快,可部署性强,但效果一般。因此,结合两类方法的优点,该文提出一种异构设置的非监督对比蒸馏模型,用于网络评论情感分析。在相同数据集和计算资源的情况下,该模型较BERT模型参数量减少146倍,推理时间减少270倍;较DistilBERT蒸馏参数量减少88倍,推理时间减少42。3倍,效果提升1。8个百分点(68。3%vs 70。1%)。
UNSUPERVISED CONTRASTIVE DISTILLATION MODEL WITH HETEROGENEOUS SETTINGS FOR SENTIMENT CLASSIFICATION
With the rapid development of Internet technology,the number of online reviews is increasing.So,sentiment analysis on these reviews is of important research value.Since the BERT model was proposed,pre-trained language models became a common method for sentiment analysis tasks.However,these models had a large number of parameters and took long inference time.The common sentiment analysis methods were some simple neural network models.These models had fast training speed and strong deployability,but mediocre effect.Therefore,combining the advantages of two types of methods,this paper proposes an unsupervised comparative distillation model with heterogeneous settings.With same dataset settings and computing resources,our model reduces the number of parameters of the BERT model by 146 times,and the reasoning time is reduced by 207 times.Compared with DistilBERT,the amount of distillation parameters is reduced by 88 times,the inferencing time is reduced by 42.3 times,and the performance is increased by 1.8%(68.3%vs 70.1%).

Natural language processingSentiment classificationTransfer learningKnowledge distillation

苏肇辰、徐乐怡、王晶晶、李俊涛

展开 >

苏州大学计算机科学与技术学院 江苏苏州 215006

自然语言处理 情感分类 迁移学习 知识蒸馏

国家自然科学基金项目

62006166

2024

计算机应用与软件
上海市计算技术研究所 上海计算机软件技术开发中心

计算机应用与软件

CSTPCD北大核心
影响因子:0.615
ISSN:1000-386X
年,卷(期):2024.41(10)