重庆邮电大学学报(自然科学版)2024,Vol.36Issue(3) :524-532.DOI:10.3979/j.issn.1673-825X.202305250159

基于多尺度时序交互的第一人称行为识别方法

Egocentric action recognition method based on multi-scale temporal interaction

罗祥奎 高陈强 陈欣悦 王升伟
重庆邮电大学学报(自然科学版)2024,Vol.36Issue(3) :524-532.DOI:10.3979/j.issn.1673-825X.202305250159

基于多尺度时序交互的第一人称行为识别方法

Egocentric action recognition method based on multi-scale temporal interaction

罗祥奎 1高陈强 1陈欣悦 1王升伟1
扫码查看

作者信息

  • 1. 重庆邮电大学 通信与信息工程学院,重庆 400065;信号与信息处理重庆市重点实验室,重庆 400065
  • 折叠

摘要

对于第一人称行为识别任务,现有方法大多使用了目标边界框和人眼视线数据等非行为类别标签对深度神经网络进行辅助监督,以使其关注视频中手部及其交互物体所在区域.这既需要更多的人工标注数据,又使得视频特征的提取过程变得更为复杂.针对该问题,提出了一种多尺度时序交互模块,通过不同尺度的3D时序卷积使2D神经网络提取的视频帧特征进行时序交互,从而使得单一视频帧的特征融合其近邻帧的特征.在只需行为类别标签作监督的情况下,多尺度时序交互能够促使网络更加关注第一人称视频中手部及其交互物体所在区域.实验结果表明,提出的方法在识别准确率优于现有第一人称行为识别方法.

Abstract

For egocentric action recognition task,most existing methods use non-behavior category labels such as target bounding boxes and eye gaze data to assist in supervising deep neural networks,so that they can focus on the areas where hands and their interactive objects are located in the video.This requires more manually labeled data and makes the process of extracting video features more complex.To address this issue,a multi-scale temporal interaction module is proposed,which enables 2D neural networks to extract video frame features through different scales of 3D temporal convolution for temporal interaction,so that the features of a single video frame can be fused with those of its neighboring frames.With only behavior category labels for supervision,multi-scale temporal interaction can promote the network to pay more attention to the areas where hands and their interactive objects are located in egocentric videos.Experimental results show that the pro-posed method has better recognition accuracy than existing egocentric action recognition methods.

关键词

行为识别/第一人称视觉/时序交互/深度学习

Key words

action recognition/egocentric vision/temporal interaction/deep learning

引用本文复制引用

基金项目

重庆市教委科学技术研究计划(KJZD-K202100606)

出版年

2024
重庆邮电大学学报(自然科学版)
重庆邮电大学

重庆邮电大学学报(自然科学版)

CSTPCD北大核心
影响因子:0.66
ISSN:1673-825X
段落导航相关论文