检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:李凯歌 蔡鹏飞 周忠[1] Li Kaige;Cai Pengfei;Zhou Zhong(State Key Laboratory of Virtual Reality Technology and Systems,Beihang University,Beijing 100191)
机构地区:[1]北京航空航天大学虚拟现实技术与系统国家重点实验室,北京100191
出 处:《计算机辅助设计与图形学学报》2023年第6期903-914,共12页Journal of Computer-Aided Design & Computer Graphics
基 金:国家自然科学基金(61872024);国家重点研发计划(2018YFB2100603).
摘 要:针对现有行为识别方法缺乏对时空特征关系建模的问题,提出一种基于特征交互和聚类的行为识别方法.首先设计一种混合多尺度特征提取网络提取连续帧的时间和空间特征;然后基于Non-local操作设计一种特征交互模块实现时空特征的交互;最后基于三元组损失函数设计一种难样本选择策略来训练识别网络,实现时空特征的聚类,提高特征的鲁棒性和判别性.实验结果表明,与基线方法TSN相比,所提方法的准确度在UCF101数据集上提高了23.25个百分点,达到94.82%;在HMDB51数据集上提高了20.27个百分点,达到44.03%.To mitigate the problem that the action recognition methods lack the modeling of spatiotemporal feature relationship,an action recognition method based on feature interaction and clustering is proposed.Firstly,a mixed multi-scale feature extraction network is designed to extract spatial and temporal features of continuous frames.Secondly,a feature interaction module is designed based on non-local operation to realize spatiotemporal feature interaction.Finally,based on the triplet loss function,a hard sample selection strategy is designed to train the recognition network,thus realizing spatiotemporal feature clustering and improving the robustness and discrimination of the features.Experimental results show that compared with TSN,the accuracy of on the UCF101 dataset is increased by 23.25 percentage points to 94.82%.On the HMDB51 dataset,the accuracy is increased by 20.27 percentage points to 44.03%.
分 类 号:TP391.41[自动化与计算机技术—计算机应用技术]
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:216.73.216.80