检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:Woongkyu Park Yeongyu Choi Mahammad Shareef Mekala Gyu Sang Choi Kook-Yeol Yoo Ho-youl Jung
机构地区:[1]Department of Information and Communication Engineering,Yeungnam University,Gyeongsan,38541,Republic of Korea [2]RLRC for Autonomous Vehicle Parts and Materials Innovation,Yeungnam University,Gyeongsan,38541,Republic of Korea [3]School of Computing,Robert Gordon University,Aberdeen,AB107QB,UK
出 处:《Computers, Materials & Continua》2025年第3期3965-3981,共17页计算机、材料和连续体(英文)
基 金:supported by the Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education under Grant NRF-2021R1A6A1A03039493.
摘 要:Designing fast and accurate neural networks is becoming essential in various vision tasks.Recently,the use of attention mechanisms has increased,aimed at enhancing the vision task performance by selectively focusing on relevant parts of the input.In this paper,we concentrate on squeeze-and-excitation(SE)-based channel attention,considering the trade-off between latency and accuracy.We propose a variation of the SE module,called squeeze-and-excitation with layer normalization(SELN),in which layer normalization(LN)replaces the sigmoid activation function.This approach reduces the vanishing gradient problem while enhancing feature diversity and discriminability of channel attention.In addition,we propose a latency-efficient model named SELNeXt,where the LN typically used in the ConvNext block is replaced by SELN to minimize additional latency-impacting operations.Through classification simulations on ImageNet-1k,we show that the top-1 accuracy of the proposed SELNeXt outperforms other ConvNeXt-based models in terms of latency efficiency.SELNeXt also achieves better object detection and instance segmentation performance on COCO than Swin Transformer and ConvNeXt for small-sized models.Our results indicate that LN could be a considerable candidate for replacing the activation function in attention mechanisms.In addition,SELNeXt achieves a better accuracy-latency trade-off,making it favorable for real-time applications and edge computing.The code is available at https://github.com/oto-q/SELNeXt(accessed on 06 December 2024).
关 键 词:Attention mechanism convolutional neural networks image classification object detection semantic segmentation
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:3.131.95.159