用于语义分割的自监督对比式表征学习  被引量:2

Self-supervised contrastive representation learning for semantic segmentation

在线阅读下载全文

作  者:刘博翀 蔡怀宇[1] 汪毅[1] 陈晓冬[1] LIU Bochong;CAI Huaiyu;WANG Yi;CHEN Xiaodong(Ministry of Education Key Laboratory of Optoelectronic Information Technology,School of Precision Instrument and Optoelectronic Engineering,Tianjin University,Tianjin 300072,China)

机构地区:[1]天津大学精密仪器与光电子工程学院光电信息技术教育部重点实验室,天津300072

出  处:《西安电子科技大学学报》2024年第1期125-134,共10页Journal of Xidian University

基  金:天津市科技计划项目(17ZXGGX00140)。

摘  要:为了提升语义分割模型的精度,并减少逐像素标注大规模语义分割数据集的人力和时间成本,研究了自监督对比式表征学习的预训练方法,并结合语义分割任务的特点,设计了全局-局部交叉对比学习(GLCCL)方法。该方法将全局图像和局部分块后的一系列图像块输入到网络中分别编码全局和局部视觉表征,并通过构建包含全局对比、局部对比和全局-局部交叉对比的损失函数来指导模型训练,使得模型能够同时学习全局和局部区域的视觉表征以及跨区域语义相关性。使用该方法预训练BiSeNet再迁移到语义分割任务时,对比现有的自监督对比式表征学习和有监督预训练方法分别具有0.24%和0.9%平均交并比(MIoU)的性能提升。实验结果表明,该方法能够采用无标注的数据训练语义分割模型而实现分割效果的提升,具有一定的实用价值。To improve the accuracy of the semantic segmentation models and avoid the labor and time costs of pixel-wise image annotation for large-scale semantic segmentation datasets,this paper studies the pre-training methods of self-supervised contrastive representation learning,and designs the Global-Local Cross Contrastive Learning(GLCCL)method based on the characteristics of the semantic segmentation task.This method feeds global images and a series of image patches after local chunking into the network to extract global and local visual representations respectively,and guides the network training by constructing loss function that includes global contrast,local contrast,and global-local cross contrast,enabling the network to learn both global and local visual representations as well as cross-regional semantic correlations.When using this method to pre-train BiSeNet and transfer to the semantic segmentation task,compared with the existing self-supervised contrastive representational learning and supervised pre-training methods,the performance improvement of 0.24%and 0.9%mean intersection over union(MIoU)is achieved.Experimental results show that this method can improve the segmentation results by pre-training the semantic segmentation model with unlabeled data,which has a certain practical value.

关 键 词:语义分割 自监督表征学习 对比学习 深度学习 

分 类 号:TP391.4[自动化与计算机技术—计算机应用技术]

 

参考文献:

正在载入数据...

 

二级参考文献:

正在载入数据...

 

耦合文献:

正在载入数据...

 

引证文献:

正在载入数据...

 

二级引证文献:

正在载入数据...

 

同被引文献:

正在载入数据...

 

相关期刊文献:

正在载入数据...

相关的主题
相关的作者对象
相关的机构对象