检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:Jaydeep Kishore Snehasis Mukherjee
机构地区:[1]Department of Computer Science and Engineering,Shiv Nadar University,Delhi NCR 201314,India
出 处:《Machine Intelligence Research》2024年第6期1201-1213,共13页机器智能研究(英文版)
摘 要:With the recent advancements of deep learning-based methods in image classification,the requirement of a huge amount of training data is inevitable to avoid overfitting problems.Moreover,supervised deep learning models require labelled datasets for train-ing.Preparing such a huge amount of labelled data requires considerable human effort and time.In this scenario,self-supervised models are becoming popular because of their ability to learn even from unlabelled datasets.However,the efficient transfer of knowledge learned by self-supervised models into a target task,is an unsolved problem.This paper proposes a method for the efficient transfer of know-ledge learned by a self-supervised model,into a target task.Hyperparameters such as the number of layers,the number of units in each layer,learning rate,and dropout are automatically tuned in these fully connected(FC)layers using a Bayesian optimization technique called the tree-structured parzen estimator(TPE)approach algorithm.To evaluate the performance of the proposed method,state-of-the-art self-supervised models such as SimClr and SWAV are used to extract the learned features.Experiments are carried out on the CIFAR-10,CIFAR-100,and Tiny ImageNet datasets.The proposed method outperforms the baseline approach with margins of 2.97%,2.45%,and 0.91%for the CIFAR-100,Tiny ImageNet,and CIFAR-10 datasets,respectively.
关 键 词:Neural architecture search AUTOTUNING self-supervised learning hyperparameter optimization tree-structured parzen estimator(TPE)
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:216.73.216.13