检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:Keshi GE Yiming ZHANG Yongquan FU Zhiquan LAI Xiaoge DENG Dongsheng LI
机构地区:[1]College of Computer,National University of Defense Technology,Changsha,410073,China
出 处:《Science China(Information Sciences)》2023年第6期134-150,共17页中国科学(信息科学)(英文版)
基 金:supported in part by National Natural Science Foundation of China (Grant Nos.62025208,61972409);National Key Research Development Program of China (Grant No.2021YFB0301200)。
摘 要:Gradient quantization has been widely used in distributed training of deep neural network(DNN)models to reduce communication cost.However,existing quantization methods overlook that gradients have a nonuniform distribution changing over time,which can lead to significant compression error that not only increases the number of training iterations but also requires a higher number of quantization bits(and consequently higher delay for each iteration)to keep the validation accuracy as high as the original stochastic gradient descent(SGD)approach.To address this problem,in this paper we propose cluster-aware sketch quantization(CASQ),a novel sketch-based gradient quantization method for SGD with convergence guarantees.CASQ models the nonuniform distribution of gradients via clustering,and adaptively allocates appropriate numbers of hash buckets based on the statistics of different clusters to compress gradients.Extensive evaluation shows that compared to existing quantization methods,CASQ-based SGD(i)achieves the same validation accuracy when decreasing quantization level from 3 bits to 2 bits,and(ii)reduces the training time to convergence by up to 43%for the same training loss.
关 键 词:distributed training deep learning COMMUNICATION SKETCH QUANTIZATION
分 类 号:TP18[自动化与计算机技术—控制理论与控制工程]
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:216.73.216.7