数学科学研究所
Insitute of Mathematical Science

Colloquium: Sparse Deep Neural Networks Through L_{1,\infty}-Weight Normalization

Colloquium| Institute of Mathematical Sciences

Time:16:00-17:00, June 24, Mondday

LocationRoom S407, IMS


Speaker: Zhouwang Yang, University of Science and Technology of China  

Abstract: Deep neural networks have recently demonstrated an amazing performance on many challenging tasks. Overtting is one of the notorious features for DNNs. Empirical evidence suggests that inducing sparsity can relieve overtting, and weight normalization can accelerate the algorithm convergence. In this talk, we report our work on L_{1,\infty}-weight normalization for deep neural networks with bias neurons to achieve the sparse architecture. We theoretically establish the generalization error bounds for both regression and classication under the L_{1,\infty}-weight normalization. It is shown that the upper bounds are independent of the network width and sqrt(k)-dependence on the network depth k, which are the best available bounds for networks with bias neurons. These results provide theoretical justications on the usage of such weight normalization to reduce the generalization error. We also develop an easily implemented gradient projection descent algorithm to practically obtain a sparse neural network. We perform various experiments to validate our theory and demonstrate the eectiveness of the resulting approach.

  


  


地址:上海市浦东新区华夏中路393号
邮编:201210
上海市徐汇区岳阳路319号8号楼
200031(岳阳路校区)