This work addresses the problem of network pruning and proposes a novel joint training method based on a multiobjective optimization model. Most of the state-of-the-art pruning methods rely on user experience for selecting the sparsity ratio of the weight matrices or tensors, and thus suffer from severe performance reduction with inappropriate user-defined parameters. Moreover, networks might be inferior due to the inefficient connecting architecture search, especially when it is highly sparse. It is revealed in this work that the network model might maintain sparse characteristic in the early stage of the backpropagation (BP) training process, and evolutionary computation-based algorithms can accurately discover the connecting architecture with satisfying network performance. In particular, we establish a multiobjective sparse model for network pruning and propose an efficient approach that combines BP training and two modified multiobjective evolutionary algorithms (MOEAs). The BP algorithm converges quickly, and the two MOEAs can search for the optimal sparse structure and refine the weights, respectively. Experiments are also included to prove the benefits of the proposed algorithm. We show that the proposed method can obtain a desired Pareto front (PF), leading to a better pruning result comparing to the state-of-the-art methods, especially when the network structure is highly sparse.
Skip Nav Destination
Article navigation
March 2021
March 26 2021
Joint Structure and Parameter Optimization of Multiobjective Sparse Neural Network
In Special Collection:
CogNet
Junhao Huang,
Junhao Huang
Guangdong Key Laboratory of Intelligent Information Processing, College of Electronics and Information Engineering, Shenzhen University, Shenzhen 518060, China, [email protected]
Search for other works by this author on:
Weize Sun,
Weize Sun
*
Guangdong Key Laboratory of Intelligent Information Processing, College of Electronics and Information Engineering, Shenzhen University, Shenzhen 518060, China, [email protected]
*Corresponding author: [email protected]
Search for other works by this author on:
Lei Huang
Lei Huang
Guangdong Key Laboratory of Intelligent Information Processing, College of Electronics and Information Engineering, Shenzhen University, Shenzhen 518060, China, [email protected]
Search for other works by this author on:
Junhao Huang
Guangdong Key Laboratory of Intelligent Information Processing, College of Electronics and Information Engineering, Shenzhen University, Shenzhen 518060, China, [email protected]
Weize Sun
*
Guangdong Key Laboratory of Intelligent Information Processing, College of Electronics and Information Engineering, Shenzhen University, Shenzhen 518060, China, [email protected]
Lei Huang
Guangdong Key Laboratory of Intelligent Information Processing, College of Electronics and Information Engineering, Shenzhen University, Shenzhen 518060, China, [email protected]
*Corresponding author: [email protected]
Received:
April 16 2020
Accepted:
November 06 2020
Online ISSN: 1530-888X
Print ISSN: 0899-7667
© 2021 Massachusetts Institute of Technology
2021
Massachusetts Institute of Technology
Neural Computation (2021) 33 (4): 1113–1143.
Article history
Received:
April 16 2020
Accepted:
November 06 2020
Citation
Junhao Huang, Weize Sun, Lei Huang; Joint Structure and Parameter Optimization of Multiobjective Sparse Neural Network. Neural Comput 2021; 33 (4): 1113–1143. doi: https://doi.org/10.1162/neco_a_01368
Download citation file:
Sign in
Don't already have an account? Register
Client Account
You could not be signed in. Please check your email address / username and password and try again.
Could not validate captcha. Please try again.
Sign in via your Institution
Sign in via your InstitutionEmail alerts
Advertisement
Cited By
Related Articles
VSD-MOEA: A Dominance-Based Multiobjective Evolutionary Algorithm with Explicit Variable Space Diversity Management
Evol Comput (June,2022)
Difficulty Adjustable and Scalable Constrained Multiobjective Test Problem Toolkit
Evol Comput (September,2020)
Multiobjective Evolutionary Algorithms: Analyzing the State-of-the-Art
Evol Comput (June,2000)
Combining Convergence and Diversity in Evolutionary Multiobjective Optimization
Evol Comput (September,2002)
Related Book Chapters
Joint Attention
The Shared World: Perceptual Common Knowledge, Demonstrative Communication, and Social Space
The Joint Target-Motion Estimation Problem
Measurements-Based Radar Signature Modeling: An Analysis Framework
What Is Joint Attention? Comparative, Developmental, and Definitional Issues
Joint Attention: New Developments in Psychology, Philosophy of Mind, and Social Neuroscience
How Does Joint Attention Work? Mechanisms and Processes
Joint Attention: New Developments in Psychology, Philosophy of Mind, and Social Neuroscience