Technical Name Sparse Neural Network Learning
Project Operator National Taiwan University
Project Host 林守德
Summary
A trained deep learning model usually contains millionseven trillions of parameters, which can prohibit it from being executed on lightweight end-devices such as IoT sensorsmobile devices due to the constraint of memorystorage. To addresses the issue, we develop a novel sparse learning technique that not only reduces the computational cost, but also retains the original model performan
Scientific Breakthrough
To tackle the issue of high computational costs, we conduct L1-regularization technique to sparsify deep neural networks. With the designated L1-regularized objective function, we develop a novel algorithm for the optimization. Based on the current experiment results, the proposed method can massively reduce the number of parameters of commonly used networks (e.g., VGG networks).
Industrial Applicability
"Our sparse neural network learning method can be deployed on various applications. The computational cost of model inferences on deep neural networks can be massively reducedit can be beneficial to process neural network applications on IoT edge-devicespersonal mobile devices. 
Based on the concept, the PI has co-funded a company Omnieyes that applies the idea to Edge AI for video data 
Keyword Machine Learning Deep Learning Sparse Models Edge AI AI Applications Video Data Analysis IoT Artificial Neural Network Resource Constrained Situation Big Data Mining
Notes
  • Contact
other people also saw