Exploring Accumulated Gradient-Based Quantization and Compression for Deep Neural Networks
The growing complexity of neural networks makes their deployment on resource-constrained embedded or mobile devices challenging. With millions of weights and biases, modern deep neural networks can be computationally intensive, with large memory, power and computational requirements. In this thesis,...
Main Author: | Gaopande, Meghana Laxmidhar |
---|---|
Other Authors: | Electrical and Computer Engineering |
Format: | Others |
Published: |
Virginia Tech
2020
|
Subjects: | |
Online Access: | http://hdl.handle.net/10919/98617 |
Similar Items
-
Learning Sparse Low-Precision Neural Networks With Learnable Regularization
by: Yoojin Choi, et al.
Published: (2020-01-01) -
Zero-Centered Fixed-Point Quantization With Iterative Retraining for Deep Convolutional Neural Network-Based Object Detectors
by: Sungrae Kim, et al.
Published: (2021-01-01) -
A Deep Learning Framework of Quantized Compressed Sensing for Wireless Neural Recording
by: Biao Sun, et al.
Published: (2016-01-01) -
Ps and Qs: Quantization-Aware Pruning for Efficient Low Latency Neural Network Inference
by: Benjamin Hawks, et al.
Published: (2021-07-01) -
Deep Learning Models Compression for Agricultural Plants
by: Arnauld Nzegha Fountsop, et al.
Published: (2020-09-01)