Author: | Luo, Boyuan |
Title: | Int8-based fully quantization-aware training |
Advisors: | Guo, Song (COMP) |
Degree: | M.Sc. |
Year: | 2021 |
Subject: | Neural networks (Computer science) Machine learning Hong Kong Polytechnic University -- Dissertations |
Department: | Department of Computing |
Pages: | [32] pages : color illustrations |
Language: | English |
Abstract: | Deep Neural Networks have shown great success to handle various real-world applications, where huge computational overhead is required to drive the resource-hungry training procedure. With the increase of model complexity, the limited computational capacity becomes the performance bottleneck of modern learning tasks, especially when dealing with a great number of tensor-level arithmetic operations. Recently, quantizing number into low-precision data formats is a promising research direction to address the above challenge. However, most existing methods focus on post-training quantization and ultra-low-bit neural networks, where the computational primitives cannot be fully utilized. A natural manner to make the quantization algorithm hardware-friendly is to exploit the power of 8-bit fixed-point instructions, which hold fewer resource demands over the conventional 32-bit floating-point operations. This property motivates us to propose a novel INT8-based Fully quantization-aware training(FQAT) algorithm, which quantizes model parameters in both forward and backward pass, including weights, activations and gradients. The proposed FQAT can be deployed in extensive usage scenarios, including powerful Nvidia's GPU, Intel's CPU, and resource-constrained devices, such as Raspberry Pi development board. Compared with the non-uniform and uniform quantization scheme, I choose the uniform linear quantization scheme to match the limited on-device computational capacity. Besides, I jointly design batch normalization and range clipping by simplifying these operations into a single function, named Clipping Batch normalization. I implement the proposed algorithms on commodity Deep Learning frameworks based on Python and Numpy, which is the first to enable on-device training from scratch. Experimental results show that FQAT can effectively handle training tasks on resource-constrained embedded devices on MNIST and CIFAR10 datasets. |
Rights: | All rights reserved |
Access: | restricted access |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
5827.pdf | For All Users (off-campus access for PolyU Staff & Students only) | 909.91 kB | Adobe PDF | View/Open |
Copyright Undertaking
As a bona fide Library user, I declare that:
- I will abide by the rules and legal ordinances governing copyright regarding the use of the Database.
- I will use the Database for the purpose of my research or private study only and not for circulation or further reproduction or any other purpose.
- I agree to indemnify and hold the University harmless from and against any loss, damage, cost, liability or expenses arising from copyright infringement or unauthorized usage.
By downloading any item(s) listed above, you acknowledge that you have read and understood the copyright undertaking as stated above, and agree to be bound by all of its terms.
Please use this identifier to cite or link to this item:
https://theses.lib.polyu.edu.hk/handle/200/11389