Full metadata record
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.contributor.advisorGuo, Song (COMP)en_US
dc.contributor.advisorXiao, Bin (COMP)en_US
dc.creatorWang, Yingchun-
dc.identifier.urihttps://theses.lib.polyu.edu.hk/handle/200/13522-
dc.languageEnglishen_US
dc.publisherHong Kong Polytechnic Universityen_US
dc.rightsAll rights reserveden_US
dc.titleTowards sample bias-aware deep neural network compressionen_US
dcterms.abstractDeep neural networks (DNNs) have achieved remarkable success in various fields like image classification, natural language processing, and speech synthesis. Their success often relies on a large number of parameters well organized to perform com­plex computations, introducing significant resource overhead. Over the past decade, research on DNN compression has proliferated, focusing on efficient architecture rep­resentations while overlooking the impact of inter-sample variations. In fact, model redundancy is highly sample-dependent, influenced by factors such as object type, en­vironmental context, and data quality. Therefore, this thesis systematically explores sample-oriented deep model compression methods, aiming to improve the performance of lightweight models in complex, real-world data environments.en_US
dcterms.abstractFirst, to address the channel mis-deletion caused by sample object bias and its re­sulting channel importance variations, we propose a global channel attention-based model pruning method, named GlobalPru, to improve the performance of statically pruned models. The overall pipeline can be divided into two stages: First, GlobalPru identifies a global channel ranking by a majority-voting-based strategy. Then, dur­ing sparse training, it pushes all sample-wise (local) channel attention to the global one via the learn-to-rank regularization. Finally, GlobalPru can execute a fit-to­-all (samples) pruning statically since all samples share the same ranking of channel relative importance. Extensive experiments demonstrate the effectiveness of the pro­posed method. For example, when benchmarked on ImageNet, GlobalPru can reduce 54.0% FLOPs of ResNet-50 with only 0.2% top-1 accuracy degradation, showing bet­ter performance in terms of both accuracy and computational cost compared to the state-of-the-art methods.en_US
dcterms.abstractSecond, to address the erroneous channel sparsity caused by sample environment (domain) bias and the resulting differences in pruning demands, we propose a novel spurious features-targeted model pruning method, named SFP, to achieve feature-targeted model pruning through a two-stage pipeline. During each iteration, SFP first identifies in-domain samples entangled with spurious features using a theoretical loss threshold. Then, it weakens the feature projection of these identified samples in the model space via a regularization term, thereby sparsifying branches that fit spurious features and effectively aligning the pruned model with invariant feature directions. Extensive experiments demonstrate the effectiveness of the proposed method. For instance, when benchmarked on DomainBed, SFP outperforms the state-of-the-art pruning method for out-of-distribution generalization, with significant accuracy im­provements of 2.9%, 9.4%, and 3.2% on VLCS, PACS, and OfficeHome, respectively.en_US
dcterms.abstractThird, to address the vulnerable quantized models caused by sample quality bias and its resulting quantization sensitivity differences, we propose a data quality-aware mixed-precision quantization method, named DQMQ, to allocate layer-wise bit-width conditioned on input samples dynamically. We first demonstrate that the optimal bit-width configurations are highly sample quality-dependent. Then, DQMQ is mod­eled as a hybrid reinforcement learning task that combines policy optimization for bit-width decision-making with supervised quantization training. By relaxing the dis­crete bit-width sampling to a continuous probability distribution encoded by a few learnable parameters, DQMQ is differentiable and can be optimized end-to-end with a hybrid objective of task accuracy and quantitative benefit. Extensive experiments demonstrate the effectiveness of the proposed method. For instance, when bench­marked on ImageNet, DQMQ improves accuracy by 2.1% for ResNet-18, 1.7% for ResNet-50, and 0.5% for MobileNet-V2, compared to state-of-the-art methods under the same compression rate.en_US
dcterms.abstractIn summary, this thesis presents an in-depth research of sample bias-aware deep neural network pruning and quantization, mainly focusing on the challenges arising from sample object bias, sample environment bias, and sample quality bias. The research provides theoretical and technical support for model compression in complex and dynamic real-world data environments. Extensive experiments demonstrate the effectiveness of our proposed methods.en_US
dcterms.extentxvii, 165 pages : color illustrationsen_US
dcterms.isPartOfPolyU Electronic Thesesen_US
dcterms.issued2024en_US
dcterms.educationalLevelPh.D.en_US
dcterms.educationalLevelAll Doctorateen_US
dcterms.LCSHNeural networks (Computer science)en_US
dcterms.LCSHDeep learning (Machine learning)en_US
dcterms.LCSHData compression (Computer science)en_US
dcterms.LCSHHong Kong Polytechnic University -- Dissertationsen_US
dcterms.accessRightsopen accessen_US

Files in This Item:
File Description SizeFormat 
7973.pdfFor All Users8.78 MBAdobe PDFView/Open


Copyright Undertaking

As a bona fide Library user, I declare that:

  1. I will abide by the rules and legal ordinances governing copyright regarding the use of the Database.
  2. I will use the Database for the purpose of my research or private study only and not for circulation or further reproduction or any other purpose.
  3. I agree to indemnify and hold the University harmless from and against any loss, damage, cost, liability or expenses arising from copyright infringement or unauthorized usage.

By downloading any item(s) listed above, you acknowledge that you have read and understood the copyright undertaking as stated above, and agree to be bound by all of its terms.

Show simple item record

Please use this identifier to cite or link to this item: https://theses.lib.polyu.edu.hk/handle/200/13522