Full metadata record
DC FieldValueLanguage
dc.contributorDepartment of Electrical and Electronic Engineeringen_US
dc.contributor.advisorLam, Kin-man (EEE)en_US
dc.contributor.advisorLoo, Ka-hong (EEE)en_US
dc.creatorYeung, Ching Chi-
dc.identifier.urihttps://theses.lib.polyu.edu.hk/handle/200/13272-
dc.languageEnglishen_US
dc.publisherHong Kong Polytechnic Universityen_US
dc.rightsAll rights reserveden_US
dc.titleDeep learning for vision-based defect inspectionen_US
dcterms.abstractVision-based defect inspection is an essential quality control task in various industries. With the development of deep learning, deep learning-based visual defect inspection meth­ods have achieved remarkable performance. However, existing deep learning-based visual defect inspection models face three main challenges according to their specific applica­tion requirements, including inspection efficiency, precise localization and classification, and generalization ability. Therefore, this thesis aims to investigate deep learning-based models to address these challenges, which are particularly relevant to three specific appli­cations of vision-based defect inspection. These applications include steel surface defect detection, defect semantic segmentation, and pavement crack detection.en_US
dcterms.abstractIn this thesis, we study the efficient design of steel surface defect detection models. We propose a fused-attention network (FANet) to balance the trade-off between accuracy and speed. This model applies an attention mechanism to a single balanced feature map to improve accuracy while maintaining detection speed. Moreover, it introduces a feature fusion and an attention module to handle defects with multiple scales and shape variations.en_US
dcterms.abstractFurthermore, we investigate the model design to boost the localization and classifi­cation performance for defect semantic segmentation. We propose an attentive boundary-aware transformer framework, namely ABFormer, to precisely segment different types of defects. This framework introduces a feature fusion scheme to split and fuse the boundary and context features with two different attention modules. This facilitates the different learning aspects of the attention modules. In addition, the two attention modules cap­ture the spatial and channel interdependencies of the features, respectively, to address the intraclass difference and interclass indiscrimination problems.en_US
dcterms.abstractFinally, we focus on improving the generalization ability of pavement crack detec­tion models. We propose a contrastive decoupling network (CDNet) to effectively detect cracks in seen and unseen domains. This framework separately extracts global and local features with contrastive learning to produce generalized and discriminative representa­tions. Besides, it introduces a semantic enhancement module, detail refinement module, and feature aggregation scheme to tackle diverse cracks with complex backgrounds in input images.en_US
dcterms.abstractThe vision-based defect inspection models proposed in this thesis are evaluated by comparing them with other state-of-the-art methods on different defect inspection datasets. Experimental results validate that our models can achieve promising perfor­mance. These models have great potential to advance deep learning-based methods for various applications of vision-based defect inspection.en_US
dcterms.extentxxvi, 124 pages : color illustrationsen_US
dcterms.isPartOfPolyU Electronic Thesesen_US
dcterms.issued2024en_US
dcterms.educationalLevelPh.D.en_US
dcterms.educationalLevelAll Doctorateen_US
dcterms.LCSHDeep learning (Machine learning)en_US
dcterms.LCSHEngineering inspection -- Automationen_US
dcterms.LCSHComputer vision -- Industrial applicationsen_US
dcterms.LCSHSteel -- Surfaces --Defectsen_US
dcterms.LCSHImage segmentationen_US
dcterms.LCSHPavements --Crackingen_US
dcterms.LCSHHong Kong Polytechnic University -- Dissertationsen_US
dcterms.accessRightsopen accessen_US

Files in This Item:
File Description SizeFormat 
7718.pdfFor All Users7.78 MBAdobe PDFView/Open


Copyright Undertaking

As a bona fide Library user, I declare that:

  1. I will abide by the rules and legal ordinances governing copyright regarding the use of the Database.
  2. I will use the Database for the purpose of my research or private study only and not for circulation or further reproduction or any other purpose.
  3. I agree to indemnify and hold the University harmless from and against any loss, damage, cost, liability or expenses arising from copyright infringement or unauthorized usage.

By downloading any item(s) listed above, you acknowledge that you have read and understood the copyright undertaking as stated above, and agree to be bound by all of its terms.

Show simple item record

Please use this identifier to cite or link to this item: https://theses.lib.polyu.edu.hk/handle/200/13272