Full metadata record
DC FieldValueLanguage
dc.contributorDepartment of Electronic and Information Engineeringen_US
dc.contributor.advisorLun, Daniel (EIE)en_US
dc.creatorPan, Yikun-
dc.identifier.urihttps://theses.lib.polyu.edu.hk/handle/200/10767-
dc.languageEnglishen_US
dc.publisherHong Kong Polytechnic Universityen_US
dc.rightsAll rights reserveden_US
dc.titleImage segmentation and inpainting based on deep learningen_US
dcterms.abstractBased on the deep learning methods, many intelligent image processing algorithms are employed in the field of computer vision in recent years. One of them is image segmentation, which is a commonly used technique in digital image processing for partitioning an image into multiple parts or regions based on the characteristics of the pixels in the image. Via learning the image features from batches of training samples, the convolutional neural network has become one of the powerful tools in the spatially dense prediction tasks of semantic segmentation. To make the network keep the global contextual information of the whole image, a pyramid pooling module is used in this project to accomplish different region-based context aggregation. Another application that the deep learning methods have been successfully applied is image inpainting, which aims at reconstructing the missing or corrupted regions within a picture or an image. For image inpainting, the Generative Adversarial Network (GAN) has been commonly used for generating the predicted outputs with a synthesis model trained based on the competition between a generator and a discriminator. For example, the two-stage adversarial model of the method EdgeConnect comprises an edge generator followed by an image completion network. It is adopted in this project for reproducing the missing regions exhibiting fine details. In this project, an entertaining application is developed by using image segmentation and an inpainting model that can interact with the users for removing the unwanted object in an image and filling with meaningful content. Users can mask out the unwanted object in an image through the segmentation model and fill the masked region with meaningful content through the inpainting model. During the stage of inpainting, users are allowed to sketch the outline of the objects inside the masked region. After receiving the input from the user, the system performs much better and can give the desired results that the user wants.en_US
dcterms.extent53 pages : color illustrationsen_US
dcterms.isPartOfPolyU Electronic Thesesen_US
dcterms.issued2020en_US
dcterms.educationalLevelM.Sc.en_US
dcterms.educationalLevelAll Masteren_US
dcterms.LCSHImage analysisen_US
dcterms.LCSHImage processingen_US
dcterms.LCSHMachine learningen_US
dcterms.LCSHHong Kong Polytechnic University -- Dissertationsen_US
dcterms.accessRightsrestricted accessen_US

Files in This Item:
File Description SizeFormat 
5169.pdfFor All Users (off-campus access for PolyU Staff & Students only)3.35 MBAdobe PDFView/Open


Copyright Undertaking

As a bona fide Library user, I declare that:

  1. I will abide by the rules and legal ordinances governing copyright regarding the use of the Database.
  2. I will use the Database for the purpose of my research or private study only and not for circulation or further reproduction or any other purpose.
  3. I agree to indemnify and hold the University harmless from and against any loss, damage, cost, liability or expenses arising from copyright infringement or unauthorized usage.

By downloading any item(s) listed above, you acknowledge that you have read and understood the copyright undertaking as stated above, and agree to be bound by all of its terms.

Show simple item record

Please use this identifier to cite or link to this item: https://theses.lib.polyu.edu.hk/handle/200/10767