Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor | Department of Electronic and Information Engineering | en_US |
dc.contributor.advisor | Lun, Daniel (EIE) | en_US |
dc.creator | Pan, Yikun | - |
dc.identifier.uri | https://theses.lib.polyu.edu.hk/handle/200/10767 | - |
dc.language | English | en_US |
dc.publisher | Hong Kong Polytechnic University | en_US |
dc.rights | All rights reserved | en_US |
dc.title | Image segmentation and inpainting based on deep learning | en_US |
dcterms.abstract | Based on the deep learning methods, many intelligent image processing algorithms are employed in the field of computer vision in recent years. One of them is image segmentation, which is a commonly used technique in digital image processing for partitioning an image into multiple parts or regions based on the characteristics of the pixels in the image. Via learning the image features from batches of training samples, the convolutional neural network has become one of the powerful tools in the spatially dense prediction tasks of semantic segmentation. To make the network keep the global contextual information of the whole image, a pyramid pooling module is used in this project to accomplish different region-based context aggregation. Another application that the deep learning methods have been successfully applied is image inpainting, which aims at reconstructing the missing or corrupted regions within a picture or an image. For image inpainting, the Generative Adversarial Network (GAN) has been commonly used for generating the predicted outputs with a synthesis model trained based on the competition between a generator and a discriminator. For example, the two-stage adversarial model of the method EdgeConnect comprises an edge generator followed by an image completion network. It is adopted in this project for reproducing the missing regions exhibiting fine details. In this project, an entertaining application is developed by using image segmentation and an inpainting model that can interact with the users for removing the unwanted object in an image and filling with meaningful content. Users can mask out the unwanted object in an image through the segmentation model and fill the masked region with meaningful content through the inpainting model. During the stage of inpainting, users are allowed to sketch the outline of the objects inside the masked region. After receiving the input from the user, the system performs much better and can give the desired results that the user wants. | en_US |
dcterms.extent | 53 pages : color illustrations | en_US |
dcterms.isPartOf | PolyU Electronic Theses | en_US |
dcterms.issued | 2020 | en_US |
dcterms.educationalLevel | M.Sc. | en_US |
dcterms.educationalLevel | All Master | en_US |
dcterms.LCSH | Image analysis | en_US |
dcterms.LCSH | Image processing | en_US |
dcterms.LCSH | Machine learning | en_US |
dcterms.LCSH | Hong Kong Polytechnic University -- Dissertations | en_US |
dcterms.accessRights | restricted access | en_US |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
5169.pdf | For All Users (off-campus access for PolyU Staff & Students only) | 3.35 MB | Adobe PDF | View/Open |
Copyright Undertaking
As a bona fide Library user, I declare that:
- I will abide by the rules and legal ordinances governing copyright regarding the use of the Database.
- I will use the Database for the purpose of my research or private study only and not for circulation or further reproduction or any other purpose.
- I agree to indemnify and hold the University harmless from and against any loss, damage, cost, liability or expenses arising from copyright infringement or unauthorized usage.
By downloading any item(s) listed above, you acknowledge that you have read and understood the copyright undertaking as stated above, and agree to be bound by all of its terms.
Please use this identifier to cite or link to this item:
https://theses.lib.polyu.edu.hk/handle/200/10767