Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor | Department of Electronic and Information Engineering | en_US |
dc.contributor.advisor | Siu, Wan-chi (EIE) | en_US |
dc.contributor.advisor | Lun, P. K. Daniel (EIE) | en_US |
dc.creator | Wang, Liwen | - |
dc.identifier.uri | https://theses.lib.polyu.edu.hk/handle/200/13033 | - |
dc.language | English | en_US |
dc.publisher | Hong Kong Polytechnic University | en_US |
dc.rights | All rights reserved | en_US |
dc.title | Deep learning technologies for scene enlightening and road safety | en_US |
dcterms.abstract | Vision is the most important sense of our human beings. Similarly, image/video signal is one of the most popular ways for machines to precept the external world. Benefited from the increasement of the computational power, both conventional and deep learning based machine learning methods have drawn great attention in the computer vision field. These learning-based methods make it possible to utilize the captured digital vision-based signals automatically. However, because of the insufficient illumination, the camera sometimes cannot produce pleasing pictures in low-light conditions, which limits the use-value of the obtained visual signal. This thesis gives focuses on image enhancement in low-light conditions, and scene analysis with emphasis on object recognition. | en_US |
dcterms.abstract | Low-light image enhancement is a challenging task. To address the problem, we regard the low-light enhancement as a residual learning problem that is to estimate the residual between low- and normal-light images. We propose a novel Deep Lightening Network (DLN) that consists of several Lightening Back-Projection (LBP) blocks. The LBPs perform lightening and darkening processes iteratively to learn the residual for normal-light estimations. To effectively utilize the local and global features, we also propose a Feature Aggregation (FA) block that adaptively fuses the results of different LBPs. Experiments show the proposed method is very effective. | en_US |
dcterms.abstract | Because the raw signal captured by the Color Filter Array (CFA) contains the richest digital information, we further extend our method to work on the camera sensors to process raw signals, which produce standard RGB (sRGB) images with more pleasing visual quality. This helps the night surveillance system or autonomous cars with better input signals. Because the initial raw signal is noisy under the dark environment, a denoising operator is proposed to remove the influence of the noises. And a weighting operation is introduced to adjust the compensation adaptively. Experimental works has been done, results of which demonstrate that the proposed method is very effective. | en_US |
dcterms.abstract | Besides working on the image processing task, we explored automatic video analysis methods especially for the road videos. Locating the distance between two points on the road should be extremely useful for autonomous driving, intelligent traffic system, etc. To enable the road cameras with real-world perception, we propose a novel scheme that can automatically build the connection between pixel measures and real-world distances. In our design, a pixel is assigned to have a weight unit which consists of two orthogonal weights on the scene. Then, an efficient Distance Estimation Network (DEN) is proposed to generate two maps of the distance weights on the road. Based on prior knowledge of the road videos, we propose a set of constraints to assist the training of the DEN. The approach is novel, which does not need the user to explicitly enter any parameter into the system. We have evaluated the proposed method in different scenes. Numerical results show that the proposed method can estimate distances on the roads automatically and correctly, and actually this approach is being used in one of our recent consultancy projects. | en_US |
dcterms.extent | 237 pages : color illustrations | en_US |
dcterms.isPartOf | PolyU Electronic Theses | en_US |
dcterms.issued | 2022 | en_US |
dcterms.educationalLevel | Ph.D. | en_US |
dcterms.educationalLevel | All Doctorate | en_US |
dcterms.LCSH | Computer vision | en_US |
dcterms.LCSH | Deep learning (Machine learning) | en_US |
dcterms.LCSH | Image processing -- Digital techniques | en_US |
dcterms.LCSH | Hong Kong Polytechnic University -- Dissertations | en_US |
dcterms.accessRights | open access | en_US |
Copyright Undertaking
As a bona fide Library user, I declare that:
- I will abide by the rules and legal ordinances governing copyright regarding the use of the Database.
- I will use the Database for the purpose of my research or private study only and not for circulation or further reproduction or any other purpose.
- I agree to indemnify and hold the University harmless from and against any loss, damage, cost, liability or expenses arising from copyright infringement or unauthorized usage.
By downloading any item(s) listed above, you acknowledge that you have read and understood the copyright undertaking as stated above, and agree to be bound by all of its terms.
Please use this identifier to cite or link to this item:
https://theses.lib.polyu.edu.hk/handle/200/13033