Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor | Department of Computing | en_US |
dc.creator | Yang, Zhongqi | - |
dc.identifier.uri | https://theses.lib.polyu.edu.hk/handle/200/11834 | - |
dc.language | English | en_US |
dc.publisher | Hong Kong Polytechnic University | en_US |
dc.rights | All rights reserved | en_US |
dc.title | Analysis of smartphone images for vision screening of refractive errors | en_US |
dcterms.abstract | Refractive error is the most common of visual impairments and impacts millions of people globally. Regular vision screening is the recommended best strategy to ensure timely diagnosis and treatment, however, many people do not have access to optometric care and a comprehensive vision examination is inaccessible to many people. There is therefore a need for fast, lowcost and easilyoperate vision screening approaches. In this thesis, we aim to investigate the possibility of conducting photorefraction, a common vision screening procedure, on the mobile platform, to address the challenge. | en_US |
dcterms.abstract | Our approach exploits machine learning algorithms and computer vision techniques. Starting from principles from optometry and prior studies, we create several handcrafted features corresponding detection methods. The experiment results indicate that our detection methods outperform contemporary approaches, leading to a better performance of refractive error measurement and amblyopia risk factor detection. We then move on to pretrained features extracted by convolutional neural networks (CNN). We employ the convolutional layers from multiple pretrained CNN models to encode features and train machine learning models to predict the refractive error. The experiments show promising results, even though the CNN models were not trained on photorefraction datasets. | en_US |
dcterms.abstract | Given these encouraging results, we further investigate the possibility of data augmentation. One of our challenges is that it is not possible to collect a large amount of data which is enough to train a wellperforming CNN model from scratch. Therefore, we investigate the use of synthetic data for augmentation. We develop a model of the eye based on the principle of photorefraction, and use it to generate synthetic pupil images with predetermined refractive errors. Evaluation results show that models trained on these synthetic pupil images can achieve similar performance as real images on multiple experiments, which provides solid evidence for the correctness of our photorefraction model. | en_US |
dcterms.abstract | We finally apply transfer learning to solve the insufficient data issue. CNN models pretrained on largescale public image datasets are finetuned with photorefraction images and the experiments results show large improvement. The CNN models are then trained on more than 10,000 images of synthetic eyes generated via our eye model, and finetuned using real images, achieving performances that outperform all of the previous models. These results support the feasibility of the proposed photorefraction model, and provides a novel direction to obtain training data, which may be extensible to other similar domains. | en_US |
dcterms.extent | xiv, 112 pages : color illustrations | en_US |
dcterms.isPartOf | PolyU Electronic Theses | en_US |
dcterms.issued | 2022 | en_US |
dcterms.educationalLevel | M.Phil. | en_US |
dcterms.educationalLevel | All Master | en_US |
dcterms.LCSH | Eye -- Refractive errors -- Diagnosis -- Data processing | en_US |
dcterms.LCSH | Vision -- Testing -- Data processing | en_US |
dcterms.LCSH | Smartphones -- Programming | en_US |
dcterms.LCSH | Machine learning | en_US |
dcterms.LCSH | Hong Kong Polytechnic University -- Dissertations | en_US |
dcterms.accessRights | open access | en_US |
Copyright Undertaking
As a bona fide Library user, I declare that:
- I will abide by the rules and legal ordinances governing copyright regarding the use of the Database.
- I will use the Database for the purpose of my research or private study only and not for circulation or further reproduction or any other purpose.
- I agree to indemnify and hold the University harmless from and against any loss, damage, cost, liability or expenses arising from copyright infringement or unauthorized usage.
By downloading any item(s) listed above, you acknowledge that you have read and understood the copyright undertaking as stated above, and agree to be bound by all of its terms.
Please use this identifier to cite or link to this item:
https://theses.lib.polyu.edu.hk/handle/200/11834