VSG-GAN: A high-fidelity image synthesis method with semantic manipulation in retinal fundus image.
Date
2024-09
Journal Title
Journal ISSN
Volume Title
Repository Usage Stats
views
downloads
Citation Stats
Attention Stats
Abstract
In recent years, advancements in retinal image analysis, driven by machine learning and deep learning techniques, have enhanced disease detection and diagnosis through automated feature extraction. However, challenges persist, including limited data set diversity due to privacy concerns and imbalanced sample pairs, hindering effective model training. To address these issues, we introduce the vessel and style guided generative adversarial network (VSG-GAN), an innovative algorithm building upon the foundational concept of GAN. In VSG-GAN, a generator and discriminator engage in an adversarial process to produce realistic retinal images. Our approach decouples retinal image generation into distinct modules: the vascular skeleton and background style. Leveraging style transformation and GAN inversion, our proposed hierarchical variational autoencoder module generates retinal images with diverse morphological traits. In addition, the spatially adaptive denormalization module ensures consistency between input and generated images. We evaluate our model on MESSIDOR and RITE data sets using various metrics, including structural similarity index measure, inception score, Fréchet inception distance, and kernel inception distance. Our results demonstrate the superiority of VSG-GAN, outperforming existing methods across all evaluation assessments. This underscores its effectiveness in addressing data set limitations and imbalances. Our algorithm provides a novel solution to challenges in retinal image analysis by offering diverse and realistic retinal image generation. Implementing the VSG-GAN augmentation approach on downstream diabetic retinopathy classification tasks has shown enhanced disease diagnosis accuracy, further advancing the utility of machine learning in this domain.
Type
Department
Description
Provenance
Subjects
Citation
Permalink
Published Version (Please cite this version)
Publication Info
Liu, Junjie, Shixin Xu, Ping He, Sirong Wu, Xi Luo, Yuhui Deng and Huaxiong Huang (2024). VSG-GAN: A high-fidelity image synthesis method with semantic manipulation in retinal fundus image. Biophysical journal, 123(17). pp. 2815–2829. 10.1016/j.bpj.2024.02.019 Retrieved from https://hdl.handle.net/10161/33540.
This is constructed from limited available data and may be imprecise. To cite this article, please review & use the official citation provided by the journal.
Collections
Scholars@Duke
Shixin Xu
Shixin Xu is an Assistant Professor of Mathematics whose research spans several dynamic and interconnected fields. His primary interests include machine learning and data-driven models for disease prediction, multiscale modeling of complex fluids, neurovascular coupling, homogenization theory, and numerical analysis. His current projects reflect a diverse and impactful portfolio:
- Developing predictive models based on image data to identify hemorrhagic transformation in acute ischemic stroke.
- Conducting electrodynamics modeling of saltatory conduction along myelinated axons to understand nerve impulse transmission.
- Engaging in electrochemical modeling to explore the interactions between electric fields and chemical processes.
- Investigating fluid-structure interactions with mass transport and reactions, crucial for understanding physiological and engineering systems.
These projects demonstrate his commitment to addressing complex problems through interdisciplinary approaches that bridge mathematics with biological and physical sciences.
Unless otherwise indicated, scholarly articles published by Duke faculty members are made available here with a CC-BY-NC (Creative Commons Attribution Non-Commercial) license, as enabled by the Duke Open Access Policy. If you wish to use the materials in ways not already permitted under CC-BY-NC, please consult the copyright owner. Other materials are made available here through the author’s grant of a non-exclusive license to make their work openly accessible.
