Articles producció científica> Enginyeria Informàtica i Matemàtiques

FGR-Net: Interpretable fundus image gradeability classification based on deep reconstruction learning

  • Datos identificativos

    Identificador: imarina:9331246
    Autores:
    Khalid, SaifRashwan, Hatem AAbdulwahab, SaddamAbdel-Nasser, MohamedQuiroga, Facundo ManuelPuig, Domenec
    Resumen:
    The performance of diagnostic Computer-Aided Design (CAD) systems for retinal diseases depends on the quality of the retinal images being screened. Thus, many studies have been developed to evaluate and assess the quality of such retinal images. However, most of them did not investigate the relationship between the accuracy of the developed models and the quality of the visualization of interpretability methods for distinguishing between gradable and non-gradable retinal images. Consequently, this paper presents a novel framework called “FGR-Net” to automatically assess and interpret underlying fundus image quality by merging an autoencoder network with a classifier network. The FGR-Net model also provides an interpretable quality assessment through visualizations. In particular, FGR-Net uses a deep autoencoder to reconstruct the input image in order to extract the visual characteristics of the input fundus images based on self-supervised learning. The extracted features by the autoencoder are then fed into a deep classifier network to distinguish between gradable and ungradable fundus images. FGR-Net is evaluated with different interpretability methods, which indicates that the autoencoder is a key factor in forcing the classifier to focus on the relevant structures of the fundus images, such as the fovea, optic disk, and prominent blood vessels. Additionally, the interpretability methods can provide visual feedback for ophthalmologists to understand how our model evaluates the quality of fundus images. The experimental results showed the superiority of FGR-Net over the state-of-the-art quality assessment methods, with an accuracy of >89% and an F1-score of >87%. The code is publicly available at https://github.com/saifalkh/FGR-Net.
  • Otros:

    Autor según el artículo: Khalid, Saif; Rashwan, Hatem A; Abdulwahab, Saddam; Abdel-Nasser, Mohamed; Quiroga, Facundo Manuel; Puig, Domenec
    Departamento: Enginyeria Informàtica i Matemàtiques
    Autor/es de la URV: Abdellatif Fatahallah Ibrahim Mahmoud, Hatem / Abdelnasser Mohamed Mahmoud, Mohamed / Abdulwahab, Saddam Abdulrhman Hamed / Puig Valls, Domènec Savi
    Palabras clave: Retinal image Quality assessment Ocular diseases Neural-network model Intepretability Gradability Explainability Deep learning Autoencoder network segmentation quality assessment ocular diseases intepretability gradability explainability diabetic-retinopathy deep learning autoencoder network
    Resumen: The performance of diagnostic Computer-Aided Design (CAD) systems for retinal diseases depends on the quality of the retinal images being screened. Thus, many studies have been developed to evaluate and assess the quality of such retinal images. However, most of them did not investigate the relationship between the accuracy of the developed models and the quality of the visualization of interpretability methods for distinguishing between gradable and non-gradable retinal images. Consequently, this paper presents a novel framework called “FGR-Net” to automatically assess and interpret underlying fundus image quality by merging an autoencoder network with a classifier network. The FGR-Net model also provides an interpretable quality assessment through visualizations. In particular, FGR-Net uses a deep autoencoder to reconstruct the input image in order to extract the visual characteristics of the input fundus images based on self-supervised learning. The extracted features by the autoencoder are then fed into a deep classifier network to distinguish between gradable and ungradable fundus images. FGR-Net is evaluated with different interpretability methods, which indicates that the autoencoder is a key factor in forcing the classifier to focus on the relevant structures of the fundus images, such as the fovea, optic disk, and prominent blood vessels. Additionally, the interpretability methods can provide visual feedback for ophthalmologists to understand how our model evaluates the quality of fundus images. The experimental results showed the superiority of FGR-Net over the state-of-the-art quality assessment methods, with an accuracy of >89% and an F1-score of >87%. The code is publicly available at https://github.com/saifalkh/FGR-Net.
    Áreas temáticas: Química Operations research & management science Medicina iii Medicina ii Medicina i Materiais Matemática / probabilidade e estatística Interdisciplinar Geociências General engineering Farmacia Engineering, electrical & electronic Engineering (miscellaneous) Engineering (all) Engenharias iv Engenharias iii Engenharias ii Engenharias i Enfermagem Educação Economia Direito Computer science, artificial intelligence Computer science applications Ciências sociais aplicadas i Ciências biológicas iii Ciências biológicas ii Ciências biológicas i Ciências ambientais Ciências agrárias i Ciência da computação Biotecnología Biodiversidade Astronomia / física Artificial intelligence Arquitetura, urbanismo e design Administração, ciências contábeis e turismo Administração pública e de empresas, ciências contábeis e turismo
    Acceso a la licencia de uso: https://creativecommons.org/licenses/by/3.0/es/
    Direcció de correo del autor: saddam.abdulwahab@urv.cat mohamed.abdelnasser@urv.cat hatem.abdellatif@urv.cat saddam.abdulwahab@urv.cat domenec.puig@urv.cat
    Identificador del autor: 0000-0002-1074-2441 0000-0001-5421-1637 0000-0002-0562-4205
    Fecha de alta del registro: 2024-10-12
    Versión del articulo depositado: info:eu-repo/semantics/publishedVersion
    URL Documento de licencia: https://repositori.urv.cat/ca/proteccio-de-dades/
    Referencia al articulo segun fuente origial: Expert Systems With Applications. 238 121644-
    Referencia de l'ítem segons les normes APA: Khalid, Saif; Rashwan, Hatem A; Abdulwahab, Saddam; Abdel-Nasser, Mohamed; Quiroga, Facundo Manuel; Puig, Domenec (2024). FGR-Net: Interpretable fundus image gradeability classification based on deep reconstruction learning. Expert Systems With Applications, 238(), 121644-. DOI: 10.1016/j.eswa.2023.121644
    Entidad: Universitat Rovira i Virgili
    Año de publicación de la revista: 2024
    Tipo de publicación: Journal Publications
  • Palabras clave:

    Artificial Intelligence,Computer Science Applications,Computer Science, Artificial Intelligence,Engineering (Miscellaneous),Engineering, Electrical & Electronic,Operations Research & Management Science
    Retinal image
    Quality assessment
    Ocular diseases
    Neural-network model
    Intepretability
    Gradability
    Explainability
    Deep learning
    Autoencoder network
    segmentation
    quality assessment
    ocular diseases
    intepretability
    gradability
    explainability
    diabetic-retinopathy
    deep learning
    autoencoder network
    Química
    Operations research & management science
    Medicina iii
    Medicina ii
    Medicina i
    Materiais
    Matemática / probabilidade e estatística
    Interdisciplinar
    Geociências
    General engineering
    Farmacia
    Engineering, electrical & electronic
    Engineering (miscellaneous)
    Engineering (all)
    Engenharias iv
    Engenharias iii
    Engenharias ii
    Engenharias i
    Enfermagem
    Educação
    Economia
    Direito
    Computer science, artificial intelligence
    Computer science applications
    Ciências sociais aplicadas i
    Ciências biológicas iii
    Ciências biológicas ii
    Ciências biológicas i
    Ciências ambientais
    Ciências agrárias i
    Ciência da computação
    Biotecnología
    Biodiversidade
    Astronomia / física
    Artificial intelligence
    Arquitetura, urbanismo e design
    Administração, ciências contábeis e turismo
    Administração pública e de empresas, ciências contábeis e turismo
  • Documentos:

  • Cerca a google

    Search to google scholar