Skip to main navigation Skip to search Skip to main content

BraNet: a mobil application for breast image classification based on deep learning algorithms

  • Yuliana Jiménez Gaona (First Author)
  • , María José Rodríguez Álvarez
  • , Darwin Castillo Malla
  • , Santiago García Jaen
  • , Diana Carrión Figueroa
  • , Patricio Corral Domínguez
  • , Vasudevan Lakshminarayanan (Last Author)
  • Universidad Técnica Particular de Loja
  • Universitat Politécnica de Valencia
  • School of Opto ΩN2L3G1
  • Hospital-IESS del Sur de Quito
  • Universidad de Cuenca
  • University of Waterloo

Research output: Contribution to journalArticlepeer-review

11 Scopus citations

Abstract

Mobile health apps are widely used for breast cancer detection using artificial intelligence algorithms, providing radiologists with second opinions and reducing false diagnoses. This study aims to develop an open-source mobile app named “BraNet” for 2D breast imaging segmentation and classification using deep learning algorithms. During the phase off-line, an SNGAN model was previously trained for synthetic image generation, and subsequently, these images were used to pre-trained SAM and ResNet18 segmentation and classification models. During phase online, the BraNet app was developed using the react native framework, offering a modular deep-learning pipeline for mammography (DM) and ultrasound (US) breast imaging classification. This application operates on a client–server architecture and was implemented in Python for iOS and Android devices. Then, two diagnostic radiologists were given a reading test of 290 total original RoI images to assign the perceived breast tissue type. The reader’s agreement was assessed using the kappa coefficient. The BraNet App Mobil exhibited the highest accuracy in benign and malignant US images (94.7%/93.6%) classification compared to DM during training I (80.9%/76.9%) and training II (73.7/72.3%). The information contrasts with radiological experts’ accuracy, with DM classification being 29%, concerning US 70% for both readers, because they achieved a higher accuracy in US ROI classification than DM images. The kappa value indicates a fair agreement (0.3) for DM images and moderate agreement (0.4) for US images in both readers. It means that not only the amount of data is essential in training deep learning algorithms. Also, it is vital to consider the variety of abnormalities, especially in the mammography data, where several BI-RADS categories are present (microcalcifications, nodules, mass, asymmetry, and dense breasts) and can affect the API accuracy model. Graphical abstract: (Figure presented.)

Original languageEnglish
Pages (from-to)2737-2756
Number of pages20
JournalMedical and Biological Engineering and Computing
Volume62
Issue number9
DOIs
StatePublished - Sep 2024
Externally publishedYes

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 3 - Good Health and Well-being
    SDG 3 Good Health and Well-being

Keywords

  • Breast cancer
  • Deep learning
  • Mammography
  • Mobil app
  • Ultrasound

Fingerprint

Dive into the research topics of 'BraNet: a mobil application for breast image classification based on deep learning algorithms'. Together they form a unique fingerprint.

Cite this