AI Model Achieves 99.24 % Accuracy in Breast Ultrasound Classification

A new AI system can detect breast cancer with near-perfect accuracy while also surfacing similar past cases for clinicians, adding context, clarity, and confidence to the diagnostic process.

Female doctor looking at a Mammogram film image.

Study: Ensemble-based high-performance deep learning models for medical image retrieval in breast cancer detection. Image Credit: Chompoo Suriyo/Shutterstock.com

In an article published in Scientific Reports, researchers introduced a medical image retrieval system built on deep learning. By combining convolutional neural networks (CNNs), recurrent neural networks (RNNs), and explainable artificial intelligence (XAI), the system aims to bridge the gap between visual data and clinical interpretation.

Trained on breast ultrasound images (BUSI), the hybrid model achieved 99.24% classification accuracy while also improving content-based image retrieval performance.

Background

Breast cancer is the second most commonly diagnosed cancer worldwide, and ultrasound remains a safe and accessible imaging modality for detection. As medical image databases continue to grow, content-based medical image retrieval (CBMIR) systems are becoming increasingly valuable. These systems enable clinicians to identify similar cases using visual features, supporting both diagnosis and treatment planning.

Previous deep learning approaches, particularly those using CNNs and attention mechanisms, have demonstrated strong performance. However, several limitations remain. Many studies rely on relatively small datasets, which can restrict generalizability. Cross-dataset validation is often limited, and explainability is not always integrated into model design. In addition, computational complexity can make real-world clinical deployment more challenging.

To address these gaps, the researchers of this study proposed a hybrid CNN-RNN architecture integrated with XAI. In this approach, CNNs handle spatial feature extraction, while RNNs model relationships within image representations. The addition of explainability techniques helps make predictions more transparent, improving both performance and clinical usability.

Model Architecture and Implementation

The study used the BUSI dataset, which includes 830 grayscale ultrasound images across three categories: normal (10.7 %), benign (62.4 %), and malignant (26.9 %). Originally stored in DICOM format, the images underwent extensive preprocessing. The dataset was organized into clearly labeled categories, and mask files were created to distinguish cancerous and non-cancerous regions. An 80:20 split was used for training and testing.

To evaluate performance, the researchers implemented four deep learning architectures.

The CNN model consisted of four convolutional layers with increasing filter sizes, ReLU activation, and max-pooling, followed by fully connected layers with dropout and a softmax output. The RNN model treated image data as sequences, using a long short-term memory (LSTM) layer to capture dependencies across pixel rows. The XAI model applied transfer learning with MobileNetV2 and used Grad-CAM to generate heatmaps highlighting regions that influenced predictions, alongside standard data augmentation techniques.

The proposed hybrid model integrates VGG16 for spatial feature extraction with an RNN to capture relationships within those features, while also incorporating Grad-CAM for interpretability. It also enables content-based image retrieval by using features from the penultimate layer as image signatures and comparing them using Euclidean distance to identify clinically similar cases.

Data augmentation techniques, including rotation, flipping, scaling, and contrast adjustment, expanded the training set significantly. Training was optimized using the Adam optimizer, categorical cross-entropy loss, early stopping, and adaptive learning rate adjustments.

Results and Performance Analysis

Performance was evaluated using accuracy, precision, recall, and F1-score, along with confusion matrix analysis to better understand misclassification patterns. The CNN model achieved 98.85 % accuracy with strong overall metrics and minimal errors.

The RNN model reached 94.58 % accuracy but showed some confusion between benign and normal cases. The XAI-based MobileNetV2 model achieved 89.68 % accuracy, offering interpretability at the cost of slightly lower classification performance.

The hybrid model delivered the strongest results, achieving 99.24 % accuracy along with the lowest training loss. Its confusion matrix showed near-perfect classification, with only minimal misclassifications across categories. Cross-validation further confirmed consistency, with a mean accuracy of 98.92 % and low variance. Grad-CAM visualizations aligned closely with clinically relevant regions, reinforcing confidence in the model’s predictions.

Comparative analysis showed that the hybrid model outperformed standalone approaches (CNN, RNN, and XAI) and compared favorably with previously reported methods such as DDA-Net and attention U-Net. The combination of spatial feature extraction, sequential modeling, and explainability appears to drive these improvements, while the retrieval component adds practical value by enabling clinicians to reference similar cases during diagnosis.

Conclusion

Overall, this study presents a hybrid deep learning model that combines CNNs, RNNs, and XAI for breast ultrasound image classification and retrieval.

Tested on the BUSI dataset, the model achieved 99.24 % accuracy, outperforming individual architectures and several comparable approaches. By integrating high performance with interpretability, the system moves closer to practical use in clinical workflows.

That said, the study remains limited to a single dataset and does not yet include external validation or clinical trial evaluation. Further work will need to expand testing across datasets and imaging modalities, incorporate patient-level data, explore newer architectures such as transformers, and validate performance through clinical studies involving radiologists.

Journal Reference

Fawzy, A. E., Almandouh, M. E., Herajy, M., & Eisa, M. (2026). Ensemble-based high-performance deep learning models for medical image retrieval in breast cancer detection. Scientific Reports, 16(1). DOI:10.1038/s41598-026-38218-y. https://www.nature.com/articles/s41598-026-38218-y

Disclaimer: The views expressed here are those of the author expressed in their private capacity and do not necessarily represent the views of AZoM.com Limited T/A AZoNetwork the owner and operator of this website. This disclaimer forms part of the Terms and conditions of use of this website.

Sources:

Citations

Please use one of the following formats to cite this article in your essay, paper or report:

  • APA

    Nandi, Soham. (2026, March 20). AI Model Achieves 99.24 % Accuracy in Breast Ultrasound Classification. AZoRobotics. Retrieved on March 20, 2026 from https://www.azorobotics.com/News.aspx?newsID=16359.

  • MLA

    Nandi, Soham. "AI Model Achieves 99.24 % Accuracy in Breast Ultrasound Classification". AZoRobotics. 20 March 2026. <https://www.azorobotics.com/News.aspx?newsID=16359>.

  • Chicago

    Nandi, Soham. "AI Model Achieves 99.24 % Accuracy in Breast Ultrasound Classification". AZoRobotics. https://www.azorobotics.com/News.aspx?newsID=16359. (accessed March 20, 2026).

  • Harvard

    Nandi, Soham. 2026. AI Model Achieves 99.24 % Accuracy in Breast Ultrasound Classification. AZoRobotics, viewed 20 March 2026, https://www.azorobotics.com/News.aspx?newsID=16359.

Tell Us What You Think

Do you have a review, update or anything you would like to add to this news story?

Leave your feedback
Your comment type
Submit

Sign in to keep reading

We're committed to providing free access to quality science. By registering and providing insight into your preferences you're joining a community of over 1m science interested individuals and help us to provide you with insightful content whilst keeping our service free.

or

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.