Next Article in Journal
Socioeconomic Deprivation and Invasive Breast Cancer Incidence by Stage at Diagnosis: A Possible Explanation to the Breast Cancer Social Paradox
Previous Article in Journal
Rehabilitation in People Living with Glioblastoma: A Narrative Review of the Literature
Previous Article in Special Issue
Endoscopic Contrast-Enhanced Ultrasound and Fine-Needle Aspiration or Biopsy for the Diagnosis of Pancreatic Solid Lesions: A Systematic Review and Meta-Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Deep Learning for Real-Time Ablation Zone Measurement in Ultrasound Imaging

1
Erbe Elektromedizin GmbH, 72072 Tübingen, Germany
2
Faculty of Medicine, University of Tuzla, 75000 Tuzla, Bosnia and Herzegovina
*
Author to whom correspondence should be addressed.
Submission received: 3 April 2024 / Revised: 24 April 2024 / Accepted: 26 April 2024 / Published: 27 April 2024
(This article belongs to the Special Issue Endoscopic Ultrasound in Cancer Research)

Abstract

:

Simple Summary

The manual measurement of ablation zones (AZs) in radiofrequency ablation (RFA) therapy is prone to inaccuracies, highlighting the need for automated methods. Our study investigated the effectiveness of an Artificial Intelligence (AI) model, Mask2Former, in automating AZ measurements from ultrasound images, comparing its performance against manual techniques. Conducted on chicken breast and liver samples, the study found the AI model to achieve high accuracy, particularly in chicken breast tissue, with no significant difference in measurements between AI and manual methods. These results suggest that the Mask2Former model can significantly reduce variability in manual measurements, marking a step forward in the automation of AZ measurement in RFA therapy research and potentially improving the precision of treatment assessments.

Abstract

Background: The accurate delineation of ablation zones (AZs) is crucial for assessing radiofrequency ablation (RFA) therapy’s efficacy. Manual measurement, the current standard, is subject to variability and potential inaccuracies. Aim: This study aims to assess the effectiveness of Artificial Intelligence (AI) in automating AZ measurements in ultrasound images and compare its accuracy with manual measurements in ultrasound images. Methods: An in vitro study was conducted using chicken breast and liver samples subjected to bipolar RFA. Ultrasound images were captured every 15 s, with the AI model Mask2Former trained for AZ segmentation. The measurements were compared across all methods, focusing on short-axis (SA) metrics. Results: We performed 308 RFA procedures, generating 7275 ultrasound images across liver and chicken breast tissues. Manual and AI measurement comparisons for ablation zone diameters revealed no significant differences, with correlation coefficients exceeding 0.96 in both tissues (p < 0.001). Bland–Altman plots and a Deming regression analysis demonstrated a very close alignment between AI predictions and manual measurements, with the average difference between the two methods being −0.259 and −0.243 mm, for bovine liver and chicken breast tissue, respectively. Conclusion: The study validates the Mask2Former model as a promising tool for automating AZ measurement in RFA research, offering a significant step towards reducing manual measurement variability.

1. Introduction

The battle against cancer remains a formidable challenge in the quest to extend the human lifespan into the 21st century [1]. In this context, minimally invasive thermal ablation techniques, such as radiofrequency ablation (RFA), microwave ablation (MWA), and high-intensity focused ultrasound (HIFU), have emerged as effective options for the treatment of tumors in organs like the liver, lung, kidney, and bone [2]. These techniques employ the targeted thermal destruction of cancerous tissues, offering a critical advantage by minimizing damage to surrounding healthy tissues [3].
Radiofrequency ablation has ascended as a prominent treatment modality, suitable for both curative and palliative objectives [4,5]. Its efficacy is most pronounced in tumors less than 3 cm in diameter, where a precise application of alternating current generates significant hyperthermia to induce tumor cell death [5,6]. The affected area, known as the ablation zone (AZ), is typically modeled as a three-dimensional spheroid characterized by one long axis and two short axes (SAs), which correspond to the dimensions covered by the ablation [7,8,9].
Despite its benefits, the success of RFA and other ablation methods hinges on the accurate assessment and monitoring of the AZ, necessitating advanced imaging techniques for real-time guidance [10]. The advent of image-guided interventions has significantly enhanced the precision of ablation therapies [11]. Although traditional imaging modalities like computed tomography (CT) and magnetic resonance imaging (MRI) provide detailed anatomical visualization, their utility is limited by the lack of real-time feedback and high operational costs [5,12,13]. Ultrasound (US) imaging, by contrast, offers a cost-effective and real-time alternative for monitoring ablation procedures [14]. Nevertheless, challenges such as image artifacts and inter-operator variability underscore the need for improvements in imaging accuracy and interpretation [5,14,15,16].
The integration of deep-learning algorithms with ultrasound imaging presents a promising avenue for overcoming these limitations [17,18,19,20]. By automating the recognition and measurement of the ablation zone, deep-learning (DL) models can potentially enhance the precision, reproducibility, and efficiency of thermal ablation therapies [21,22]. Convolutional neural networks (CNNs), as a particularly appropriate method of DL, were proven as a good method for object recognition and characterization [23,24,25,26]. Recent advancements in computer vision have seen a transition from CNNs to Transformer-based architectures, presenting a paradigm shift in imaging tasks. CNNs excel in image tasks but struggle with long-range dependencies vital for granular recognition. Transformer-based architectures, initially designed for natural language processing, offer a promising alternative. These models excel at processing complex spatial relationships within images and offer significant improvements in how imaging data are interpreted, providing a detailed semantic segmentation that can enhance real-time monitoring and procedural accuracy [27].
The objective of this study is, therefore, to explore the efficacy of employing a Transformer-based architecture for the automatic segmentation of the AZ in US images. This investigation is also aimed at facilitating the real-time monitoring of RFA progress, with the additional goal of developing an experimental setup that integrates US imaging with DL technologies.

2. Materials and Methods

2.1. Data Acquisition and Experimental Setup

The experimental setup for this study was designed to investigate the efficacy of RFA using a bipolar RFA probe (Erbe Elektromedizin GmbH, Tübingen, Germany). The primary objective was to create ablation zones (AZs) of varying sizes to assess the performance of US imaging in conjunction with deep learning for AZ segmentation and measurement.
The RFA procedures were conducted on two types of tissue: bovine liver tissue, a commonly used surrogate for human liver in ex vivo RFA studies, and chicken breast tissue. These tissues were chosen based on their electrical properties, availability, different tissue density, and the distinct color change upon coagulation at temperatures above 60 °C, facilitating visual identification of the AZ [7,28,29]. Both tissue types were maintained at room temperature throughout the experiments.
A self-designed test stand that can ensure precise positioning of the RFA probe and the US transducer was developed (Figure 1). The RFA probe was horizontally inserted into the tissue, while the US transducer, attached to a handheld wireless linear US scanner (L7HD, Clarius Mobile Health Corp., Vancouver, BC, Canada), was positioned perpendicularly above the probe. This setup allowed for uniform imaging conditions and minimized operator-induced variability.
We used five distinct RFA durations (30, 120, 300, 600, and 900 s) to generate AZs of different sizes. The “muscle setting” on the Clarius scanner was used, and the imaging depth was set to 3 cm to ensure a consistent scale across all images. US images were captured automatically every 15 s during RFA activation, providing a detailed temporal record of the AZ development.

2.2. Image Processing, Deep-Learning Model, and Analysis

In the essential phase of image processing and analysis, this study employed an advanced deep-learning architecture, specifically the Mask2Former model, to analyze ultrasound (US) images captured during radiofrequency ablation (RFA) procedures. Mask2Former was chosen because of its high accuracy for ultrasound imaging processing compared to other architectures like Mask R-CNN or SOLO [30]. The initial step involved adopting the foundational code of the Mask2Former model to handle the segmentation of the AZ within US images. The architecture consists of an encoder–decoder structure and integrates a pixel decoder and Transformer decoder in the decoder stage. In the encoder, a Swin-B Transformer is used as a backbone, resulting in an overall model size of 107M parameters [27,31].
Given the limited number of images accessible for model training, the study leveraged transfer learning techniques, utilizing pre-trained weights from the Microsoft Common Objects in Context (MS COCO) dataset [32]. This approach was complemented by data augmentation strategies, including image cropping, vertical mirroring, and contrast adjustments, to enrich the dataset artificially, ensuring a robust training process despite the dataset’s constraints [33].
The model’s training was conducted through supervised learning, utilizing two distinct datasets of US images. Each dataset underwent meticulous annotation by an expert, ensuring accurate delineation of the AZ. Prior to training, individual normalization procedures were applied to the images within each dataset to cover their respective ranges. Separate models were trained for liver and chicken breast to maintain consistent accuracy across different tissue types. The datasets were then partitioned into training, validation, and test sets, following an 80/10/10 split, resulting in a distribution of 2397 training, 303 validations, and 299 test images for chicken breast tissue, and 3421 training, 457 validation, and 398 test images for liver tissue.

2.3. Short-Axis (SA) Measurement

The accurate measurement of the AZ short axis is essential for evaluating the effectiveness of RFA procedures. We employed several distinct approaches to SA measurement. Manual measurement was performed on the horizontally sectioned tissue using a caliper, but, as we were able to do so only after the RFA was completed, without the possibility for real-time serial measurement, this was dismissed as a possibility for a ground-truth standard.
Instead, we used the value of SA measured directly in each individual US image taken over the entire RFA run. These measurements were referred to as US diameters.
Finally, the approach that we developed for the purpose of this research involved analyzing US images through a Mask2Former model (AI method), which predicted the AZ’s mask. This method identified the largest horizontal span within the predicted mask as the SA. Subsequently, the pixel values obtained were converted to millimeters using the scale established by the ZenCore (ZenCore v2.7, Zeiss, Oberkochen, Germany) software, allowing for an accurate comparison with manual methods.

2.4. Statistical Methods

Prior to analysis, data were normalized to ensure uniformity across different measurement scales. This included converting all measurements to a common unit (millimeters) and aligning data points according to predefined RFA durations and tissue types. Descriptive statistics (mean, median, standard deviation, and range) were calculated for each measurement method across both tissue types (liver and chicken breast) to summarize the central tendency and variability of AZ dimensions. The Shapiro–Wilk test was employed to assess the normality of data distributions.
Comparisons between measurement methods were performed using paired t-tests or Wilcoxon signed-rank tests, depending on the normality of the data. The Bland–Altman analysis was utilized to assess the agreement between methods, with limits of agreement defined as mean difference ± 1.96 standard deviations. We assessed the data by using correlation and Deming regression to explore the alignment between two measuring methods. The significance of differences in slopes and intercepts was tested to assess method-specific biases.
A p-value of <0.05 was considered statistically significant for all tests. Statistical analyses were performed using GraphPad Prism 9.0.

3. Results

We performed a total of 308 RFA runs, out of which 59.7% were performed in liver tissue and the remaining 40.3% were performed in chicken breast tissue (Table 1). During these runs, we acquired a total of 7275 images, with a similar percentage split (58.7% and 41.3%) between liver and chicken breast tissue, as with the RFA runs. The training, validation, and testing sets of images were split in an 8:1:1 ratio.
Thirty manual measurements of the ablation zone were performed at the end of the RFA run using a caliper. The average diameter (SD) was 16.3 mm (3.8 mm). Serial US manual measurements of the ablation zone in the liver tissue were made on US images, as previously described. A total of 398 measurements were performed, with a mean of 15.8 mm (5.9 mm), ranging from 1.83 to 29.2 mm. In the chicken breast tissue, 299 measurements were made, with a mean of 10.5 mm (4.2 mm), ranging from 2.2 to 19.4 mm.

3.1. Segmentation Performance of the AI Model

To evaluate the AI model’s performance, four pixel-based metrics were used. The AZ exhibited different segmentation performance for different tissue types (Table 2).
Notably, the AZ in chicken breast tissue appeared less hyperechoic in US images compared to liver tissue, as depicted in Figure 2. Conversely, the growth of the AZ around the RFA probe was shaped as an oval in chicken breast tissue, whereas, in liver tissue, the AZ assumed a more elliptical shape. As RFA progressed, increased image artifacts arose due to tissue heating and the forming of gas bubbles, thus creating a posterior shadowing effect, and complicating the assessment of the lower AZ contour. Figure 2 illustrates examples of these image artifacts.

3.2. Comparison of AI vs. US Manual Measurements

To begin our analysis, we compared the mean values of AI and US manual measurements in both liver and chicken breast tissue. The average diameter for AI in liver tissue was 15.6 mm (5.9 mm), while, for US, it was 15.8 mm (5.9 mm). This difference was not significant (t-test; p = 0.54). In chicken breast tissue, the mean predicted diameter for AI was 10.2 mm (4.1 mm), while the mean for US manual measurements was 10.5 mm (4.2 mm). This difference was also not significant (p = 0.48). Figure 3 presents a graphical comparison of the average values obtained from both methods in chicken and liver tissue.
Correlation coefficients were calculated between the diameters measured by AI and US. In bovine liver, the Pearson’s correlation coefficient was 0.965 (95% CI: 0.957–0.971; p < 0.001), with R2 = 0.931, while, in chicken breast tissue, it was 0.962 (95% CI: 0.953–0.970), with R2 = 0.925.
To test the alignment in measurements between the AI and US manual mode, we used a Bland–Altman plot (Figure 4). The average difference between the two methods in liver tissue was −0.259 mm (95% CI: −0.414 to −0.104), with the lower boundary at −3.339 mm and the upper boundary at 2.820 mm. In chicken breast tissue, the average difference between the two methods was −0.243 mm (95% CI: −0.374 to −0.112 mm), with the lower boundary at −2.771 mm and the upper boundary at 1.777 mm.
Finally, we utilized Deming regression to compare the AI-predicted diameter of the ablation zone in bovine liver against the US-measured diameter as the ground truth. The analysis yielded the regression equation, AI diameter = 0.9953 × US diameter + 0.3332. The equation slope of 0.9953 indicates a near one-to-one relationship between the AI predictions and the US measurements, thus demonstrating the AI model’s accuracy (with a small underestimation) in assessing the diameter of the ablation zone. The Y-intercept was found to be 0.3332, suggesting a small systematic bias in the AI predictions.
The same analysis was performed for chicken breast tissue and the resultant regression formula, AI diameter = 1.014 × US diameter + 0.09540, signifies a slope of 1.014. Here, the AI model performed with a slight overestimation. The Y-intercept, determined to be 0.09540, hints at a minor systematic bias in the AI predictions.
Both final analyses underline the precision of the AZ diameter prediction by our AI algorithm (Figure 4).

4. Discussion

This study explores the potential of automated AZ measurements in US images through the application of a Mask2Former model, aiming to enhance both laboratory processes and the outcomes of RFA research. The model underwent supervised training to identify and delineate AZs in US images captured at 15 s intervals during bipolar RFA activations, focusing separately on chicken and liver samples. A custom-built test stand facilitated straightforward and consistent laboratory procedures. AZ labelling within these images was conducted by a single expert to train the Mask2Former model. Short-axis measurements were conducted to compare the performance of the AI-predicted diameter in comparison to the manually measured US diameter.
Machine learning (ML) has found extensive applications within the medical sector, notably in image analysis and diagnostics [34,35,36]. Different AI models trained via supervised learning are frequently employed for their capabilities in US image tasks [14,25,36]. Nonetheless, a significant challenge is the limited amount of available training samples, potentially hindering the effectiveness of AI training [17,25]. To address this issue, data augmentation strategies were employed to enrich the dataset and prevent model overfitting, a practice supported by various studies. This step was essential in enhancing the model’s ability to accurately recognize and delineate the AZ across different tissue types [36,37].
The manual measurement, serving as the primary benchmark against which other methods are evaluated, is a staple in laboratory practices. It relies on the tissue color change upon coagulation, an indicator of histopathological damage from thermal exposure [38,39]. However, we decided against using it, primarily as it does not allow for obtaining a large number of dynamic measurements needed for accurate comparison and for the real-time tracking of the size of ablation zone, which is one of the major clinical requirements in practical use.
Additional factors influencing manual measurements include the specifics of tissue sectioning and inherent tissue characteristics, such as the presence of vessels or muscle fibers, which may distort the AZ. Accurate slicing at the RFA probe’s height is crucial, as the AZ typically forms spherically around it [7]. This requirement underscores the potential advantages of automated measurement techniques, which promise consistent and precise SA measurements in undisturbed tissue, mitigating the impact of human variance. AI-enhanced image segmentation in conjunction with US imaging emerges as a promising approach to address these limitations [5,25,33].
Ultrasound imaging is a cornerstone of diagnostic and therapeutic procedures, including image-guided thermal ablation, prized for its real-time imaging capabilities [5,12]. However, US’s grayscale imaging limits lesion clarity compared to CT or MRI, intensified by challenges in tissue visibility and the effects of hyperechoic structures and gas. Despite these limitations, the US remains preferable for routine laboratory use over CT and MRI, given the latter’s extensive technical and material requirements [12,13].
The efficacy and reliability of the SA measurement via US have been substantiated across studies, though outcomes can vary based on equipment brand, scanner orientation, and distance from the transducer to the target [40,41,42,43]. The Clarius US scanner, a modern handheld, wireless device, demonstrates no significant compromise in image quality, maintaining measurement accuracy on par with traditional stationary systems [44].
The main goal of the present study was to compare the AI-predicted diameter against the ground truth—in our study, the manually measured US diameter. AI-derived SA measurements were extracted from the largest horizontal span across the predicted mask within the US image. This approach was grounded on the assumption that the AZ uniformly encircles the RFA probe’s center, although deviations caused by nearby vascular structures or other anomalies could challenge this assumption [8]. Given that AI utilized the same image set as the US method, it was subject to similar challenges of image artifacts, which could lead to overestimations of the SA in comparison to manual measurements. This is consistent with the broader literature noting the complexities of interpreting US images with AI due to image artifacts [16,45]. It is important to emphasize that overestimation of the SA size regularly occurs with US imaging; therefore, all methods that rely on US will have a certain amount of overestimation, which needs to be taken into account, especially in future research in real-life clinical settings.
When examining AI’s performance relative to manual measurements, it was noted that AI more accurately delineated the AZ in chicken breast tissue compared to liver tissue (Figure 2). This discrepancy may be attributed to the differential impact of US image artifacts on tissue visibility, with chicken breast tissue presenting fewer disturbances. This observation is supported by comparisons with existing literature, where AI’s efficacy in AZ detection showcases the potential for more accurate assessments in conditions with less ultrasonic interference [46]. The challenges associated with ultrasound image artifacts, such as shadowing and speckle noise, were minimized by allowing the model to learn the representation of such artifacts during the training process. As the model was trained with the expert’s labels, it behaves in a similar fashion regarding the AZ.
The results gained with Bland–Altman plots and Deming´s regression demonstrate an excellent alignment of the AI-predicted diameter with manual measurements, with negligible systematic bias and a very slight propensity for underestimation. However, it appears that this alignment is more variable in small AZs, notably, AZ < 10 mm in liver tissue and AZ < 5 mm in chicken breast tissue (Figure 4). Conversely, AI demonstrated a higher accuracy for larger SAs, especially in chicken breast tissue, likely due to the high acoustic impedance of chicken breast tissue affecting AZ visibility [47,48,49]. A smaller AZ corresponds with the start of RFA, where the acoustic properties of tissue are prone to the huge variability due to uneven heating of the tissue or steam formation next to the RFA electrodes which may at least partially explain this phenomenon. For future research into the clinical use of our method, it is important to consider that AI-predicted measurements of AZ at the beginning of the ablation may be less reliable.
It is important to mention several limitations of the present study, which primarily stem from its in vitro nature, affecting the generalizability of the findings to clinical settings. Other potential limitations include the controlled environment not fully replicating the complexity of live tissue characteristics and responses to RFA, and the use of a specific set of tissues (chicken breast and liver) which may not represent the diversity found in human pathology. Another limitation is the fact that only homogenous imaging conditions were present, without movement artifacts or blood flow. Additionally, the study’s focus on a single AI model (Mask2Former) limits exploration of alternative or potentially more effective AI approaches. Yet, the clear potential of our approach proves its applicability in laboratory conditions but warrants further investigations regarding the clinical applicability.
The evaluation of clinical applicability is, of course, a further necessary step in the exploration of this field, particularly in terms of assessing its clinical potential for both percutaneous ultrasound-guided (i.e., in the liver) and endoscopic ultrasound-guided (i.e., in the lung and pancreas) RFA procedures. Based on the results of the present study, we strongly believe that this approach offers clear potential for the reliable real-time tracking of ablation zone size in the clinical setting.

5. Conclusions

In conclusion, our study demonstrated the potential of using a Mask2Former model for the automatic delineation and measurement of ablation zones in ultrasound images, offering a promising tool for enhancing the accuracy and efficiency of radiofrequency ablation research. While the AI model showed notable precision compared to manual methods, variations across tissue types and the variability with smaller ablation zones highlight areas for further development. Future work should focus on refining the AI model’s adaptability and accuracy to fully leverage AI in clinical RFA applications. This will improve patient outcomes by optimizing therapeutic interventions and transferring and using the model for the real-time monitoring of US-guided RFA ablation in clinical settings.

Author Contributions

Conceptualization, C.Z. and W.L.; methodology, C.Z.; software, A.M. and Y.D.; validation, C.Z., Y.D., W.L., M.D.E. and N.S.; formal analysis, C.Z., A.M., Y.D., M.D.E., N.S. and W.L.; investigation, C.Z., A.M., Y.D. and W.L.; resources, W.L. and M.D.E.; data curation, C.Z., A.M., Y.D., N.S. and W.L.; writing—original draft preparation, C.Z., A.M., N.S., M.D.E. and W.L.; visualization, C.Z., A.M., N.S. and W.L. supervision, M.D.E., N.S. and W.L.; project administration, M.D.E. and W.L.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author due to the proprietary algorithms created.

Conflicts of Interest

All authors are employees at Erbe Elektromedizin GmbH.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA A Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  2. Brace, C.L. Radiofrequency and microwave ablation of the liver, lung, kidney, and bone: What are the differences? Curr. Probl. Diagn. Radiol. 2009, 38, 135–143. [Google Scholar] [CrossRef] [PubMed]
  3. Ahmed, M. Image-Guided Tumor Ablation: Standardization of Terminology and Reporting Criteria—A 10-Year Update: Supplement to the Consensus Document. J. Vasc. Interv. Radiol. 2014, 25, 1706–1708. [Google Scholar] [CrossRef] [PubMed]
  4. Kaur, J.; Mohanti, B.K. Transition from curative to palliative care in cancer. Indian J. Palliat. Care 2011, 17, 1–5. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, B.-D.; Ye, X.; Fan, W.-J.; Li, X.-G.; Feng, W.-J.; Lu, Q.; Mao, Y.; Lin, Z.-Y.; Li, L.; Zhuang, Y.-P.; et al. Expert consensus on image-guided radiofrequency ablation of pulmonary tumors: 2018 edition. Thorac. Cancer 2018, 9, 1194–1208. [Google Scholar] [CrossRef] [PubMed]
  6. Chi, J.; Ding, M.; Shi, Y.; Wang, T.; Cui, D.; Tang, X.; Li, P.; Zhai, B. Comparison study of computed tomography-guided radiofrequency and microwave ablation for pulmonary tumors: A retrospective, case-controlled observational study. Thorac. Cancer 2018, 9, 1241–1248. [Google Scholar] [CrossRef] [PubMed]
  7. Nahum Goldberg, S.; Scott Gazelle, G.; Dawson, S.L.; Rittman, W.J.; Mueller, P.R.; Rosenthal, D.I. Tissue ablation with radiofrequency: Effect of probe size, gauge, duration, and temperature on lesion volume. Acad. Radiol. 1995, 2, 399–404. [Google Scholar] [CrossRef] [PubMed]
  8. Rempp, H.; Mezger, D.; Voigtlaender, M.; Scharpf, M.; Hoffmann, R.; Pereira, P.L.; Enderle, M.D.; Claussen, C.D.; Clasen, S. A Comparison of Internally Water-perfused and Cryogenically Cooled Monopolar and Bipolar Radiofrequency Applicators in Ex Vivo Liver Samples. Acad. Radiol. 2014, 21, 661–666. [Google Scholar] [CrossRef] [PubMed]
  9. Goldberg, S. Radiofrequency tumor ablation: Principles and techniques. Eur. J. Ultrasound 2001, 13, 129–147. [Google Scholar] [CrossRef]
  10. Reinhardt, M.; Brandmaier, P.; Seider, D.; Kolesnik, M.; Jenniskens, S.; Sequeiros, R.B.; Eibisberger, M.; Voglreiter, P.; Flanagan, R.; Mariappan, P.; et al. A prospective development study of software-guided radio-frequency ablation of primary and secondary liver tumors: Clinical intervention modelling, planning and proof for ablation cancer treatment (ClinicIMPPACT). Contemp. Clin. Trials Commun. 2017, 8, 25–32. [Google Scholar] [CrossRef]
  11. Ziemlewicz, T.J.; Hinshaw, J.L.; Lubner, M.G.; Knott, E.A.; Willey, B.J.; Lee Fred, T., Jr.; Brace, C.L. Radiofrequency and microwave ablation in a porcine liver model: Non-contrast CT and ultrasound radiologic-pathologic correlation. Int. J. Hyperth. 2020, 37, 799–807. [Google Scholar] [CrossRef] [PubMed]
  12. Goldberg, S.N.; Gazelle, G.S.; Mueller, P.R. Thermal Ablation Therapy for Focal Malignancy. Am. J. Roentgenol. 2000, 174, 323–331. [Google Scholar] [CrossRef] [PubMed]
  13. McWilliams, J.P.; Lee, E.W.; Yamamoto, S.; Loh, C.T.; Kee, S.T. Image-guided tumor ablation: Emerging technologies and future directions. Semin. Intervent. Radiol. 2010, 27, 302–313. [Google Scholar] [CrossRef] [PubMed]
  14. Brattain, L.J.; Telfer, B.A.; Dhyani, M.; Grajo, J.R.; Samir, A.E. Machine learning for medical ultrasound: Status, methods, and future opportunities. Abdom. Radiol. 2018, 43, 786–799. [Google Scholar] [CrossRef] [PubMed]
  15. Ferrara, R.; Mansi, L. Paul Suetens (ed): Fundamentals of Medical Imaging (2nd edition). Eur. J. Nucl. Med. Mol. Imaging 2010, 38, 409. [Google Scholar] [CrossRef]
  16. Yuan, Z.; Puyol-Antón, E.; Jogeesvaran, H.; Reid, C.; Inusa, B.; King, A.P. Deep Learning for Automatic Spleen Length Measurement in Sickle Cell Disease Patients; Springer International Publishing: Cham, Switzerland, 2020. [Google Scholar]
  17. Huang, Q.; Zhang, F.; Li, X. Machine Learning in Ultrasound Computer-Aided Diagnostic Systems: A Survey. Biomed Res. Int. 2018, 2018, 5137904. [Google Scholar] [CrossRef] [PubMed]
  18. Ghorbani, A.; Ouyang, D.; Abid, A.; He, B.; Chen, J.H.; Harrington, R.A.; Liang, D.H.; Ashley, E.A.; Zou, J.Y. Deep learning interpretation of echocardiograms. NPJ Digit. Med. 2020, 3, 10. [Google Scholar] [CrossRef] [PubMed]
  19. Namburete, A.I.L.; Xie, W.; Noble, J.A. Robust Regression of Brain Maturation from 3D Fetal Neurosonography Using CRNs; Springer International Publishing: Cham, Switzerland, 2017. [Google Scholar]
  20. Jovanovic, P.; Salkic, N.N.; Zerem, E. Artificial neural network predicts the need for therapeutic ERCP in patients with suspected choledocholithiasis. Gastrointest. Endosc. 2014, 80, 260–268. [Google Scholar] [CrossRef] [PubMed]
  21. Chen, L.-P. Mehryar Mohri, Afshin Rostamizadeh, and Ameet Talwalkar: Foundations of machine learning, second edition. Stat. Pap. 2019, 60, 1793–1795. [Google Scholar] [CrossRef]
  22. Rajkomar, A.; Dean, J.; Kohane, I. Machine Learning in Medicine. N. Engl. J. Med. 2019, 380, 1347–1358. [Google Scholar] [CrossRef] [PubMed]
  23. Chiao, J.-Y.; Chen, K.-Y.; Liao, K.Y.-K.; Hsieh, P.-H.; Zhang, G.; Huang, T.-C. Detection and classification the breast tumors using mask R-CNN on sonograms. Medicine 2019, 98, e15200. [Google Scholar] [CrossRef] [PubMed]
  24. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 25 December 2017. [Google Scholar]
  25. Liu, S.; Wang, Y.; Yang, X.; Lei, B.; Liu, L.; Li, S.X.; Ni, D.; Wang, T. Deep Learning in Medical Ultrasound Analysis: A Review. Engineering 2019, 5, 261–275. [Google Scholar] [CrossRef]
  26. Liu, J.; Li, P. A Mask R-CNN Model with Improved Region Proposal Network for Medical Ultrasound Image; Springer International Publishing: Cham, Switzerland, 2018. [Google Scholar]
  27. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-attention Mask Transformer for Universal Image Segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  28. Rathke, H.; Hamm, B.; Güttler, F.; Rathke, J.; Rump, J.; Teichgräber, U.; de Bucourt, M. Comparison of four radiofrequency ablation systems at two target volumes in an ex vivo bovine liver model. Diagn. Interv. Radiol. 2014, 20, 251–258. [Google Scholar] [CrossRef] [PubMed]
  29. Tanaka, T.; Penzkofer, T.; Isfort, P.; Bruners, P.; Disselhorst-Klug, C.; Junker, E.; Kichikawa, K.; Schmitz-Rode, T.; Mahnken, A.H. Direct Current Combined with Bipolar Radiofrequency Ablation: An Ex Vivo Feasibility Study. CardioVascular Interv. Radiol. 2010, 34, 631–636. [Google Scholar] [CrossRef] [PubMed]
  30. Yuan, Y.; Hou, S.; Wu, X.; Wang, Y.; Sun, Y.; Yang, Z.; Yin, S.; Zhang, F. Application of deep-learning to the automatic segmentation and classification of lateral lymph nodes on ultrasound images of papillary thyroid carcinoma. Asian J. Surg. 2024. [Google Scholar] [CrossRef] [PubMed]
  31. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  32. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Computer Vision—ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 740–755. ISBN 978-3-319-10601-4. [Google Scholar]
  33. El Naqa, I.; Murphy, M.J. What Is Machine Learning? Springer International Publishing: Cham, Switzerland, 2015. [Google Scholar]
  34. Cao, Z.; Duan, L.; Yang, G.; Yue, T.; Chen, Q. An experimental study on breast lesion detection and classification from ultrasound images using deep learning architectures. BMC Med Imaging 2019, 19, 51. [Google Scholar] [CrossRef] [PubMed]
  35. Li, S.; Wei, J.; Chan, H.-P.; Helvie, M.A.; Roubidoux, M.A.; Lu, Y.; Zhou, C.; Hadjiiski, L.M.; Samala, R.K. Computer-aided assessment of breast density: Comparison of supervised deep learning and feature-based statistical learning. Phys. Med. Biol. 2018, 63, 25005. [Google Scholar] [CrossRef] [PubMed]
  36. Lei, Y.; He, X.; Yao, J.; Wang, T.; Wang, L.; Li, W.; Curran, W.J.; Liu, T.; Xu, D.; Yang, X. Breast tumor segmentation in 3D automatic breast ultrasound using Mask scoring R-CNN. Med. Phys. 2020, 48, 204–214. [Google Scholar] [CrossRef] [PubMed]
  37. Zhao, Z.-Q.; Zheng, P.; Xu, S.-T.; Wu, X. Object Detection with Deep Learning: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed]
  38. McGahan, J.P.; Brock, J.M.; Tesluk, H.; Gu, W.-Z.; Schneider, P.; Browning, P.D. Hepatic Ablation with Use of Radio-Frequency Electrocautery in the Animal Model. J. Vasc. Interv. Radiol. 1992, 3, 291–297. [Google Scholar] [CrossRef] [PubMed]
  39. Rossi, S.; Fornati, F.; Pathies, C.; Buscarini, L. Thermal Lesions Induced by 480 KHz Localized Current Field in Guinea Pig and Pig Liver. Tumori J. 1990, 76, 54–57. [Google Scholar] [CrossRef]
  40. Gamba, J.L.; Bowie, J.D.; Dodson, W.C.; Hedlund, L.W. Accuracy of Ultrasound in Fetal Femur Length Determination Ultrasound Phantom Study. Investig. Radiol. 1985, 20, 316–323. [Google Scholar] [CrossRef] [PubMed]
  41. Kwah, L.K.; Pinto, R.Z.; Diong, J.; Herbert, R.D. Reliability and validity of ultrasound measurements of muscle fascicle length and pennation in humans: A systematic review. J. Appl. Physiol. 2013, 114, 761–769. [Google Scholar] [CrossRef] [PubMed]
  42. Tanaka, K.; Carlier, S.G.; Mintz, G.S.; Sano, K.; Liu, X.; Fujii, K.; de Ribamar Costa, J.; Lui, J.; Moses, J.W.; Stone, G.W.; et al. The accuracy of length measurements using different intravascular ultrasound motorized transducer pullback systems. Int. J. Cardiovasc. Imaging 2007, 23, 733–738. [Google Scholar] [CrossRef] [PubMed]
  43. Scorza, A.; Conforto, S.; D’Anna, C.; Sciuto, S.A. A Comparative Study on the Influence of Probe Placement on Quality Assurance Measurements in B-mode Ultrasound by Means of Ultrasound Phantoms. Open Biomed. Eng. J. 2015, 9, 164–178. [Google Scholar] [CrossRef] [PubMed]
  44. Zardi, E.M.; Franceschetti, E.; Giorgi, C.; Palumbo, A.; Franceschi, F. Accuracy and performance of a new handheld ultrasound machine with wireless system. Sci. Rep. 2019, 9, 14599. [Google Scholar] [CrossRef] [PubMed]
  45. Chen, X.; He, M.; Dan, T.; Wang, N.; Lin, M.; Zhang, L.; Xian, J.; Cai, H.; Xie, H. Automatic Measurements of Fetal Lateral Ventricles in 2D Ultrasound Images Using Deep Learning. Front. Neurol. 2020, 11, 526. [Google Scholar] [CrossRef] [PubMed]
  46. Wei, Z.; Zhang, B.; Liu, P. Object Dimension Measurement Based on Mask R-CNN; Springer International Publishing: Cham, Switzerland, 2019. [Google Scholar]
  47. Shishitani, T.; Matsuzawa, R.; Yoshizawa, S.; Umemura, S. Changes in backscatter of liver tissue due to thermal coagulation induced by focused ultrasound. J. Acoust. Soc. Am. 2013, 134, 1724–1730. [Google Scholar] [CrossRef] [PubMed]
  48. Chan, V.; Perlas, A. Basics of Ultrasound Imaging; Springer: New York, NY, USA, 2010. [Google Scholar]
  49. Shishitani, T.; Yoshizawa, S.; Umemura, S. Acoustic Impedance Evaluation of High-Intensity-Focused-Ultrasound Exposed Chicken Breast Muscle Using Ultrasonic Microscopy. Jpn. J. Appl. Phys. 2010, 49, 07HF04. [Google Scholar] [CrossRef]
Figure 1. Self-designed test stand with liver tissue. The RFA probe (1) was horizontally introduced in the tissue which was placed in the tissue cup (3). The US transducer (2) was placed horizontally and perpendicular to the RFA probe at the location of the separator, as shown in the schematic on the right. For the adjustment the US transducer could be moved in three dimensions indicated by the white arrows.
Figure 1. Self-designed test stand with liver tissue. The RFA probe (1) was horizontally introduced in the tissue which was placed in the tissue cup (3). The US transducer (2) was placed horizontally and perpendicular to the RFA probe at the location of the separator, as shown in the schematic on the right. For the adjustment the US transducer could be moved in three dimensions indicated by the white arrows.
Cancers 16 01700 g001
Figure 2. Representation of AZ in US images of liver (first row) and chicken tissue (second row) and the labelled mask (green) and predicted mask (orange). The AZ is less hyperechoic represented in chicken breast tissue as in liver tissue. Larger image artifacts can be observed underneath the AZ in liver tissue with ongoing RFA, making the assessment of the under AZ contour difficult. An acoustic shadow can be observed in the case of chicken breast.
Figure 2. Representation of AZ in US images of liver (first row) and chicken tissue (second row) and the labelled mask (green) and predicted mask (orange). The AZ is less hyperechoic represented in chicken breast tissue as in liver tissue. Larger image artifacts can be observed underneath the AZ in liver tissue with ongoing RFA, making the assessment of the under AZ contour difficult. An acoustic shadow can be observed in the case of chicken breast.
Cancers 16 01700 g002
Figure 3. Box-plot diagrams comparing the mean values of AI predicted and US manual measured diameter of RFA ablation zones in bovine liver tissue (left) and chicken breast tissue (right).
Figure 3. Box-plot diagrams comparing the mean values of AI predicted and US manual measured diameter of RFA ablation zones in bovine liver tissue (left) and chicken breast tissue (right).
Cancers 16 01700 g003
Figure 4. Bland - Altman plots (left) for chicken breast tissue (red) and bovine liver (blue) depicting the reliability of AI predicted diameter in comparison with US manual measurement as ground truth. Deming regression plots (right) for chicken breast tissue (red) and bovine liver (blue), demonstrating the reliability of AI−predicted diameter in comparison with US manual measurement as ground truth.
Figure 4. Bland - Altman plots (left) for chicken breast tissue (red) and bovine liver (blue) depicting the reliability of AI predicted diameter in comparison with US manual measurement as ground truth. Deming regression plots (right) for chicken breast tissue (red) and bovine liver (blue), demonstrating the reliability of AI−predicted diameter in comparison with US manual measurement as ground truth.
Cancers 16 01700 g004
Table 1. Baseline characteristics of number of RFA and sample split for AI training.
Table 1. Baseline characteristics of number of RFA and sample split for AI training.
Total Number of RFA RunsTraining SetValidation SetTest SetTotal Number of US Images
Liver18414720174276
Chicken1249815112999
Table 2. Evaluation metrics of the trained Mask2Former model according to the tissue type. Higher metrics were scored for chicken breast tissue. The amount of test images used for each tissue type is given by n.
Table 2. Evaluation metrics of the trained Mask2Former model according to the tissue type. Higher metrics were scored for chicken breast tissue. The amount of test images used for each tissue type is given by n.
Tissue TypenAccuracy [%]Sensitivity [%]Specificity [%]F1-Score [%]
Liver39898.588.899.389.7
Chicken breast29999.491.999.792.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zimmermann, C.; Michelmann, A.; Daniel, Y.; Enderle, M.D.; Salkic, N.; Linzenbold, W. Application of Deep Learning for Real-Time Ablation Zone Measurement in Ultrasound Imaging. Cancers 2024, 16, 1700. https://0-doi-org.brum.beds.ac.uk/10.3390/cancers16091700

AMA Style

Zimmermann C, Michelmann A, Daniel Y, Enderle MD, Salkic N, Linzenbold W. Application of Deep Learning for Real-Time Ablation Zone Measurement in Ultrasound Imaging. Cancers. 2024; 16(9):1700. https://0-doi-org.brum.beds.ac.uk/10.3390/cancers16091700

Chicago/Turabian Style

Zimmermann, Corinna, Adrian Michelmann, Yannick Daniel, Markus D. Enderle, Nermin Salkic, and Walter Linzenbold. 2024. "Application of Deep Learning for Real-Time Ablation Zone Measurement in Ultrasound Imaging" Cancers 16, no. 9: 1700. https://0-doi-org.brum.beds.ac.uk/10.3390/cancers16091700

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop