PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
SPIE publishes accepted journal articles as soon as they are approved for publication. Journal issues are considered In Progress until all articles for an issue have been published. Articles published ahead of the completed issue are fully citable.
Our study investigates the potential benefits of incorporating prior anatomical knowledge into a deep learning (DL) method designed for the automated segmentation of lung lobes in chest CT scans.
Approach
We introduce an automated DL-based approach that leverages anatomical information from the lung’s vascular system to guide and enhance the segmentation process. This involves utilizing a lung vessel connectivity (LVC) map, which encodes relevant lung vessel anatomical data. Our study explores the performance of three different neural network architectures within the nnU-Net framework: a standalone U-Net, a multitasking U-Net, and a cascade U-Net.
Results
Experimental findings suggest that the inclusion of LVC information in the DL model can lead to improved segmentation accuracy, particularly, in the challenging boundary regions of expiration chest CT volumes. Furthermore, our study demonstrates the potential for LVC to enhance the model’s generalization capabilities. Finally, the method’s robustness is evaluated through the segmentation of lung lobes in 10 cases of COVID-19, demonstrating its applicability in the presence of pulmonary diseases.
Conclusions
Incorporating prior anatomical information, such as LVC, into the DL model shows promise for enhancing segmentation performance, particularly in the boundary regions. However, the extent of this improvement has limitations, prompting further exploration of its practical applicability.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Deep learning is the standard for medical image segmentation. However, it may encounter difficulties when the training set is small. Also, it may generate anatomically aberrant segmentations. Anatomical knowledge can be potentially useful as a constraint in deep learning segmentation methods. We propose a loss function based on projected pooling to introduce soft topological constraints. Our main application is the segmentation of the red nucleus from quantitative susceptibility mapping (QSM) which is of interest in parkinsonian syndromes.
Approach
This new loss function introduces soft constraints on the topology by magnifying small parts of the structure to segment to avoid that they are discarded in the segmentation process. To that purpose, we use projection of the structure onto the three planes and then use a series of MaxPooling operations with increasing kernel sizes. These operations are performed both for the ground truth and the prediction and the difference is computed to obtain the loss function. As a result, it can reduce topological errors as well as defects in the structure boundary. The approach is easy to implement and computationally efficient.
Results
When applied to the segmentation of the red nucleus from QSM data, the approach led to a very high accuracy (Dice 89.9%) and no topological errors. Moreover, the proposed loss function improved the Dice accuracy over the baseline when the training set was small. We also studied three tasks from the medical segmentation decathlon challenge (MSD) (heart, spleen, and hippocampus). For the MSD tasks, the Dice accuracies were similar for both approaches but the topological errors were reduced.
Conclusions
We propose an effective method to automatically segment the red nucleus which is based on a new loss for introducing topology constraints in deep learning segmentation.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Monitoring radiation dose and time parameters during radiological interventions is crucial, especially in neurointerventional procedures, such as aneurysm treatment with embolization coils. The algorithm presented detects the presence of these embolization coils in medical images. It establishes a bounding box as a reference for automated collimation, with the primary objective being to enhance the efficiency and safety of neurointerventional procedures by actively optimizing image quality while minimizing patient dose.
Methods
Two distinct methodologies are evaluated in our study. The first involves deep learning, employing the Faster R-CNN model with a ResNet-50 FPN as a backbone and a RetinaNet model. The second method utilizes a classical blob detection approach, serving as a benchmark for comparison.
Results
We performed a fivefold cross-validation, and our top-performing model achieved mean mAP@75 of 0.84 across all folds on validation data and mean mAP@75 of 0.94 on independent test data. Since we use an upscaled bounding box, achieving 100% overlap between ground truth and prediction is not necessary. To highlight the real-world applications of our algorithm, we conducted a simulation featuring a coil constructed from an alloy wire, effectively showcasing the implementation of automatic collimation. This resulted in a notable reduction in the dose area product, signifying the reduction of stochastic risks for both patients and medical staff by minimizing scatter radiation. Additionally, our algorithm assists in avoiding extreme brightness or darkness in X-ray angiography images during narrow collimation, ultimately streamlining the collimation process for physicians.
Conclusion
To our knowledge, this marks the initial attempt at an approach successfully detecting embolization coils, showcasing the extended applications of integrating detection results into the X-ray angiography system. The method we present has the potential for broader application, allowing its extension to detect other medical objects utilized in interventional procedures.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Medical imaging-based machine learning (ML) for computer-aided diagnosis of in vivo lesions consists of two basic components or modules of (i) feature extraction from non-invasively acquired medical images and (ii) feature classification for prediction of malignancy of lesions detected or localized in the medical images. This study investigates their individual performances for diagnosis of low-dose computed tomography (CT) screening-detected lesions of pulmonary nodules and colorectal polyps.
Approach
Three feature extraction methods were investigated. One uses the mathematical descriptor of gray-level co-occurrence image texture measure to extract the Haralick image texture features (HFs). One uses the convolutional neural network (CNN) architecture to extract deep learning (DL) image abstractive features (DFs). The third one uses the interactions between lesion tissues and X-ray energy of CT to extract tissue-energy specific characteristic features (TFs). All the above three categories of extracted features were classified by the random forest (RF) classifier with comparison to the DL-CNN method, which reads the images, extracts the DFs, and classifies the DFs in an end-to-end manner. The ML diagnosis of lesions or prediction of lesion malignancy was measured by the area under the receiver operating characteristic curve (AUC). Three lesion image datasets were used. The lesions’ tissue pathological reports were used as the learning labels.
Results
Experiments on the three datasets produced AUC values of 0.724 to 0.878 for the HFs, 0.652 to 0.965 for the DFs, and 0.985 to 0.996 for the TFs, compared to the DL-CNN of 0.694 to 0.964. These experimental outcomes indicate that the RF classifier performed comparably to the DL-CNN classification module and the extraction of tissue-energy specific characteristic features dramatically improved AUC value.
Conclusions
The feature extraction module is more important than the feature classification module. Extraction of tissue-energy specific characteristic features is more important than extraction of image abstractive and characteristic features.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Lung cancer is the second most common cancer and the leading cause of cancer death globally. Low dose computed tomography (LDCT) is the recommended imaging screening tool for the early detection of lung cancer. A fully automated computer-aided detection method for LDCT will greatly improve the existing clinical workflow. Most of the existing methods for lung detection are designed for high-dose CTs (HDCTs), and those methods cannot be directly applied to LDCTs due to domain shifts and inferior quality of LDCT images. In this work, we describe a semi-automated transfer learning-based approach for the early detection of lung nodules using LDCTs.
Approach
In this work, we developed an algorithm based on the object detection model, you only look once (YOLO) to detect lung nodules. The YOLO model was first trained on CTs, and the pre-trained weights were used as initial weights during the retraining of the model on LDCTs using a medical-to-medical transfer learning approach. The dataset for this study was from a screening trial consisting of LDCTs acquired from 50 biopsy-confirmed lung cancer patients obtained over 3 consecutive years (T1, T2, and T3). About 60 lung cancer patients’ HDCTs were obtained from a public dataset. The developed model was evaluated using a hold-out test set comprising 15 patient cases (93 slices with cancerous nodules) using precision, specificity, recall, and F1-score. The evaluation metrics were reported patient-wise on a per-year basis and averaged for 3 years. For comparative analysis, the proposed detection model was trained using pre-trained weights from the COCO dataset as the initial weights. A paired t-test and chi-squared test with an alpha value of 0.05 were used for statistical significance testing.
Results
The results were reported by comparing the proposed model developed using HDCT pre-trained weights with COCO pre-trained weights. The former approach versus the latter approach obtained a precision of 0.982 versus 0.93 in detecting cancerous nodules, specificity of 0.923 versus 0.849 in identifying slices with no cancerous nodules, recall of 0.87 versus 0.886, and F1-score of 0.924 versus 0.903. As the nodule progressed, the former approach achieved a precision of 1, specificity of 0.92, and sensitivity of 0.930. The statistical analysis performed in the comparative study resulted in a p-value of 0.0054 for precision and a p-value of 0.00034 for specificity.
Conclusions
In this study, a semi-automated method was developed to detect lung nodules in LDCTs using HDCT pre-trained weights as the initial weights and retraining the model. Further, the results were compared by replacing HDCT pre-trained weights in the above approach with COCO pre-trained weights. The proposed method may identify early lung nodules during the screening program, reduce overdiagnosis and follow-ups due to misdiagnosis in LDCTs, start treatment options in the affected patients, and lower the mortality rate.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
TOPICS: Image segmentation, Arteries, Education and training, Magnetic resonance imaging, Data modeling, Independent component analysis, Simulation of CCA and DLA aggregates, 3D image processing, Scanners, Performance modeling
Atherosclerosis of the carotid artery is a major risk factor for stroke. Quantitative assessment of the carotid vessel wall can be based on cross-sections of three-dimensional (3D) black-blood magnetic resonance imaging (MRI). To increase reproducibility, a reliable automatic segmentation in these cross-sections is essential.
Approach
We propose an automatic segmentation of the carotid artery in cross-sections perpendicular to the centerline to make the segmentation invariant to the image plane orientation and allow a correct assessment of the vessel wall thickness (VWT). We trained a residual U-Net on eight sparsely sampled cross-sections per carotid artery and evaluated if the model can segment areas that are not represented in the training data. We used 218 MRI datasets of 121 subjects that show hypertension and plaque in the ICA or CCA measuring ≥1.5mm in ultrasound.
Results
The model achieves a high mean Dice coefficient of 0.948/0.859 for the vessel’s lumen/wall, a low mean Hausdorff distance of 0.417/0.660mm, and a low mean average contour distance of 0.094/0.119mm on the test set. The model reaches similar results for regions of the carotid artery that are not incorporated in the training set and on MRI of young, healthy subjects. The model also achieves a low median Hausdorff distance of 0.437/0.552mm on the 2021 Carotid Artery Vessel Wall Segmentation Challenge test set.
Conclusions
The proposed method can reduce the effort for carotid artery vessel wall assessment. Together with human supervision, it can be used for clinical applications, as it allows a reliable measurement of the VWT for different patient demographics and MRI acquisition settings.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Image Perception, Observer Performance, and Technology Assessment
TOPICS: Signal detection, Computer aided detection, Digital breast tomosynthesis, 3D modeling, 3D image processing, Eye, Visualization, Education and training, Deep learning, Picture Archiving and Communication System
Radiologists are tasked with visually scrutinizing large amounts of data produced by 3D volumetric imaging modalities. Small signals can go unnoticed during the 3D search because they are hard to detect in the visual periphery. Recent advances in machine learning and computer vision have led to effective computer-aided detection (CADe) support systems with the potential to mitigate perceptual errors.
Approach
Sixteen nonexpert observers searched through digital breast tomosynthesis (DBT) phantoms and single cross-sectional slices of the DBT phantoms. The 3D/2D searches occurred with and without a convolutional neural network (CNN)-based CADe support system. The model provided observers with bounding boxes superimposed on the image stimuli while they looked for a small microcalcification signal and a large mass signal. Eye gaze positions were recorded and correlated with changes in the area under the ROC curve (AUC).
Results
The CNN-CADe improved the 3D search for the small microcalcification signal (ΔAUC=0.098, p=0.0002) and the 2D search for the large mass signal (ΔAUC=0.076, p=0.002). The CNN-CADe benefit in 3D for the small signal was markedly greater than in 2D (ΔΔAUC=0.066, p=0.035). Analysis of individual differences suggests that those who explored the least with eye movements benefited the most from the CNN-CADe (r=−0.528, p=0.036). However, for the large signal, the 2D benefit was not significantly greater than the 3D benefit (ΔΔAUC=0.033, p=0.133).
Conclusion
The CNN-CADe brings unique performance benefits to the 3D (versus 2D) search of small signals by reducing errors caused by the underexploration of the volumetric data.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Biomedical Applications in Molecular, Structural, and Functional Imaging
Our objective was to train machine-learning algorithms on hyperpolarized He3 magnetic resonance imaging (MRI) datasets to generate models of accelerated lung function decline in participants with and without chronic-obstructive-pulmonary-disease. We hypothesized that hyperpolarized gas MRI ventilation, machine-learning, and multivariate modeling could be combined to predict clinically-relevant changes in forced expiratory volume in 1 s (FEV1) across 3 years.
Approach
Hyperpolarized He3 MRI was acquired using a coronal Cartesian fast gradient recalled echo sequence with a partial echo and segmented using a k-means clustering algorithm. A maximum entropy mask was used to generate a region-of-interest for texture feature extraction using a custom-developed algorithm and the PyRadiomics platform. The principal component and Boruta analyses were used for feature selection. Ensemble-based and single machine-learning classifiers were evaluated using area-under-the-receiver-operator-curve and sensitivity-specificity analysis.
Results
We evaluated 88 ex-smoker participants with 31±7 months follow-up data, 57 of whom (22 females/35 males, 70±9 years) had negligible changes in FEV1 and 31 participants (7 females/24 males, 68±9 years) with worsening FEV1≥60mL/year. In addition, 3/88 ex-smokers reported a change in smoking status. We generated machine-learning models to predict FEV1 decline using demographics, spirometry, and texture features, with the later yielding the highest classification accuracy of 81%. The combined model (trained on all available measurements) achieved the overall best classification accuracy of 82%; however, it was not significantly different from the model trained on MRI texture features alone.
Conclusion
For the first time, we have employed hyperpolarized He3 MRI ventilation texture features and machine-learning to identify ex-smokers with accelerated decline in FEV1 with 82% accuracy.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.