1.IntroductionIn recent years, the number of computed tomography (CT) examinations has increased significantly, resulting in a high workload for radiologists. The resulting time pressure increases the risk of reduced quality and safety of diagnoses and radiologist burnout.1 Therefore, artificial intelligence solutions have been developed to address this issue and provide quality diagnosis as a second opinion or pre-diagnosis. In this work, we deal with CT brain examinations and aim to support radiologists in finding relevant pathologies, such as hemorrhages, tumors, or ischemia. Various methods have been proposed to assist radiologists in this task. In particular, deep learning-based approaches have shown very good results in many applications and are now the technology of choice. Deep learning methods for automated anomaly detection can generally be divided into two types. The first type is designed to find specific pathologies. Therefore, a model is trained to detect or segment a particular type of abnormality, usually with a supervised or semi-supervised approach using annotated training data.2 Good detection performance of such methods has been demonstrated in the literature, but they are inherently limited to the specific types of findings for which they have been trained, and at best, unknown abnormalities are detected as false positives. The second type of method takes a fundamentally different approach. They do not make any assumptions and do not try to find a specific pathology but are designed to detect any deviation from “normal.” Therefore, training such models does not require reference segmentations of pathologies but is usually based on (unsupervised) learning a representation of normal anatomy. Thus, it potentially has fewer limitations, and a trained model should be less error-prone for unseen anomalies. Due to that advantage, we follow the latter approach for our method of general anomaly detection in brain CT images. Architectures used for this type of anomaly detection include generative adversarial networks (GANs), variational autoencoders (VAE), autoencoders (AE), and combinations of these methods.3–6 Reconstruction of a normal image is then generally done by mapping an input image with an encoding network to a latent space and reconstructing the output image with a decoding network. In the so-called restoration approaches, the encoding is not performed by a trained network but through iterative optimization with respect to the input in the latent space of the trained decoding network, such that the decoded (or restored) output image is most similar to the given input image. Baur et al.7 compared the different unsupervised approaches for magnetic resonance imaging (MRI) brain images and found that a VAE with restoration and the “f-AnoGAN,”8 originally proposed for optical coherence tomography data and combining a Wasserstein GAN9 with encoder training, performed best. In addition, diffusion models have recently been investigated for anomaly detection. Wyatt et al.10 obtained Dice values of around 0.38 on MRI images with brain tumors. Much research has been done on MRI images, but simply using these models as is does not solve the problem of detecting abnormalities in CT images of the brain. The imaging properties of CT imply a high contrast between the skull and the brain tissue with much more subtle gray versus white matter differences. Thus, the generated reconstructions must also reflect these properties while especially preserving visible details inside the brain with lower grayscale contrast. For brain CT data, Viana et al.11 utilized a 3D implementation of the f-AnoGAN to detect traumatic brain injuries. Toikkanen et al.12 developed a method to segment intracranial hemorrhages in brain CT images and combined a GAN with an encoder to reconstruct pseudo-normal versions of abnormal input images. Their reconstructions however do not correct structural changes of the brain, such as ventricle deformations, that might occur due to hemorrhages. They achieve Dice scores of around 0.7 while focusing on the segmentation of hemorrhages only. In a recent publication, Lee et al.13 described a combination of GAN and encoder training for CT emergency triage. They collected a large dataset of more than 34,000 patients for training and used iterative latent code and noise optimization during inference, showing very promising results. In summary, there is very little work on unsupervised detection of brain anomalies in CT. Autoencoders often tend to produce blurry results that lack detail. Diffusion models are expected to result in higher image quality than VAEs and require less training data than GAN approaches. On the other hand, reconstruction times are very long, and high image resolutions, which have been shown to be beneficial for anomaly detection,4 might be hard to obtain due to computational requirements. Furthermore, when using diffusion models to reconstruct normal versions of an image, the choice of noise distribution used is critical in determining what sizes of anomalies can be recovered (see Wyatt et al.10). The contribution of this work is that we propose an efficient unsupervised GAN-based method to localize different types of lesions in CT scans of the brain. We choose a GAN-based architecture over others because of the above-mentioned disadvantages. We aim to generate high-resolution healthy image reconstructions and want to ensure that a representation of normal anatomy is learned first so that our reconstructions will not contain abnormalities. In this work, we show that with the right choice of GAN training, we can achieve high-quality reconstruction results even on a moderate-sized training dataset. The paper is structured by first describing the GAN training and our two image reconstruction approaches, followed by the generation of anomaly segmentation maps and an evaluation of CT images with hemorrhages and tumors. We also compare our method with two non-GAN-based approaches. 2.Material and Methods2.1.DataWe use the publicly available challenge data for intracranial hemorrhage detection provided by the Radiological Society of North America (RSNA).14 Originally, this dataset contains more than 25,000 CT head examinations with slice-wise annotations for the type of hemorrhage present, collected at various sites in California, Philadelphia, and Brazil using scanners from different manufacturers. The resolution of the images in the axial plane ranges from 0.4 to 0.6 mm with a slice thickness of 5 mm. In addition to the annotations provided, two of our radiologists (I.G. and K.V.) with more than 10 years of experience in neurological imaging also annotated a subset of these data by marking slices without findings as “normal.” The RSNA annotations did not include such information but only hemorrhage annotations. We export more than 1000 normal 2D slices with a size of . We also select a subset of 30 images with hemorrhages for testing, where our radiologists have delineated the hemorrhages present. We restrict both subsets to mid-axial slices of the brain to limit the complexity of the problem in the first step. The number of slices that can be used from the RSNA dataset is reduced because of the limited annotation time and because we only use center slices of the brain. The image slices were grouped according to patients. Images from one patient are only used for either training or testing. For testing, we also include tumor data from the “Glioma Image Segmentation for Radiotherapy” study from the Cancer Imaging Archive15 and extract 23 slices that have been annotated by one of our radiologists. We therefore obtain reference anomaly masks for an independent dataset. Thus, in total, we used 1180 “normal” 2D axial images to train our GAN. In addition, we obtain a total of 53 segmented images with pathologies (30 with hemorrhages and 23 with tumors). Subsequently, we determined an operating point for the anomaly detection on a validation set of 27 axial images with segmented pathologies (15 scans with hemorrhages and 12 scans with tumors). Then, we tested our method on 26 test images (15 scans with hemorrhages and 11 scans with tumors). A summary of the data split for testing and validation is shown in Table 1. Table 1Composition of validation and test dataset with pathologies. A total of 53 images were manually segmented by two radiologists. The images show hemorrhages or tumors and were split into test and validation sets as shown.
2.2.Architecture and TrainingThe architecture of our network is schematically shown in Fig. 1. To ensure learning a representation of normal brain anatomy, we decided to combine a GAN architecture with an image encoding step (using an encoder network or an iterative optimization). Thus, we can later compare two different encoding approaches. As shown in Fig. 1, we have two subsequent training steps: First, the GAN is trained to generate normal CT slices from a latent space vector [see Fig. 1(a)], and second, the encoder is trained to map healthy input images to the latent space [see Fig. 1(b)]. During inference, an input CT is mapped with the encoder to the latent space and from there back to the image space with the generator. As an alternative that requires no training of an encoder network, we implement a restoration approach. Here, we iteratively optimize over the latent space to find an optimal input vector for the decoder so that a given image is reconstructed in the best possible way [see Fig. 1(c)]. We describe all steps in detail in Secs. 2.2.1–2.2.4. 2.2.1.GAN training: generating normal CT imagesThe first part of our method is to train a GAN to learn representations of normal CT images of the brain. Based on literature research, we use a network architecture after Liu et al.,16 which was proposed for non-medical images and has been shown to be capable of generating high-resolution images from even small training sets. Also, we verified in our own initial tests that this architecture leads to better results compared with a Wasserstein GAN9 architecture. The code base for the discriminator and generator can be found at https://github.com/odegeasslbc/FastGAN-pytorch/tree/main. We use the proposed architecture with noise injection, spectral normalization,17 and image augmentation (random mirroring, contrast changes, and translation). Only a small change was made to the architecture as we use smaller feature maps (four instead of eight) as input to the decoder that is part of the discriminator. In addition, we leave one generator upsampling block out as we only work with an image size of 512. The model is trained with a batch size of 8 for 100,000 steps, using the learning rate of and Adam optimizer. For our experiments, we used a training dataset consisting of 1180 brain CT slices. Our code is written in PyTorch, and we train on an NVIDIA GeForce RTX 2080 Ti GPU. In the first step, we alternatingly train the generator and the discriminator (see Liu et al.16 for model implementation details). This is done by minimizing the generator loss and the discriminator loss where denotes the distribution of real normal CT scans, are latent vectors with components uniformly distributed over the interval , and is an additional reconstruction loss that originally was introduced by Liu et al.16 to enforce the extraction of meaningful image features in the discriminator (for details, see Ref. 16). Here, we use the reconstruction loss as proposed in Ref. 18.After training, we can use the generator to obtain normal brain images from a random uniformly distributed latent vector . 2.2.2.Image reconstruction: encoder trainingTo reconstruct a normal representation of a test image, we need to encode the image by a latent vector , such that , i.e., the generated image is similar to the input image . Therefore, we train an encoder network to obtain from . For the encoder training [Fig. 1(c)], the parameters of the generator remain fixed, and the discriminator is not used. We train the encoder by minimizing the encoder loss with so-called loss ,19 mean squared error loss , and empirically chosen fixed weights , balancing both loss terms. We use the same training dataset of 1180 normal brain CT slices as for the GAN training. In addition, we augment the images during training, such that random rectangular parts of the image are erased, meaning that the image in this area is set to the background value with an empirically chosen probability of 0.25. This should prevent the encoder from learning the identity function. The network is trained for 20,000 steps. The encoder architecture consists of eight down-sampling blocks that each include 2D convolution, batch normalization, and gated linear unit activation layers. A schematic representation of the encoder architecture is shown in Fig. 2.2.2.3.Image restoration: optimization of zAs an alternative to training an encoder and using the trained network during inference, we compute a solution to the minimization problem with the loss ,19 mean squared error loss , and empirically chosen fixed weights , balancing both loss terms. We compute a numerical solution through iterative optimization of the objective function w.r.t. to the latent variable starting from the initial guess . We used Adam optimizer, and is clipped component-wise to the range at each step. The iteration is stopped after a fixed number of steps (8000), taking around 110 s. So, in this method, we do not train a network, but for each image, we iteratively compute our own individually optimized solution for inference.2.2.4.Obtaining anomaly mapsAfter training the networks, we are able to reconstruct a healthy version of a given input image using our iterative encoding method or the encoder, resulting in image pairs of the input image and a normal representation . We then obtain a difference map with a residual function , e.g., for detecting any change, which results in , or if we are trying to detect certain abnormalities that appear brighter than normal tissue, such as hemorrhages, we could choose , which results in . Finally, we use a threshold to generate a binary anomaly mask. The mask is refined by applying a morphological closing operation and using a network ensemble. This means that we train our GAN three times with the same settings and thus also get three binary maps after reconstruction, which are then combined by a majority vote. By doing so, we can neglect regions that are falsely segmented as anomalies because of missing details in a single reconstruction. 2.3.EvaluationWe evaluate our method with two datasets. The first set includes 30 images with hemorrhages, whereas the second additionally contains 23 images with brain tumors. We show receiver operating characteristics (ROC) curves plotting pixel-wise false-positive rate against true-positive rate for different thresholds (see Fig. 7), and we report boxplots showing Dice overlap (Fig. 6). Other segmentation metrics such as the Haustorff distance would not provide meaningful results in our scenario, as there are often very small false-positive areas in the generated anomaly masks. Furthermore, we evaluate the realism of the images generated by our GAN through a visual Turing test.21,22 For this purpose, we mixed randomly generated and real images and asked radiologists and two medical imaging scientists to classify which images are generated and which are not. 3.Experiments and ResultsFor generating normal CT scans, we empirically set the dimension of the latent space to 512, i.e., we consider vectors . Smaller dimensions led to decreased quality of the generated images showing repetitive image artifacts. Larger latent vectors did not further improve the results. As the method relies on the capability of the model to generate data that follow the distribution of normal healthy images, it is of interest to use a good generator. To give a quantitative analysis of the generated normal CT scans, we compute the unbiased Fréchet inception distance.23 Comparing the generated samples from the GAN, we obtain for real healthy examples, whereas for real disease images. Thus, our generated images are more similar to healthy images. We evaluate the quality of the GAN images through a visual Turing test, where we randomly shuffled 25 real CT images and 25 CT images generated by our GAN, and two experienced radiologists and two experienced scientists from our medical imaging community evaluate which images were generated images. The results are summarized in Table 2. The radiologists rated the image quality as very good and were unable to correctly classify of the images. Our medical imaging colleagues were unable to distinguish between fake and real images. Table 2Turing test results.
TP, number of correctly classified generated images; TN, number of correctly classified real images. In addition, qualitative examples of the generated images are shown in Fig. 3. It is visible that a variety of ventricle and skull shapes are represented by the generated images. Brain tissue details such as sulci or center line are recognizable, which lets the images appear realistic. To obtain a further quantitative value for the reconstruction quality, we calculated the root mean square error (RMSE) between the input and output images for each of the 100 real CT slices with and without anomalies. For this purpose, the intensity range of the images was normalized to the interval . Although we obtain a mean and standard deviation of the absolute difference of and an RMSE of 0.089 for the image pairs with anomalies, the mean ± standard deviation and the RMSE for normal image pairs are only and 0.076, respectively. Moreover, the 90% percentile of the pixel difference was 0.074 for normal images and 0.106 for images with abnormalities. To summarize, we achieve good reconstruction quality for healthy images. The measured reconstruction quality of the healthy images is significantly better compared with images with abnormalities, which is due to the fact that abnormal areas are poorly or “healthy” reconstructed. In addition, the quality of healthy image reconstruction is sufficient so that the images can be differentiated into two classes (normal and abnormal), as in Walluscheck et al.18 A qualitative impression of our results for three test images is given in Fig. 4. The shown reconstructions are average images of the reconstructions of three GAN models and therefore appear blurrier than the individual reconstructed images. The shown residual is color-coded to make it visible in the overlay. Stronger deviations appear in a darker orange color. We also test the performance of our method on CT volumes. Figure 5 shows the segmentation overlay on several consecutive axial slices for two test volumes. We observe that using our method on 2D slices of CT scans leads to sensible results when applied to subsequent slices. As done here, a connected component analysis on the results of multiple slices can further reduce small false-positive findings. 3.1.Hemorrhage DatasetWhen testing on hemorrhage data, only positive values are considered because hemorrhages appear brighter than healthy tissue. In Fig. 6, we present the Dice overlap for increasing the anomaly threshold for our model with encoder and with restoration on the validation set (15 images). We observe that our results highly depend on the threshold. After choosing a threshold on the validation set (), the network with restoration achieved a median Dice overlap of 0.71 (mean Dice ) on the test dataset (15 images). The Dice for the version with an encoder is slightly lower with a median of 0.66 (mean Dice ) leading to a -value of 0.96 in a statistical Wilcoxon test. We detect 91.7% of the reference lesions. A lesion is counted as detected when the Dice is for both network versions. We also tested the supervised hemorrhage segmentation tool BlastCT20 on our test data and obtained a median Dice score of 0.73. The difference in our results is not statistically significant (-value ). Therefore, our methods achieve similar results as this supervised approach. 3.2.Combined DatasetAs our approach is not trained to detect a specific type of lesion, it can be used on data, including multiple anomalies. When testing on data with tumors and hemorrhages, we do not consider any prior knowledge of image values. Thus, the anomaly threshold is applied to the absolute values of the difference image. First, we determine a threshold on the validation set (27 images) and use this operating point to report results on our test data (26 images). The best result is obtained with a median Dice overlap of 0.43 (mean Dice ) for the restoration and 0.30 (mean Dice ) for the encoder network version (Table 3). A statistical test on both results leads to a -value of 0.02 showing statistical significance. We observe that, overall, the Dice values are lower than the previously presented results for the hemorrhage dataset. However, we would like to emphasize that we did not develop a method for the exclusive detection of hemorrhage and therefore include these data in our evaluation. Furthermore, we achieved an accuracy of 92.9% on the tumor data in our test set, which means that our method is suitable for application to multiple types of abnormalities. In Fig. 7, we show the ROC curves for the model with encoder and restoration for two test sets. We see that the restoration method performs better than the encoder model. Also, the performance for both models is better on the hemorrhage test set than on the combined test set. In addition, we mark the performance of the BlastCT tool20 (tested on hemorrhage data) as a point in the graph and see that it is similar to our models. Table 3Summary of the anomaly segmentation for two network variants on two test sets showing the mean Dice values with standard deviation.
3.3.Comparison with Other ModelsA natural candidate for comparison to our method would be the work of Lee et al.,13 which has been proposed for emergency triage in brain CT scans. Their results are of high quality and very promising. We applied the available trained network to our test data but obtained unsatisfactory reconstruction results that showed strong deviations from our input images. Therefore, we refrain from further quantitative analysis for comparison here. In addition, we compare our model with two state-of-the-art models, namely, an autoencoder24 and a diffusion model called “AnoDDPM,”10 that were both proposed for anomaly detection in brain MR images. We trained both methods on our training dataset and applied them to our hemorrhage test data (15 images). Figure 8 shows ROC curves for the autoencoder, diffusion model, and our two network versions on our hemorrhage test dataset. It shows that our method performs slightly better. To get a visual impression of the performance difference, we show the results for all four models on a disease test case with prominent hemorrhages in Fig. 9. We note that the hemorrhage within a ventricle is still visible in the reconstruction of the AnoDDPM. We found that the noise distribution used for the diffusion model has a strong influence on what kind of abnormalities are removed in the reconstruction. This could be a disadvantage as it cannot be ensured that the reconstruction always represents normal anatomy. We note that the hemorrhage within a ventricle is still visible in the reconstruction of the AnoDDPM. We found that the noise distribution used for the diffusion model has a strong influence on what kind of abnormalities are removed in the reconstruction. This could be a disadvantage as it cannot be ensured that the reconstruction always represents normal anatomy. The image reconstructed by the autoencoder does not show details inside the brain tissue and is generally blurred. In this example, we see that our model provides images with normal anatomy, even for this example image that shows a brain heavily affected by bleeding. Our network produces sharp and detailed images that contain no abnormalities. 4.Discussion and ConclusionWe present a method for detecting nonspecific abnormalities in brain CT scans that does not require annotated data. Unlike many GAN approaches, the architecture we have chosen requires only a very small amount of training data and produces high-resolution images that are less blurred than the VAE or AE methods. Instead of learning to detect specific abnormalities, we learn normal representations of brain CT scans and detect the deviations from normal in unseen images. We train with healthy 2D CT slices only and show that we can detect and localize hemorrhages and tumors. When testing our unsupervised approach with hemorrhage data, we obtain good results comparable to a recent supervised CT segmentation tool (Monteiro et al.20) for hemorrhages. Thus, we show that our approach can localize lesions that were not specified during training. Our network is not susceptible to the size of the anomalies as we compare the images per pixel and do not filter or constrain the output according to the residual size. By training a GAN to produce normal images in the first step, we ensure that we only obtain images that show normal anatomy, regardless of the type or severity of the abnormality in the input image. At the same time, our reconstructions have good image contrast and show more details of the brain tissue than comparable autoencoder approaches. For our model, we compare two approaches to image reconstruction, namely, training an encoder and iterative optimization for image restoration. We find that both methods produce very similar reconstructions. In the quantitative evaluation, the restoration method seems to perform slightly better, but the difference is not statistically significant. Also, the restoration model requires 110 s to reconstruct a 2D image, whereas the encoder model takes only around 0.02 s, which could be advantageous in an application. As our method is based on grayscale comparisons, the performance is limited by the gray value deviation of the lesion compared with the healthy tissue. Very subtle lesions are therefore likely to be missed. However, this is a general observation that affects any approach based on detecting lesions by analyzing image value intensities and also includes manual detection by radiologists. However, a limitation of our work is the limited training dataset of 1180 images and the small test dataset of 26 images. It would be a great future addition to further test the model on a larger amount of data if more images were available. We have shown that our method produces consistent results when applied to consecutive axial slices and is therefore suitable to locate lesions in CT volumes. By learning a representation of normal CT scans, we can reconstruct the overall anatomy of a healthy brain. Thus, structural changes such as ventricular deformations are also corrected in our reconstruction. In the future, detection of such changes may also be possible with our method. Anatomical details such as the exact location of the sulci of a particular patient are not optimally reconstructed. Therefore, these differences between input and reconstruction are currently detected as anomalies in some cases, as can also be seen in Fig. 4 (small extra areas in the segmentation mask). In future work, we would like to investigate how to distinguish between these “irrelevant” differences and true anomalous differences. Possible research could include developing learning-based methods to obtain anomaly maps from the reconstructed images rather than computing the difference image. DisclosuresThe authors have no competing interests to declare that are relevant to the content of this article. Code and Data AvailabilityIn this study, we used two publicly available datasets. The RSNA data that we used for training are available from Ref. 14. For testing, we utilized a glioma dataset that is available from the Cancer Imaging Archive.15 Due to the research being part of a running project funded by the Federal Ministry of Education and Research of Germany (BMBF), the code developed for this work cannot be made publicly available. AcknowledgmentsThis work was funded by the Federal Ministry of Education and Research of Germany as part of AutoRAD (Project No. 13GW0491B). ReferencesR. J. M. Bruls and R. M. Kwee,
“Workload for radiologists during on-call hours: dramatic increase in the past 15 years,”
Insights Imaging, 11
(1), 121 https://doi.org/10.1186/s13244-020-00925-z
(2020).
Google Scholar
S. Wang et al.,
“Brain stroke lesion segmentation using consistent perception generative adversarial network,”
Neural Comput. Appl., 34 8657
–8669 https://doi.org/10.1007/s00521-021-06816-8
(2022).
Google Scholar
C. Baur et al.,
“Deep autoencoding models for unsupervised anomaly segmentation in brain MR images,”
Lect. Notes Comput. Sci., 11383 161
–169 https://doi.org/10.1007/978-3-030-11723-8_16 LNCSD9 0302-9743
(2019).
Google Scholar
C. Baur et al.,
“Modeling healthy anatomy with artificial intelligence for unsupervised anomaly detection in brain MRI,”
Radiol.: Artif. Intell., 3
(3), e190169 https://doi.org/10.1148/ryai.2021190169 AINTBB 0004-3702
(2021).
Google Scholar
S. Akcay, A. Atapour-Abarghouei and T. P. Breckon,
“Ganomaly: semi-supervised anomaly detection via adversarial training,”
Lect. Notes Comput. Sci., 11363 622
–637 https://doi.org/10.1007/978-3-030-20893-6_39 LNCSD9 0302-9743
(2019).
Google Scholar
C. Gong et al.,
“Generative AI for brain image computing and brain network computing: a review,”
Front. Neurosci., 17 1203104 https://doi.org/10.3389/fnins.2023.1203104 1662-453X
(2023).
Google Scholar
C. Baur et al.,
“Autoencoders for unsupervised anomaly segmentation in brain MR images: a comparative study,”
Med. Image Anal., 69 101952 https://doi.org/10.1016/j.media.2020.101952
(2021).
Google Scholar
T. Schlegl et al.,
“f-AnoGAN: fast unsupervised anomaly detection with generative adversarial networks,”
Med. Image Anal., 54 30
–44 https://doi.org/10.1016/j.media.2019.01.010
(2019).
Google Scholar
M. Arjovsky, S. Chintala, L. Bottou,
“Wasserstein generative adversarial networks,”
in Proc. 34th Int. Conf. Mach. Learn.,
214
–223
(2017). Google Scholar
J. Wyatt et al.,
“AnoDDPM: anomaly detection with denoising diffusion probabilistic models using simplex noise,”
in IEEE/CVF Conf. Comput. Vision and Pattern Recognit. Workshops (CVPRW),
649
–655
(2022). https://doi.org/10.1109/CVPRW56347.2022.00080 Google Scholar
J. S. Viana et al.,
“Unsupervised 3D brain anomaly detection,”
Lect. Notes Comput. Sci., 12658 133
–142 https://doi.org/10.1007/978-3-030-72084-1_13 LNCSD9 0302-9743
(2021).
Google Scholar
M. Toikkanen, D. Kwon and M. Lee,
“ReSGAN: intracranial hemorrhage segmentation with residuals of synthetic brain CT scans,”
Lect. Notes Comput. Sci., 12901 400
–409 https://doi.org/10.1007/978-3-030-87193-2_38 LNCSD9 0302-9743
(2021).
Google Scholar
S. Lee et al.,
“Emergency triage of brain computed tomography via anomaly detection with a deep generative model,”
Nat. Commun., 13 4251 https://doi.org/10.1038/s41467-022-31808-0 NCAOBW 2041-1723
(2022).
Google Scholar
A. E. Flanders et al.,
“Construction of a machine learning dataset through collaboration: the RSNA 2019 brain CT hemorrhage challenge,”
Radiol.: Artif. Intell., 2
(3), e190211 https://doi.org/10.1148/ryai.2020209002 AINTBB 0004-3702
(2020).
Google Scholar
N. Shusharina and T. Bortfeld,
“Glioma image segmentation for radiotherapy: RT targets, barriers to cancer spread, and organs at risk (GLIS-RT) [data set],”
(2021). https://doi.org/10.7937/TCIA.T905-ZQ20 Google Scholar
B. Liu et al.,
“Towards faster and stabilized GAN training for high-fidelity few-shot image synthesis,”
in Int. Conf. Learn. Represent.,
(2021). Google Scholar
T. Miyato et al.,
“Spectral normalization for generative adversarial networks,”
in Int. Conf. Learn. Represent.,
(2018). Google Scholar
S. Walluscheck et al.,
“Unsupervised learning of healthy anatomy for anomaly detection in brain CT scans,”
Proc. SPIE, 12465 1246504 https://doi.org/10.1117/12.2653889 PSISDG 0277-786X
(2023).
Google Scholar
J. Johnson, A. Alahi and L. Fei-Fei,
“Perceptual losses for real-time style transfer and super-resolution,”
Lect. Notes Comput. Sci., 9906 694
–711 https://doi.org/10.1007/978-3-319-46475-6_43 LNCSD9 0302-9743
(2016).
Google Scholar
M. Monteiro et al.,
“Multiclass semantic segmentation and quantification of traumatic brain injury lesions on head CT using deep learning: an algorithm development and multicentre validation study,”
Lancet Digital Health, 2 https://doi.org/10.1016/S2589-7500(20)30085-6
(2020).
Google Scholar
D. Geman et al.,
“Visual Turing test for computer vision systems,”
Proc. Natl. Acad. Sci. U. S. A., 112 3618
–3623 https://doi.org/10.1073/pnas.1422953112
(2015).
Google Scholar
M. Jang et al.,
“Image Turing test and its applications on synthetic chest radiographs by using the progressive growing generative adversarial network,”
Sci. Rep., 13
(1), 2356 https://doi.org/10.1038/s41598-023-28175-1
(2023).
Google Scholar
M. J. Chong and D. Forsyth,
“Effectively unbiased FID and inception score and where to find them,”
in IEEE/CVF Conf. Comput. Vision and Pattern Recognit. (CVPR),
6069
–6078
(2020). https://doi.org/10.1109/CVPR42600.2020.00611 Google Scholar
G. Luo et al.,
“Unsupervised anomaly detection in brain MRI: learning abstract distribution from massive healthy brains,”
Comput. Biol. Med., 154 106610 https://doi.org/10.1016/j.compbiomed.2023.106610 CBMDAW 0010-4825
(2023).
Google Scholar
|
Education and training
Computed tomography
Brain
Image restoration
Neuroimaging
Medical image reconstruction
Gallium nitride