1.IntroductionIn the emergency room, physicians use neuroimaging to assess changes in blood irrigation in the brain and to define treatments.1 In particular, perfusion imaging is used to quantify the core and penumbra of the lesion in ischemic stroke patients and hereby tailor the treatment decision based on standardized procedures, as well as to predict effectiveness. Studies show that diffusion and perfusion weighted imaging based on magnetic resonance (DWI-MRI and PWI-MRI) are highly accurate methods for inferring the infarct core of the ischemic lesion.2–4 These techniques are very sensitive to intracellular water shifts after cell depolarization; hence it is easy to identify the core lesion.5 In practice, these scanners are often unavailable, or their acquisition time exceeds the time frame to extend a diagnosis to the patient,6 and computer tomography (CT) methods using contrast agents are preferred. Diagnosis using computer tomography perfusion (CTP) is rated as of equivalent value in trial and medical practice;7 nonetheless, CTP imaging is inherently more difficult to interpret because CTP is less sensitive to the small change in attenuation caused by water uptake in the acute ischemic brain tissue. In addition, apparent diffusion coefficient contrast increases linearly over time, and the image contrast changes at different time points during the diagnosis time. We are interested in improving the image analysis by making use of a graph neural network to automatically segment stroke lesions from CTP-parameter maps. Imaging is essential in modern medical practice among all different specializations, but its analysis has many difficulties. Practitioners use them to assess the dimensions, structures, and topology of internal organs to identify abnormalities and establish a medical treatment. Although trained professionals can evaluate the patient’s condition from unprocessed images, the precise delimitation of relevant components remains difficult and time consuming. Studies confirm that there exists an intra- and inter-observer variability of manual segmentation that depends on the complexity of the target structure.8–10 These discrepancies could be associated with the intrinsic limitations of the cognitive processes in the human vision system.11 Moreover, in the case of a stroke diagnosis, many highly debated factors make harder the estimation of final infarct lesions for perfusion imaging methods. For example, reversal of the core and penumbra is sometimes observed, albeit the reported results were regarded as non-clinically significant.12,13 The extension of collateral vascularization, genetics, and external stimuli (e.g., chronic hypoperfusion) leads to changes in preliminary estimations of volumes mismatch estimations.6 Another factor, the lack of standardized software for the calculation of stroke parameters encumbers the definition of an optimized threshold value to make a simple threshold segmentation.14–16 In addition, the location of the ischemic stroke lesion affects the vulnerability of hypoperfusion and the outcome of the treatment.15 All of these difficulties make the estimation of penumbra from CTP a burdensome task, and a successful manual segmentation of stroke lesion from CTP images depends greatly on the expertise and ability of the interpreter. As a result, in an ischemic stroke diagnosis, manual segmentation is unpractical and prone to errors. In this regard, automated segmentation methods demonstrated promising potential to support the diagnosis of stroke patients. In recent years, machine learning methods have been applied to solve the stroke lesion segmentation from CTP images problem with improved, yet not perfect, results.17–19 In general, the segmentation of brain structures from MRI or CT images is a difficult task and entails a considerable number of problems. For example, training machine learning models requires large amounts of data,20 but available datasets have only a few samples.10,21–23 Likewise, medical conditions, such as tumors, oedemas, and other lesions, introduce other problems such as fuzzy boundaries and have a high variance of shapes and locations.19,24,25 Additionally, imaging artifacts, different scanners/protocols, and anatomical variability (e.g., age and neurodegeneration) introduce contrast and intensity variations that also affect the stroke lesion datasets.26 Despite the success of convolutional neural networks (CNNs) in medical image analysis, in the case of ischemic stroke, there are many issues not thoroughly solved,17 and it is an active research area. Currently, the best solution is proposed in Ref. 27, which combines a generative model that produces a pseudo-DWI from the CTP-parameters and an attention-based loss. Additionally, other relevant solutions use U-Net on patches or 3D-convolutions in Refs. 28 and 29, respectively. However, the domain of CNN is a Euclidean domain, i.e., pixel grids, meaning that the structure of the features is limited by pixel position.30,31 An emerging field in deep learning denominated “geometric deep learning” proposes an extension of CNN to non-Euclidean structured convolution, which allows for positioning of inter-pixel features. Geometric deep learning has proven to successfully solve image classification of natural 2D images using spectral and spatial non-Euclidean convolutions.32,33 In neuroscience, graph neural networks have also been extensively used for the analysis of cortical gyrification: modeling of anatomical features,34 cortical parcellation,35–38 and understanding subtle topological dependencies in classification of functional MRI signals.39–41 Geometric deep learning leverages non-Euclidean convolution on meshes preserving cortical brain topology.35 In image segmentation, geometric deep learning is used to address the loss of feature localization.42–44 In a similar line to our proposed model, Juarez et al.43 and Lu et al.44 proposed graph fully convolutional network (GFCN) and U-Net-graph, respectively, to leverage the node connectivity but without using pooling operations. The absence of pooling introduces the disadvantage of increasing computational cost and memory footprint to process inputs. In addition, the mentioned models are based on spectral convolution, which lacks directional information and critical information to define the object boundaries.32 Our method differs from other graph encoder–decoder models,44,45 as we use spline convolution and pooling operators, allowing us to have different weights in different directions, thus extracting richer geometrical information. In this work, we defined a deep learning model with graph convolutional operations in an encoder–decoder architecture for the task of segmentation of stroke lesions. To this end, we propose an architecture for graph-structured data that resembles the fully convolutional network (FCN),46 in which the convolution blocks are replaced with the spline convolutions proposed in Ref. 32, and upsampling layers are an approximation of interpolation in graphs. We theorize that a graph neural network could leverage from a more complex feature map and its capacity of connecting inter-pixel information in different angles to detect the lesion more accurately. We evaluated this by inferring the internal functionality of a graph-based CNN on the segmentation masks generated by our algorithm. The inputs to our model are non-contrast CT and CTP-parameters from the ischemic stroke lesion segmentation 2018 (ISLES2018) challenge dataset.47,48 We specifically study the flaws and benefits of using a geometric deep learning algorithm to predict ischemic tissue from CT-perfusion parameters and non-contrast CT in the ISLES2018 dataset.47,48 The ground truth corresponds to the core lesion; thus the model predicts irreversible lesion tissue probability. The model is trained under different configurations to extrapolate the internal processes of the model in correspondence to its hyperparameters. We compare the model against the reported results of Refs. 2728.–29. In addition, we train and compare the results of a U-Net,49 FCN-8s,46 and PointNet++.50 In summary, the contributions of this work are as follows.
In Secs. 2.2 and 2.1, we discuss the dataset, preprocessing, evaluations approaches, and model architecture. In Sec. 3, we present the results of the ablation study to understand the function of different components of our algorithm. In addition, we unfold the results of the comparison of our model against the mentioned models. Finally, in Secs. 4 and 5, we discuss and present our conclusions. 2.Materials and MethodsIn this section, we describe the network architecture in terms of the different depth configurations and the pooling and unpooling methods used. Next, we present the dataset, preprocessing, and evaluations approaches that were developed. 2.1.Network ArchitectureThe model used in this work has a similar architecture to the FCN in Ref. 46. We consider three variants, FCN-32s, FCN-16s, and FCN-8s, which differ in the way that skip connections are added to the upsampling path. The FCN-32s requires a upsampling after five pooling layers of half-steps. The FCN-16s uses skip connections fusing features from previous layers by element-wise addition and requires a upsampling because the output comes from four pooling layers of half-steps. Similarly, the FCN-8s uses a two-skip connection, so the output requires a upsampling step, here the final output comes after three pooling layers of half-steps. In general, the FCN is built from local operations: convolution, pooling, and deconvolution. The deconvolution reconstructs a fine pixel a representation out of a coarser pixel structure. Subsequently, the network is divided into two parts: the downsampling path and the upsampling path.
The FCN architectures are implemented using equivalent graph-based components, namely spline convolution filters,32 graph-based pooling operators, and two graph-based unpooling operators, described in Sec. 2.1. Figure 1 shows the FCN-8s variant of the FCN with graph operations. These variants are denoted GFCN.
The argument to use upsampling operators has the same meaning as in the Euclidean domain, where upsampling allows for recovering the initial dimensionality of the input and creating the segmentation map. In the case of the GFCN, the downsampling path and upsampling path have a non-uniform field of view expansions and contractions, which might help with the problem of insufficient localization of long-range features in the standard Euclidean CNNs as described in Ref. 51. This is also implemented in Ref. 45, though we explore two different ways to recover the features: the first approach simply copies the values to the neighbor nodes and the second approach distributes the values proportionally to their feature value in previous pooling topologies. The latter looks for having a perfect reconstruction of the previous feature space. However, proportional unpooling will be an approximation because we employ a max function as aggregation in the pooling layer, and a perfect reconstruction will only have an effect if average pooling was used. We decide on maxpooling because average pooling introduced vanishing gradients during training in our preliminary experiments; we do not report any results of those experiments in this manuscript. 2.1.1.Pooling operatorsThe pooling layer reduces the number of nodes aggregating sets of similar nodes and applying a symmetry invariant operator, in this case, the max operator. Therefore, the pooling operation in layer is done in two steps: first, a clustering forms a subset and the second step aggregates them with the max operator to form the feature of the node in the next layer. We explore three pooling approaches (i) the Top-k pooling,45 (ii) the radius clustering of points selected by the farthest point sampling algorithm as in Ref. 50, and (iii) finally the Graclus algorithm.52 2.1.2.Unpooling operatorsThe unpooling operators restore the previous graph topology. We propose two approaches: the isotropic operator and the “proportional unpooling operator.” As a benchmark, we also use the KNN unpooling operator from Ref. 50, which computes the weighted sum of neighbor nodes in the current layer to the node in the next layer. The weights are inversely proportional to the square distance of the nodes. The “isotropic unpooling operator” copies the features in the positions of the previous nodes that were aggregated into the target node ; this is shown in Fig. 2, where the values from the target node topology in layer are copied to the position of the nodes of the next layer. For that, we store the pooling assignment for node in the target topology, where ; hence we write The proportional unpooling operator applies a factor that weights the feature propagation proportional to the sum of all members of the cluster ; then the propagation is written as The operations are computed without calculating the gradient of the weights to reduce the GPU memory, and the gradient is preserved in the upsampled node , i.e., is independent of the weights. 2.2.Data and Preprocessing2.2.1.ISLES2018 datasetWe use the dataset from the challenge for stroke lesion segmentation, ISLES2018, which consists of CTP images within 8 h after the stroke episode, and a DWI within 3 h after the CTP was performed. The dataset consists of the perfusion parameter maps: cerebral blood flow, cerebral blood volume, mean transit time, and time to peak (Tmax). The original partition has 94 samples for training with mask information and 63 samples for testing without mask information. In the comparison experiments, we use a 3:1 rate training and testing cross-validation scheme out of the 94 cases with a mask. The training set in the cross validation is additionally split into a 9:1 rate for unbiased best model selection. Thereby, the final dataset splits use 65 cases for training, 6 for validation, and 23 for testing. 2.2.2.PreprocessingPreprocessing is considered a critical step in training the model. We use as inputs the structural CT and the CTP parameters, as well as a min–max instance normalization in each volume in a similar line as in Ref. 53. In the case of the CT, we enhance the contrast of the brain using a mask on the non-zero values of the sum of the CTP-parameters for each sample, in a similar way as in Ref. 28. No augmentation method was employed. 2.2.3.EvaluationWe train the networks in the ablation study for 100 epochs equivalent to 4500 optimization steps with a batch of 4. We train most cases with different learning rates to ensure convergence, with the exception of variations of model architectures in which the learning rate was kept constant, as we want to evaluate the convergence speed due to the increment of features or the placement of the batch normalization layer. The training is done in a GEFORCE RTX 2080 TI with 11 GB of memory, an Intel® Xeon® CPU E5-2665 0 at 2.40 GHz, and 124 GB of RAM. The qualitative analysis considers the calculation of the Dice coefficient score (DCS): where TP, TN, and FN stand for the cardinality of the sets of true positive, true negative, and false negative voxel sets corresponding to a given segmentation with respect to a ground truth . In addition, the Hausdorff distance (HD), recall, precision, and coefficient of determination (COD) are computed:The models are trained under a fourfold cross-validation regime with splits for training, validation, and testing of 65, 6, and 23, respectively. The best model trained in the training set is selected using the validation set, and the metrics reported correspond to the unseen samples on the testing set. The significance of comparisons is done with a pair -student test on the testing set. We compute the metrics DCS, accuracy, recall, precision, HD, and COD on average for every slice in the validation set after each epoch. The calculation is done in a sample-wise manner, meaning that the values are averaged over all slices in one batch and then averaged on the whole dataset. These values were relatively small because using 2D slices makes some of them have a mask of zero, which makes the values drop. We cope with this by calculating the evaluation metrics at the end of the training in the testing set in a case-wise (volume-wise) manner, i.e., we consider all of the voxels of a corresponding case volume in the dataset and then average these values for all cases in the dataset. 3.ResultsWe structured our study in two parts. First, we made a component evaluation in which we investigated different elements of our model and the correlation with the segmentation results. The second part of the study compared the performance of our model against other representative deep learning methods in the literature. 3.1.Ablation Study: Understanding Model ComponentsWe have three degrees of freedom in the design of the model: (i) the architecture variants with different depths, i.e., 32s, 16s, and 8s; (ii) the unpooling operators, proportional and isotropic; and (iii) the pooling operators, Top-k, and Graclus. Therefore, the first experiments aim to find which is the best architecture and which is the best pooling and unpooling operators combination for the proposed GFCN model. Accordingly, the evaluation procedures are divided into two parts. The first part deals with the architecture configuration question and the second with the pooling and unpooling question, as described below. 3.1.1.Model architectures: down-sampling depth, batch normalization, and skip connectionsThe first experiment compares the model architectures GFCN-8s, GFCN-16s, and GFCN-32s by training on the ISLES2018 challenge dataset (cf., 2.2) over 100 epochs with a constant learning rate of with soft-Dice-loss as the optimization criterion.54 The training performance of the three models is compared using the same pooling and unpooling methods for all trials. We used Graclus pooling and isotropic unpooling. In addition, we investigate the placement of a batch normalization layer before and after the activation layer.55 The results in Table 1 show the performance metrics on the testing set, in which the best performing architecture is the GFCN-8s that uses batch normalization before the activation functions. The GFCN-16s shows lower performance than the GFCN-8s, which suggests that a deeper initial convolutional layer is necessary to extract better descriptors. In addition, comparing the GFCN-32s and the GFCN-16s cues the positive effect of skip connections. The GFCN-16s uses a skip connection from the pool-score of the graph topology, whereas the GFCN-32s is devoid of forwarding loops. We observe this effect as much for pre-batch normalization (pre-BN) as for post-batch normalization (post-BN). Table 1Comparison of model architectures on the ISLES2018 challenge dataset for segmentation of ischemic stoke lesions. Metric calculated per volume in average for 23 testing samples.
Best performance in bold. Figure 3 shows that the three architectures continue improving after 100 epochs; still, the GFCN-8s with deeper initial feature extraction improves more quickly than the other two architectures. Regardless of the post-BN starting at a higher value than the pre-BN, the velocity of convergence increases in the pre-BN case. This is consistent across the architecture variations, though the difference increases with a deeper architecture. For example, this can be observed by comparing the differences between pre-BN and post-BN in the GFCN-8s or GFCN-32s. 3.1.2.Pooling and unpooling methodsThe second experiment compares the pooling and unpooling methods. The model architecture used is the GFCN-8s trained from scratch on the ISLES2018 dataset for 100 epochs with early stopping. Again, the learning rate is constant. We collect the metrics after each epoch sample-wise, and at the end of the training, we evaluate volume-wise on the 23 testing samples. We defined four variants of the models that combine compatible operators, i.e., for pooling: Graclus and Top-k;45 and for unpooling: isotropic, proportional, and k-NN interpolation of Ref. 50. The isotropic and proportional unpooling employ the Graclus pooling layers, as detailed in Sec. 2.1; and in the case of the Top-k, it uses the k-NN interpolation as the pooling method. Finally, we included one model that uses no pooling operators. This adds up to four models studied in this experiment: isotropic, proportional, Top-k, and no-pooling. The results presented in Table 2 show the performance metrics of four variants of pooling and unpooling layers for a fixed architecture GFCN-8s. We observe that performance of the isotropic and proportional upsampling remains in a similar range. In all trials, the HD is lower for the isotropic pooling than for the proportional pooling ( -value). This is consistent with what is shown in Fig. 4, where the boundaries obtained with the isotropic are closer to the ground truth, though segmentation probabilities are smoother in the proportional upsampling. It is worth noting that the isotropic approach is less computationally expensive than the other approaches, which translates into less training time. Top-k and no pooling generally have better metrics than the proposed upsampling methods but considerably less sensitivity ( -value). Table 2Upsampling methods comparison on the ISLES2018 challenge dataset. Isotropic stands for the isotropic unpooling operator. Proportional stands for the proportional unpooling operator. Comparisons of fixed GFCN-8s architecture with batch normalization before activation, trained over 100 epochs.
Best performances in bold. 3.2.Performance Comparison with Other MethodsIn the second part of the experiments, we contrast the proposed model with existing models for semantic segmentation. Consequently, we train various models in the literature from scratch using the same inputs as we use for the proposed model, namely the FCN-8s,46 U-Net,49 and PointNet++.50 The evaluation is done with fourfold cross validation in the ISLES2018 dataset with the same splits as the one for the GFCN-8s. In all experiments, the models are trained for a maximum of 300 epochs with early stopping to avoid overfitting. The learning rate is reduced after 100 epochs by a factor of 10. We train the models using an Adam optimizer and a soft Dice loss.54 The proposed model, denominated GFCN-8s, uses Graclus pooling and isotropic upsampling layers. The results in Table 3 showed better metrics for the GFCN-8s compared with the other models trained under the same configuration. FCN-8s has the second best values among these four models. It is worth noticing that the used FCN-8s model is not exactly the model from Ref. 46 but a simplification with a similar architecture to the GFCN-8s. This was adopted because the low number of training samples would be insufficient to train all weights in the original configuration. In the case of the U-Net, we use a bilinear interpolation layer for the upsampling instead of a learnable deconvolution layer, but we include batch normalization. Despite these differences, we preserve the original U-Net architecture in Ref. 49. Finally, in the case of the Pointnet++, we employ the same configuration described in the original work.50 Table 3Comparison of segmentation models on the ISLES2018 challenge dataset. 2D-ARED stands for 2D asymmetric residual encoder–decoder from Ref. 28. COD and accuracy are not reported in Refs. 29, 28, and 27 in the original papers, as well as the dataset partitions. We simply exclude these values in the comparison.
In addition, we append the results reported in Ref. 27 on the SLNet, a 2D-patch-based U-Net presented in Ref. 28, and the 2018 winner algorithm with a 3D U-Net reported in Ref. 29 as reference. Comparing the GFCN-8s against the external models, we notice that a considerable difference exists in the metrics unfavorable to our model, which is especially important in DCS and HD. Notice that the models reported in these external references are extensively optimized and employ complex feature extraction pipelines, special arrangements of convolutional layers and/or advanced augmentation methods. We stay in a simple input configuration and employ no augmentation methods because we are solely interested in understanding the process of graph CNNs for detecting stroke lesions. Figure 5 shows a comparison of the segmentation boundaries for the trained models. The PointNet++, despite being able to successfully capture small structures, has several regions of false positives and therefore has the lowest average accuracy. Comparing the FCN-8s and U-Net, we notice that a deeper model will require more samples to train and refine the prediction as the U-Net successfully localizes the lesion but fails to correctly define the boundaries. The U-Net tends to produce fewer false positives, but it is less sensitive. On the other hand, the results of the FCN-8s are similar to the proportional unpooling from the previous experiment. The FCN-8s extracts smooth probability maps, as depicted in Fig. 5, but the prediction misses reaching the edges of the lesion. In contrast, GFCN-8s has a very flexible prediction output, and regardless of having a lower precision, compared with additional external models, it has the best average metric values among the models that we trained. Figure 6 shows the distribution of metric values calculated volume-wise and stratified into three categories according to the lesion’s volume, namely small, medium, and large lesion sizes. The sets are constructed from the distribution of the number of lesion voxels per scan split into three evenly distributed groups using a quantifier discrete cut. We observe that the PointNet++ is able to collect a higher number of lesion masks consistently along with the distribution of sizes, which leads to high recall values, yet it has the lowest precision values. The proposed GFCN-8s scores higher values compared with the other models trained. A trend of the lesion size in which smaller lesions have worse values than bigger lesions is also noticeable. This is consistent in all of the trained models. 4.DiscussionOur study focuses on identifying the relevance of GFCN in solving the segmentation problem of acute ischemic stroke lesion prediction and investigating the behavior of its components concerning the segmentation results. In this section, we analyze the results concerning the evaluation of different model architectures based on the number of convolutional blocks, pooling/unpooling operations, and batch normalization. In addition, this section also covers the comparison with selected methods trained under the same regime and state-of-the-art methods.27–29 This section discusses, in particular, the significance and limitations of the experiments, as well as points out recommendations for future research. 4.1.Understanding Model Components4.1.1.Feature extraction and perception filtersIncluding more convolutional blocks in the downsampling path improves the feature extraction and as a consequence the segmentation results. In the early days of deep learning, Jia et al.56 showed that spatial pooling allows for constructing overall semantics from the low-level features in analogy to the biological mechanics of the mammalian visual cortex.57 Moreover, images subject to inner and outer scale information exhibit the property of being invariant to small spatial shifts.58 Research suggests that in CNNs this property comes from the denominated upscaled receptive fields.59 The coarsening of the output of a convolutional layer expands the receptive field by a factor equal to the stride, as explained in the FCN original work.46 Further, the pooling factors allow for effectively calculating gradients when the receptive fields overlap. Therefore, the model with a deeper downsampling path has faster improvements and better metric values on their prediction values due to their increased receptive fields. From the analysis, it is difficult to unequivocally identify the feature propagation across nodes. However, the irregular and close adaptation to the edges of lesions might suggest a flexible feature projection, yet we do not provide enough evidence to support this. Further research should be conducted exploring the perception fields and activation maps as in Refs. 60 and 61. 4.1.2.Pooling and unpooling methodsDifferent pooling and unpooling approaches lead to vast differences in the segmentation results. This is shown in the differences in metric values as well as segmentation boundaries with and without pooling operations. Albeit the results obtained with the model “without pooling (no-pooling)” have smooth boundaries and by itself the spline CNN allows for extracting local information, the perception field does not change, which led to diminished performance. In Ref. 34, their model also used a spline CNN convolution layer and no pooling, so they also showed that it is possible to obtain good results. However, we found that the model was more difficult to train due to the high computational requirements during the calculation of gradients. Therefore, we show that pooling plays a major role and it is important to efficiently compute the predictions. Simple heuristic upsampling approaches, such as the isotropic or proportional upsampling, which are independent of the gradient of the model, obtain comparable results to optimized approaches with fewer complications during training. For example, considering the case of the denominated Top-k model, we found that it is rather unstable and prone to vanishing gradients. This problem might be due to the dependence of gradients and the inline optimization of neighbors. In fact, the learnable projection of the Top-k pooling voids the pixel location conversely to the classical 2D pooling scheme; though, to some degree, this might be compensated by the spline-CNN, the optimization remains difficult. In the original work,45 this is not an issue because the local information is not relevant for their problem. As a result, we might expect that heuristic upsampling can reach a more stable and efficient optimization in other segmentation problems. 4.1.3.Batch normalizationAs expected, the placement of batch normalization before the activation function evinced a faster optimization curve than placing it after the activation function; this is shown in Fig. 3 and Table 1, where the batch normalization placement effect was compared. Thereby, the position of normalization in graph networks is consistent with what is stated in Ref. 55 as anticipated. This might imply that, by placing the batch normalization before the activation, different neurons will activate due to a change of sign induced by the non-linearity during the training. By contrast, by placing the batch normalization after the activation, the first and second momenta will not affect the sign. Therefore, it can be inferred that the output of the spline-convolutions have a symmetric non-sparse distribution as with the Euclidean case, as stated in Ref. 55. As a limitation, it is worth mentioning that, by comparing the different architecture configurations, the results shown in Fig. 3 are not at the end of the optimization. We aimed to compare the convergence with a simplified hyperparameter configuration; therefore, we fixed the learning rate and optimization steps. However, this does not void the results shown in Table 1 because the optimization curvature will normally tend to reduce and we should not expect major changes in the optimization trends. In addition, due to the high dimension of the feature spaces, the topology of the optimization will not differ substantially as distances are small.62 Further, the models start at the exact same optimization points as shown in Fig. 3. 4.2.Comparing GFCN with Other MethodsSmaller lesions present a more difficult challenge than bigger lesions. The analysis of the distribution of metrics by size revealed that samples with medium and large sizes have better metric values than smaller samples, which is consistent with what is reported in the literature.19 This might be explained by the fact that small size lesions are associated with class imbalance. Small lesion size has somehow blended with the background inherited by the CT acquisition and resolution limitations, reported in Refs. 53 and 63. Smaller lesions will tend to blend with the background signal surrounding the lesion itself. The GFCN obtained lower metrics compared with state-of-the-art models,27–29 with absolute differences for the DCS metric ranging from 12.18% to 30.80%. It is probable that the proposed model would reach comparable metric values in a more optimized input setup. On the other hand, the GFCN obtained better metrics than the models trained under the same conditions, with absolute differences for the DCS metric ranging from 8.61% to 69.05%. The difference between the results of U-Net and FCN-8s is in line with research showing that big models require large amounts of data.25,64 Therefore, the surprisingly low metric values for the U-Net compared with the FCN-8s might be explained by the simple input and training setup adopted, as the training of bigger models require augmentation.20,65 Despite this fact, the comparison is still valid as we wanted to keep the same simplified training environment. As an outlook, a way to cope with the low number of samples could be to use a patch-wise training approach as in Ref. 28 or the generative model approach as in Ref. 27. 4.3.Medical ImplicationsThe similar behavior of our model in the results of lesion size stratification suggests that our model would obtain similar salient and activation maps as a standard U-Net, and thus we could extrapolate the results of Ref. 31; even so, further study is required to understand these maps with the graph layers of our model. In medical practice, the volume of the lesion can be calculated from the segmentation output of our model, although it is possible that the model will overlook small lesions. An explanation for this can be found in Ref. 31, where bigger lesions generate a larger response in the neural network than small lesions, which implies that these small volumes get lost in signals through the layers of the network. The studies66–68 prove the diagnostic power of penumbra regions extracted from CT perfusion and CTP-parameters in correlation with scored studies, such as Alberta Stroke Program Early CT Score, National Institutes of Health Stroke Scale, or modified Rankin Scale. The same outputs can be rapidly predicted for new unseen CTP inputs with up to 0.4553 DCS because the most time-consuming and complex process of training the model is completed at this time point. Patients will require a non-contrast CT and CTP, with the calculation of the CTP-parameters. The volumes require a simple min–max normalization calculated directly for each volume at a low computational cost. In addition, the proposed model is useful in the assessment of penumbra in cases in which the onset time is undefined in line with the findings in Refs. 69 and 67. It is shown that the determination of penumbra allows for assessing the neurological deficit and infarctic volume in patients for which the time of stroke intake is unknown. Other studies have shown that, by measuring the size of the penumbra and computing the core/penumbra rate, it is possible to identify candidate patients for rtPA perfusion within the 6 h window upon stroke intake as it is demonstrated that these patients might have an improved outcome compared with placebo patients. Therefore, the proposed model in combination with the regression models proposed in Ref. 67 could potentially allow for predicting, for example, the NIHSS at 7 days after admission. 5.ConclusionsIn this study, we focused on understanding the principles governing a graph-based FCN to estimate irreversible brain stroke lesions and to see how they differ from classical Euclidean models. Based on the ablation experiments, we observed changes in the results with deeper networks, in which more convolutional blocks enhanced the segmentation results. Furthermore, an overall view of the segmentation results showed that feature propagation in the reception fields developed into an irregular and closer adaptation to the edges of the lesion, evincing the effect of inter-pixel features. With regard to the different pooling and unpooling approaches used, we noticed that they lead to visible differences in the segmentation results (cf., Fig. 4), where simple heuristic approaches allow for fewer features and therefore less computation. In general, the model could be used in medical practice, but it will overlook small lesions. In comparison with other methods, we found that smaller lesions were more difficult to identify than bigger lesions, which is consistent with the literature. The evaluation of our model against models trained in the same regime showed that our model performed better for the metrics reported; for example, in the case of the DCS metric, we obtained improvements ranging from 8.61% to 69.05%. However, the training approach can be improved by changing the inputs to precomputed DWI from generative models as in Ref. 27 or using patch-wise training as in Ref. 28. Moreover, it might be advantageous to restructure the proposed architecture, for example, to introduce a learnable deconvolution as in Refs. 70 and 71. In addition, further visualization methods, such as activation maps or salient maps as in Refs. 39 and 61, could help to understand better the internal process of the model. The activation maps are particularly important for validating and assessing prediction in medical applications.72–74 In addition, they could shed some light on characteristics of the feature propagation across the field of view within the layers of the GFCN.75–77 DisclosuresThe authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. Code, Data, and Materials AvailabilityThe code is available at https://github.com/aiporre/gcn_segmentation AcknowledgmentsThe publication of this article was supported by the German Cancer Research Center, DKFZ, with an open access agreement. ReferencesB. K. Menon et al.,
“Aspects and other neuroimaging scores in the triage and prediction of outcome in acute stroke patients,”
Neuroimaging Clin., 21
(2), 407
–423 https://doi.org/10.1016/j.nic.2011.01.007 NCNAEO
(2011).
Google Scholar
D. C. Tong et al.,
“A standardized MRI stroke protocol: comparison with CT in hyperacute intracerebral hemorrhage,”
Stroke, 30
(9), 1974
–1981 https://doi.org/10.1161/01.STR.30.9.1974a SJCCA7 0039-2499
(1999).
Google Scholar
D. C. Tong et al.,
“Correlation of perfusion-and diffusion-weighted MRI with NIHSS score in acute (<6.5 hour) ischemic stroke,”
Neurology, 50
(4), 864
–869 https://doi.org/10.1212/WNL.50.4.864 NEURAI 0028-3878
(1998).
Google Scholar
P. D. Schellinger et al.,
“A standardized MRI stroke protocol: comparison with CT in hyperacute intracerebral hemorrhage,”
Stroke, 30
(4), 765
–768 https://doi.org/10.1161/01.STR.30.4.765 SJCCA7 0039-2499
(1999).
Google Scholar
D. Saur et al.,
“Sensitivity and interrater agreement of CT and diffusion-weighted MR imaging in hyperacute stroke,”
Am. J. Neuroradiol., 24
(5), 878
–885
(2003).
Google Scholar
K. S. Yew and E. Cheng,
“Acute stroke diagnosis,”
Am. Fam. Phys., 80
(1), 33
–40 AFPYAE 0002-838X
(2009).
Google Scholar
B. C. Campbell et al.,
“Comparison of computed tomography perfusion and magnetic resonance imaging perfusion-diffusion mismatch in ischemic stroke,”
Stroke, 43
(10), 2648
–2653 https://doi.org/10.1161/STROKEAHA.112.660548 SJCCA7 0039-2499
(2012).
Google Scholar
B. H. Menze et al.,
“The multimodal brain tumor image segmentation benchmark (BRATS),”
IEEE Trans. Med. Imaging, 34
(10), 1993
–2024 https://doi.org/10.1109/TMI.2014.2377694 ITMID4 0278-0062
(2014).
Google Scholar
M. Deeley et al.,
“Comparison of manual and automatic segmentation methods for brain structures in the presence of space-occupying lesions: a multi-expert study,”
Phys. Med. Biol., 56
(14), 4557 https://doi.org/10.1088/0031-9155/56/14/021
(2011).
Google Scholar
T. Fechter et al.,
“Esophagus segmentation in CT via 3D fully convolutional neural network and random walk,”
Med. Phys., 44
(12), 6341
–6352 https://doi.org/10.1002/mp.12593 MPHYA6 0094-2405
(2017).
Google Scholar
A. Sabih et al.,
“Image perception and interpretation of abnormalities; can we believe our eyes? Can we do something about it?,”
Insights Imaging, 2
(1), 47
–55 https://doi.org/10.1007/s13244-010-0048-1
(2011).
Google Scholar
P. Kranz and J. Eastwood,
“Does diffusion-weighted imaging represent the ischemic core? An evidence-based systematic review,”
Am. J. Neuroradiol., 30
(6), 1206
–1212 https://doi.org/10.3174/ajnr.A1547
(2009).
Google Scholar
B. C. Campbell et al.,
“The infarct core is well represented by the acute diffusion lesion: sustained reversal is infrequent,”
J. Cereb. Blood Flow Metab., 32
(1), 50
–56 https://doi.org/10.1038/jcbfm.2011.102
(2012).
Google Scholar
M. H. Selim, Acute Stroke Imaging, 93
–174 2nd ed.Cambridge University Press, (
(2013). Google Scholar
J. Demeestere et al.,
“Review of perfusion imaging in acute ischemic stroke: from time to tissue,”
Stroke, 51
(3), 1017
–1024 https://doi.org/10.1161/STROKEAHA.119.028337 SJCCA7 0039-2499
(2020).
Google Scholar
R. Gonzales et al., Digital Image Processing, Chapter 10: Image Segmentation, 719
–727 3rd ed.Prentice-Hall(
(2008). Google Scholar
S. Winzeck et al.,
“ISLES 2016 and 2017-benchmarking ischemic stroke lesion outcome prediction based on multispectral MRI,”
Front. Neurol., 9 679 https://doi.org/10.3389/fneur.2018.00679
(2018).
Google Scholar
S. K. Thiyagarajan and K. Murugan,
“A systematic review on techniques adapted for segmentation and classification of ischemic stroke lesions from brain MR images,”
Wireless Pers. Commun., 118 1225
–1244 https://doi.org/10.1007/s11277-021-08069-z
(2021).
Google Scholar
Y. Zhang et al.,
“Application of deep learning method on ischemic stroke lesion segmentation,”
J. Shanghai Jiaotong Univ. (Sci.), 27 99
–111 https://doi.org/10.1007/s12204-021-2273-9
(2021).
Google Scholar
Y. LeCun, Y. Bengio and G. Hinton,
“Deep learning,”
Nature, 521
(7553), 436
–444 https://doi.org/10.1038/nature14539
(2015).
Google Scholar
G. Litjens et al.,
“A survey on deep learning in medical image analysis,”
Med. Image Anal., 42 60
–88 https://doi.org/10.1016/j.media.2017.07.005
(2017).
Google Scholar
Z. Akkus et al.,
“Deep learning for brain MRI segmentation: state of the art and future directions,”
J. Digital Imaging, 30
(4), 449
–459 https://doi.org/10.1007/s10278-017-9983-4 JDIMEW
(2017).
Google Scholar
A. S. Lundervold and A. Lundervold,
“An overview of deep learning in medical imaging focusing on MRI,”
Zeitschr. Med. Phys., 29
(2), 102
–127 https://doi.org/10.1016/j.zemedi.2018.11.002
(2019).
Google Scholar
O. Maier et al.,
“Classifiers for ischemic stroke lesion segmentation: a comparison study,”
PLoS One, 10
(12), e0145118 https://doi.org/10.1371/journal.pone.0145118 POLNCL 1932-6203
(2015).
Google Scholar
J. Nalepa, M. Marcinkiewicz and M. Kawulok,
“Data augmentation for brain-tumor segmentation: a review,”
Front. Comput. Neurosci., 13 83 https://doi.org/10.3389/fncom.2019.00083 1662-5188
(2019).
Google Scholar
J. Kleesiek et al.,
“Deep MRI brain extraction: a 3D convolutional neural network for skull stripping,”
NeuroImage, 129 460
–469 https://doi.org/10.1016/j.neuroimage.2016.01.024 NEIMEF 1053-8119
(2016).
Google Scholar
G. Wang et al.,
“Automatic ischemic stroke lesion segmentation from computed tomography perfusion images by image synthesis and attention-based deep neural networks,”
Med. Image Anal., 65 101787 https://doi.org/10.1016/j.media.2020.101787
(2020).
Google Scholar
A. Clerigues et al.,
“Acute ischemic stroke lesion core segmentation in CT perfusion images using fully convolutional neural networks,”
Comput. Biol. Med., 115 103487 https://doi.org/10.1016/j.compbiomed.2019.103487 CBMDAW 0010-4825
(2019).
Google Scholar
A. Tureckova and A. J. Rodríguez-Sánchez,
“ISLES challenge: U-shaped convolution neural network with dilated convolution for 3D stroke lesion segmentation,”
Lect. Notes Comput. Sci., 11383 319
–327 https://doi.org/10.1007/978-3-030-11723-8_32 LNCSD9 0302-9743
(2018).
Google Scholar
M. M. Bronstein et al.,
“Geometric deep learning: going beyond Euclidean data,”
IEEE Signal Process Mag., 34
(4), 18
–42 https://doi.org/10.1109/MSP.2017.2693418 ISPRE6 1053-5888
(2017).
Google Scholar
C. Gillmann et al.,
“Visualizing multimodal deep learning for lesion prediction,”
IEEE Comput. Graphics Appl., 41
(5), 90
–98 https://doi.org/10.1109/MCG.2021.3099881 ICGADZ 0272-1716
(2021).
Google Scholar
M. Fey et al.,
“SplineCNN: fast geometric deep learning with continuous B-spline kernels,”
in Proc. IEEE Conf. Comput. Vision and Pattern Recognit.,
869
–877
(2018). https://doi.org/10.1109/CVPR.2018.00097 Google Scholar
F. Monti et al.,
“Geometric deep learning on graphs and manifolds using mixture model CNNs,”
in Proc. IEEE Conf. Comput. Vision and Pattern Recognit.,
5115
–5124
(2017). Google Scholar
F. L. Ribeiro, S. Bollmann and A. M. Puckett,
“Predicting the retinotopic organization of human visual cortex from anatomy using geometric deep learning,”
NeuroImage, 244 118624 https://doi.org/10.1016/j.neuroimage.2021.118624 NEIMEF 1053-8119
(2021).
Google Scholar
K. Gopinath, C. Desrosiers and H. Lombaert,
“Graph convolutions on spectral embeddings for cortical surface parcellation,”
Med. Image Anal., 54 297
–305 https://doi.org/10.1016/j.media.2019.03.012
(2019).
Google Scholar
Y. Lu et al.,
“CNN-G: convolutional neural network combined with graph for image segmentation with theoretical analysis,”
IEEE Trans. Cognit. Dev. Syst., 13 631
–644 https://doi.org/10.1109/TCDS.2020.2998497
(2020).
Google Scholar
R. He et al.,
“Spectral graph transformer networks for brain surface parcellation,”
in IEEE 17th Int. Symp. Biomed. Imaging (ISBI),
372
–376
(2020). https://doi.org/10.1109/ISBI45749.2020.9098737 Google Scholar
L. Z. Williams et al.,
“Geometric deep learning of the human connectome project multimodal cortical parcellation,”
Lect. Notes Comput. Sci., 13001 103
–112 https://doi.org/10.1007/978-3-030-87586-2_11 LNCSD9 0302-9743
(2021).
Google Scholar
P. Besson et al.,
“Geometric deep learning on brain shape predicts sex and age,”
Comput. Med. Imaging Graphics, 91 101939 https://doi.org/10.1016/j.compmedimag.2021.101939 CMIGEY 0895-6111
(2021).
Google Scholar
X. Zhao et al.,
“Graph convolutional network analysis for mild cognitive impairment prediction,”
in IEEE 16th Int. Symp. Biomed. Imaging (ISBI 2019),
1598
–1601
(2019). https://doi.org/10.1109/ISBI.2019.8759256 Google Scholar
T. Dissanayake et al.,
“Geometric deep learning for subject-independent epileptic seizure prediction using scalp EEG signals,”
IEEE J. Biomed. Health. Inf., 26 527
–538 https://doi.org/10.1109/JBHI.2021.3100297
(2021).
Google Scholar
Z. Guo et al.,
“Deep LOGISMOS: deep learning graph-based 3d segmentation of pancreatic tumors on CT scans,”
in IEEE 15th Int. Symp. Biomed. Imaging (ISBI 2018),
1230
–1233
(2018). https://doi.org/10.1109/ISBI.2018.8363793 Google Scholar
A. G.-U. Juarez et al.,
“A joint 3D UNet-graph neural network-based method for airway segmentation from chest CTs,”
Lect. Notes Comput. Sci., 11861 583
–591 https://doi.org/10.1007/978-3-030-32692-0_67 LNCSD9 0302-9743
(2019).
Google Scholar
Y. Lu et al.,
“Graph-FCN for image semantic segmentation,”
Lect. Notes Comput. Sci., 11554 97
–105 https://doi.org/10.1007/978-3-030-22796-8_11 LNCSD9 0302-9743
(2019).
Google Scholar
H. Gao and S. Ji,
“Graph U-nets,”
IEEE Trans. Pattern Anal. Mach. Intell., 44
(9), 4948
–4960 https://doi.org/10.1109/TPAMI.2021.3081010
(2022).
Google Scholar
J. Long, E. Shelhamer and T. Darrell,
“Fully convolutional networks for semantic segmentation,”
in Proc. IEEE Conf. Comput. Vision and Pattern Recognit.,
3431
–3440
(2015). Google Scholar
O. Maier et al.,
“ISLES 2015-a public evaluation benchmark for ischemic stroke lesion segmentation from multispectral MRI,”
Med. Image Anal., 35 250
–269 https://doi.org/10.1016/j.media.2016.07.009
(2017).
Google Scholar
M. Kistler et al.,
“The virtual skeleton database: an open access repository for biomedical research and collaboration,”
J. Med. Internet Res., 15
(11), e245 https://doi.org/10.2196/jmir.2930
(2013).
Google Scholar
O. Ronneberger, P. Fischer and T. Brox,
“U-Net: convolutional networks for biomedical image segmentation,”
Lect. Notes Comput. Sci., 9351 234
–241 https://doi.org/10.1007/978-3-319-24574-4_28 LNCSD9 0302-9743
(2015).
Google Scholar
C. R. Qi et al.,
“PointNet++: deep hierarchical feature learning on point sets in a metric space,”
in Proc. Adv. Neural Inf. Process. Syst.,
5099
–5108
(2017). Google Scholar
K. Qi et al.,
“X-Net: brain stroke lesion segmentation based on depthwise separable convolution and long-range dependencies,”
Lect. Notes Comput. Sci., 11766 247
–255 https://doi.org/10.1007/978-3-030-32248-9_28 LNCSD9 0302-9743
(2019).
Google Scholar
I. S. Dhillon, Y. Guan and B. Kulis,
“Weighted graph cuts without eigenvectors a multilevel approach,”
IEEE Trans. Pattern Anal. Mach. Intell., 29
(11), 1944
–1957 https://doi.org/10.1109/TPAMI.2007.1115 ITPIDJ 0162-8828
(2007).
Google Scholar
C. R. Gillebert, G. W. Humphreys and D. Mantini,
“Automated delineation of stroke lesions using brain CT images,”
Neuroimage Clin., 4 540
–548 https://doi.org/10.1016/j.nicl.2014.03.009
(2014).
Google Scholar
F. Milletari, N. Navab and S.-A. Ahmadi,
“ISLES: fully convolutional neural networks for volumetric medical image segmentation,”
in Fourth Int. Conf. 3D Vision (3DV),
565
–571
(2016). Google Scholar
S. Ioffe and C. Szegedy,
“Batch normalization: accelerating deep network training by reducing internal covariate shift,”
in Int. Conf. Mach. Learn.,
448
–456
(2015). Google Scholar
Y. Jia, C. Huang and T. Darrell,
“Beyond spatial pyramids: receptive field learning for pooled image features,”
in IEEE Conf. Comput. Vision and Pattern Recognit.,
3370
–3377
(2012). https://doi.org/10.1109/CVPR.2012.6248076 Google Scholar
D. H. Hubel and T. N. Wiesel,
“Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex,”
J. Physiol., 160
(1), 106
–154 https://doi.org/10.1113/jphysiol.1962.sp006837
(1962).
Google Scholar
J. J. Koenderink and A. J. Van Doorn,
“The structure of locally orderless images,”
Int. J. Comput. Vis., 31
(2), 159
–168 https://doi.org/10.1023/A:1008065931878 IJCVEQ 0920-5691
(1999).
Google Scholar
K. You et al.,
“Co-tuning for transfer learning,”
in Adv. Neural Inf. Process. Syst. 33,
(2020). Google Scholar
C. Gillmann, D. Saur and G. Scheuermann,
“How to deal with uncertainty in machine learning for medical imaging?,”
in IEEE Workshop TRust and EXpertise in Vis. Anal. (TREX),
52
–58
(2021). https://doi.org/10.1109/TREX53765.2021.00014 Google Scholar
C. Gillmann et al.,
“Uncertainty-aware visualization in medical imaging: a survey,”
Comput. Graph. Forum, 40
(3), 665
–689 https://doi.org/10.1111/cgf.14333 CGFODY 0167-7055
(2021).
Google Scholar
G. Borgefors,
“Distance transformations in arbitrary dimensions,”
Comput. Vision Graphics Image Process., 27
(3), 321
–345 https://doi.org/10.1016/0734-189X(84)90035-5 CVGPDB 0734-189X
(1984).
Google Scholar
G. Schwarzband and N. Kiryati,
“The point spread function of spiral CT,”
Phys. Med. Biol., 50
(22), 5307 https://doi.org/10.1088/0031-9155/50/22/007
(2005).
Google Scholar
J. Cho et al.,
“How much data is needed to train a medical image deep learning system to achieve necessary high accuracy?,”
(2015). https://doi.org/10.48550/arXiv.1511.06348 Google Scholar
S. M. Plis et al.,
“Deep learning for neuroimaging: a validation study,”
Front. Neurosci., 8 229 https://doi.org/10.3389/fnins.2014.00229 1662-453X
(2014).
Google Scholar
C. C. Aggarwal, Neural Networks and Deep Learning, Springer International Publishing, Cham
(2018). Google Scholar
M. Ajčević et al.,
“A CT perfusion based model predicts outcome in wake-up stroke patients treated with recombinant tissue plasminogen activator,”
Physiol. Meas., 41 075011 https://doi.org/10.1088/1361-6579/ab9c70 PMEAE3 0967-3334
(2020).
Google Scholar
T. P. Lillicrap et al.,
“Short-and long-term efficacy of modafinil at improving quality of life in stroke survivors: a post hoc sub study of the modafinil in debilitating fatigue after stroke trial,”
Front. Neurol., 9 269 https://doi.org/10.3389/fneur.2018.00269
(2018).
Google Scholar
S. Agarwal et al.,
“Collateral response modulates the time–penumbra relationship in proximal arterial occlusions,”
Neurology, 90
(4), e316
–e322 https://doi.org/10.1212/WNL.0000000000004858 NEURAI 0028-3878
(2018).
Google Scholar
Y. Xie et al.,
“Spatial clockwork recurrent neural network for muscle perimysium segmentation,”
Lect. Notes Comput. Sci., 9901 185
–193 https://doi.org/10.1007/978-3-319-46723-8_22 LNCSD9 0302-9743
(2016).
Google Scholar
J. Yang and S. Segarra,
“Enhancing geometric deep learning via graph filter deconvolution,”
in IEEE Global Conf. Signal and Inf. Process. (GlobalSIP),
758
–762
(2018). Google Scholar
D. T. Huff, A. J. Weisman and R. Jeraj,
“Interpretation and visualization techniques for deep learning models in medical imaging,”
Phys. Med. Biol., 66 04TR01 https://doi.org/10.1088/1361-6560/abcd17 PHMBA7 0031-9155
(2021).
Google Scholar
Q. Teng et al.,
“A survey on the interpretability of deep learning in medical diagnosis,”
Multimedia Syst., 28 2335
–2355 https://doi.org/10.1007/s00530-022-00960-4 MUSYEW 1432-1882
(2022).
Google Scholar
M. S. Jabal et al.,
“Interpretable machine learning modeling for ischemic stroke outcome prediction,”
Front. Neurol., 13 884693 https://doi.org/10.3389/fneur.2022.884693
(2022).
Google Scholar
K. Ahmed, M. A. Gad and A. E. Aboutabl,
“Performance evaluation of salient object detection techniques,”
Multimedia Tools Appl., 81 21741
–21777 https://doi.org/10.1007/s11042-022-12567-y
(2022).
Google Scholar
B. Ghariba, M. S. Shehata and P. McGuire,
“Visual saliency prediction based on deep learning,”
Information, 10 257 https://doi.org/10.3390/info10080257
(2019).
Google Scholar
S. Nousias et al.,
“Deep saliency mapping for 3D meshes and applications,”
ACM Trans. Multimedia Comput. Commun. Appl., 19 1
–22 https://doi.org/10.1145/3550073
(2023).
Google Scholar
BiographyAriel Iporre-Rivas is a PhD candidate at Leipzig University and Max Planck Institute for Human Cognitive and Brain Sciences. He received his master’s degree in biomedical engineering from Heidelberg University in 2019. His current research interests include geometric deep learning, and brain image analysis. Dorothee Saur received her Dr.med. degree from Friedrich-Schiller University Jena in 2001, and her Dr. habil. degree in neurology from the Albrecht-Ludwigs University Freiburg in 2010. Since 2015, she has been a full professor and co-chair of the Department of Neurology at the University of Leipzig. Her research focuses on the loss and recovery of brain function after stroke, using neuroimaging techniques such as functional and structural MRI, fiber tractography, and machine learning. Karl Rohr studied electrical engineering at the University of Karlsruhe (KIT) and received his PhD and habilitation degree in computer science from the University of Hamburg. Currently, he is head of the Biomedical Computer Vision group and associate professor at Heidelberg University. He was program chair of the IEEE International Symposium on Biomedical Imaging (ISBI) 2016. His research interests are in biomedical image analysis with focus on segmentation, tracking, and image registration. Gerik Scheuermann received the master’s degree in mathematics in 1995 and the PhD in computer science in 1999, both from the TU Kaiserslautern. He is a full professor with the Leipzig University since 2004. He has co-authored more than 300 reviewed book chapters, journal articles, and conference papers. He has served as paper cochair for EuroVis 2008, IEEE-SciVis 2011, IEEE-SciVis 2012, and IEEE-PacificVis 2015. His current research interests include visualization and visual analytics, especially on feature and topology-based methods. Christina Gillmann received her PhD in computer science from the University of Kaiserslautern in 2018. She is currently a researcher of the Signal and Image Processing Group with the University of Leipzig, leading her own subgroup on uncertainty-aware visual analytics (UAVA) in medical applications. Her research interests include UAVA, medical visualization, medical image analysis, and uncertainty analysis in general. |
Education and training
Image segmentation
Ischemic stroke
Data modeling
Brain
Deep learning
Batch normalization