Skip to main content
Top
Published in: Network Modeling Analysis in Health Informatics and Bioinformatics 1/2023

Open Access 01-12-2023 | Original Article

Development of a cerebral aneurysm segmentation method to prevent sentinel hemorrhage

Authors: Yousra Regaya, Abbes Amira, Sarada Prasad Dakua

Published in: Network Modeling Analysis in Health Informatics and Bioinformatics | Issue 1/2023

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Image segmentation being the first step is always crucial for brain aneurysm treatment planning; it is also crucial during the procedure. A robust brain aneurysm segmentation has the potential to prevent the blood leakage, also known as sentinel hemorrhage. Here, we present a method combining a multiresolution and a statistical approach in two dimensional domain to segment cerebral aneurysm in which the Contourlet transform (CT) extracts the image features, while the Hidden Markov Random Field with Expectation Maximization (HMRF-EM) segments the image, based on the spatial contextual constraints. The proposed algorithm is tested on Three-Dimensional Rotational Angiography (3DRA) datasets; the average values of segmentation accuracy, DSC, FPR, FNR, specificity, and sensitivity, are found to be 99.72%, 93.52%, 0.07%, 5.23%, 94.77%, and 99.96%, respectively.
Notes
Abbes Amira and Sarada Prasad Dakua contributed equally to this work.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

1 Background

The sentinel clot sign is a useful computed tomography finding for the evaluation of probable anatomic sites of hemorrhage. Sentinel leaks produce sudden focal or generalized head pain that may be severe. Sentinel headaches precede the aneurysm rupture. To be further specific, cerebral Aneurysm (CA) is an abnormal inflammation within the brain blood vessels that is commonly detected after the age around forty (Higashida 2003; Nikravanshalmani et al. 2013). The mortality rate of this disease is between 30% and 40%, whereas the morbidity rate ranges between 35% and 60% (Li and Li 2010). Computer Aided Diagnosis (CAD) algorithms have been helpful in reducing the bias, time consumption, increasing the CA detection accuracy, etc. Wong (2005). Medical image segmentation (MIS) algorithms are important in the diagnosis phase.
There have been several methods in image segmentation that include both traditional methods and machine-learning/deep-learning methods. In deep learning (DL)-based methods, the neural networks (NN) used for image segmentation, have received much attention due to their end-to-end nature and state-of-the-art performance. Deep neural networks gained more popularity after their success at ImageNet Large Scale Visual Recognition Challenge 2012 (Russakovsky et al. 2015). With the improvement in GPU architectures and the development of dedicated deep learning libraries [e.g., Tensorflow (Abadi et al. 2016) and PyTorch (Paszke et al. 2019)], neural networks have been in use for a wide range of computationally heavy computer vision tasks (e.g., object recognition and segmentation). For image segmentation, the basic schematic diagram of a convolutional neural networks (CNNs) is given in Fig. 1; it comprises a series of convolution and pooling layers to condense the input image into a dense feature representation. The feature representation is utilized to reconstruct the segmentation mask using deconvolution and upsampling layers. During NN training, preprocessed images (e.g., normalized images with contrast enhancement) are fed to the network to generate the segmentation probabilities (forward propagation). A loss function computes the discrepancy between the NN prediction and the ground truth. Finally, the weights of the network are updated using an optimizer (e.g., Adam Kingma and Ba 2014) in the back-propagation phase. Altogether, the type preprocessing, architecture, choice of the loss function and optimizer, determine the segmentation accuracy and inference time of the NNs.
Over the years, although several CNNs (Khan et al. 2021; Siddique et al. 2021; Torres-Velazquez et al. 2021) have been proposed, the neural networks have faced several challenges too. For instance, the neural networks are susceptible to variations in data like changes, in contrast, noise, and resolution (Thambawita et al. 2021; Orlando et al. 2022). In a clinical setting, these variations are expected in the data due to multiple machines with several acquisition parameters that can cause the data distribution to change. One technical limitation for training the neural networks is due to the limited quantity of clinical data, resulting in overfitting (i.e., poor generalizability). Additionally, the training procedure of the neural networks does not provide any convergence guarantees. Other technical challenges include the black-box nature of deep learning-based approaches, which downplays the reliability of the neural networks in clinically sensitive settings. Thus in this paper, we propose a traditional/conventional algorithm to segment the CA from 3DRA datasets using a multiresolution statistical approach, where the Contourlet Transform (CT), in conjunction with Hidden Markov Random Field, is used to segment 2D images in the Contourlet domain.
There is rich literature in conventional MIS that includes semi-automatic and automatic (Dakua 2013; Dakua and Sahambi 2011). Broadly, the semi-automatic approaches include threshold-based (Mitra and Chandra 2013), model-based (Dakua et al. 2019), graph-based (Dakua and Sahambi 2010, 2011) methods that need human intervention (Chen et al. 2014), which is tedious and prone to operator variability; these factors certainly affect the segmentation accuracy. Therefore, automatic approaches were introduced. In Mitra and Chandra (2013), spatial filtering and dual global thresholding are considered, where three parameters (i.e. two threshold values and filter mask size) are user-defined. Yang et al. present a dot enhancement filter that includes thresholding, and a region growing technique. However, it lacks the inconsistency of the sensitivity, where it ranges between 80% and 95% for CAs larger than 5 mm and between 71% and 91% otherwise. Bogunovic et al. (2011) present a Geodesic Active Regions (GAR) based algorithm. However, there is the possibility of merging either a vessel with a CA or two vessels together. This partly happens either because of the insufficiency of the imaging resolution and the low blood flow in the small vessels. In Hentschke et al. (2011), a blobness filter and a k-means technique are adapted. The algorithm results in a false positive rate reaching 20.8%. Jerman et al. (2015), present a blobness enhancement filter combined with Random Forest Decision (RFD) classifier and a grow-cut technique, where it is assumed that a saccular CA consists of more than 15 voxels. Thus, this assumption reduces the chances to detect small CAs. Suniaga et al. (2012), propose a fuzzy logic in a level set framework along with a SVM classifier to segment saccular CAs. In Zhang et al. (2012), a method on Conditional Random Field (CRF) and gentle Adaboost classifier is proposed that is trained on six datasets and tested on two datasets. However, machine Learning (specifically, Deep Learning (DL)) models need extensive data to capture the underlying distribution for generalization in real-world systems. Additionally, the black box nature of the DL models adds uncertainty to their predictions, thus limiting their use as a second opinion in clinical setting (Su et al. 2019). Lu et al. (2016) present a method, where multi-scale filtering is used to enhance the vessels suppressing the noise and a mixture model is built to fit the histogram curve of the preprocessed data. Finally, expectation maximization is used for parameter estimation. Kai et al. (Lawonn et al. 2019) present a graph-cut based method for aneurysm segmentation. Yu et al. (Chen et al. 2020) present a geodesic active contour (GAC) and Euler’s elastica model-based method for aneurysm segmentation, where GAC segments the giant aneurysms in high contrast regions and the elastica model estimates the missing boundaries in low contrast regions. Thus, the numerous automatic segmentation algorithms reported so far in the literature have potential limitations that warrant further research on automatic methods. However, the automatic methods are difficult to be controlled, as usually desired by the clinicians during their treatment planning to explore various options. Thus, this paper proposes a semi-automatic algorithm combining Contourlet Transform (CT) and Hidden Markov Random Field with Expectation Maximization (HMRF-EM) to segment CA regardless the CA shape or size.
The rest of this paper is organized as follows: in Sect. 3, the foundation and mathematical background for the proposed algorithm are presented. In Sect. 4, the proposed algorithm is discussed in detail. In Sect. 5, the objective and subjective evaluation of the proposed work are presented along with the dataset description and the environmental setup for the implementation. Section 6 includes discussion and some future work, whereas the Sect. 7 concludes the paper.

3 Mathematical background

3.1 Preprocessing

Noise is a prime obstacle in image segmentation and it is inherent in medical data. Thus, noise reduction is necessary to improve the segmentation outcome. Traditionally, filtering has been commonly used, however, it results in significant degradation in the input image quality. Stochastic resonance (SR) concept is an emergent one that constructively use the noise present in input data (Dakua et al. 2019). Stochastic resonance occurs when the ratio of signal to noise ration attains the maximum. At this state, the input/output correlation is maximum at a certain noise level. To exhibit SR, a system should possess three basic properties: a non-linearity in terms of threshold, sub-threshold signals like signals with small amplitude and a source of additive noise. This phenomenon occurs frequently in bistable systems or in systems with threshold-like behavior. The general behavior of SR mechanism shows that at lower noise intensities the weak signal is unable to cross the threshold, thus giving a very low SNR. For large noise intensities the output is dominated by the noise, also leading to a low SNR. But, for moderate noise intensities, the noise allows the signal to cross the threshold giving maximum SNR at some optimum additive noise level. Thus, a plot of SNR as a function of noise intensity shows a peak at an optimum noise level (shown in Fig. 2). There have already been many attempts to use SR in different domains such Fourier and spatial domains; however, we have chosen the maximal overlap discrete wavelet transform (MODWT) because of some of its key advantages: (1) MODWT can handle any sample size, (2) the smooth and detail coefficients of MODWT multiresolution analysis are associated with zero phase filters, (3) it is transform invariant, and (4) it produces a more asymptotically efficient wavelet variance estimator than DWT.

3.1.1 Maximal overlap discrete wavelet transform

Generally, DWT is defined by: \({\psi _{j,k}}\left( t \right) = {2^{\frac{j}{2}}}\psi \left( {{2^j}t - k} \right) \quad j,k \in {\text {Z}};\;z = \left\{ {0,1,2...} \right\}\), where \(\psi\) is a real valued function compactly supported, and \(\int _{ - \infty }^\infty {\psi \left( t \right) \text {d}t = 0}\). MODWT is evaluated using dilation equations: \(\phi \left( t \right) = \sqrt{2} \sum \nolimits _k {{l_k}} \phi \left( {2t - k} \right) ,\quad \psi \left( t \right) = \sqrt{2} \sum \nolimits _k {{h_k}} \phi \left( {2t - k} \right)\), where \(\phi \left( {2t - k} \right)\) and \(\psi \left( t \right)\) are father wavelet defining low pass filter coefficients and mother wavelet defining high pass filter coefficients \(l_k\): \({l_k} = \sqrt{2} \int _{ - \infty }^\infty {\phi \left( t \right) \phi \left( {2t - k} \right) \text {d}t,\quad {h_k} = \sqrt{2} \int _{ - \infty }^\infty {\psi \left( t \right) \psi \left( {2t - k} \right) \text {d}t} }\).

3.1.2 Denoising by MODWT

In this methodology, 2D MODWT is applied to the \(M \times N\) size image I. Applying SR to the approximation and detail coefficients, the stochastically enhanced (tuned) coefficient sets in MODWT domain are obtained as \(W_\psi ^s \left( {l,p,q} \right) _{SR}\) and \(W\left( {l_0,p,q} \right) _{SR}\). The SR in discrete form is defined as: \(\frac{{\text {d}x}}{{\text {d}t}} = \left[ {ax - ex^3 } \right] + B\sin \omega t + \sqrt{D} \xi \left( t \right)\), where \(\sqrt{D} \xi \left( t \right)\) and \(B\sin \omega t\) represent noise and input, respectively; these are replaced by MODWT sub-band coefficients. The noise term is the factor to produce SR; maximization of SNR occurs at the double well parameter a. Implementation of SR on digital images necessitates the need for solving the stochastic differential equation using Euler–Maruyama’s method (Dakua et al. 2016) that gives the iterative discrete equation:
$$\begin{aligned} {x(n+1)} = {x(n)}+\Delta {t}\left[ (ax(n)-\text {e}x^{3}(n))+{Input}(n)\right] \end{aligned}$$
(1)
where a and e are the bistable parameters, whereas n and \(\Delta {t}\) represent iteration and sampling time, respectively. Input denotes the sequence of input signal and noise, with the initial condition being \(x(0)=0\). The final stochastic simulation is obtained after some predefined number of iterations. Given the tuned (enhanced and stabilized) set of wavelet coefficients (\(X_\phi \left( {l_0,p,q} \right)\) and \(X_\psi ^s \left( {l,p,q} \right)\)), the denoised image \(I_\textrm{denoised}\) in spatial domain is obtained by inverse maximal overlap discrete wavelet transform (IMODWT) as:
$$\begin{aligned} \begin{array}{*{20}c} {I_\textrm{denoised} = \frac{1}{{\sqrt{MN} }}\sum \limits _p {\sum \limits _q {X_\phi \left( {l_0,p,q} \right) \phi _{l_0,p,q} \left( {i,j} \right) } } } \\ \nonumber \quad +{\frac{1}{{\sqrt{MN} }}\sum \limits _{s \in \left( {H,V,D} \right) } {\sum \limits _{l = l_0 } {\sum \limits _p {\sum \limits _q {X_\psi ^s \left( {l,p,q} \right) \psi _{l_0,p,q}^s \left( {i,j} \right) } } } } } \\ \end{array} \end{aligned}$$
The double well parameters a and e are determined from the SNR by differentiating SR with respect to a and equating to zero; in this way, SNR is maximized resulting in \(a = 2\sigma _0 ^2\) for maximum SNR, where \(\sigma _0\) is the noise level administered to the input image. The maximum possible value of restoring force (\(R = B\sin \omega t\)) in terms of gradient of some bistable potential function U(x), \(R = - \frac{{\text {d}U}}{{\text {d}x}} = - ax + ex^3,\frac{{\text {d}R}}{{\text {d}x}} = - a + 3ex^2 = 0\) resulting \(x = {\sqrt{a\big / 3e}}\). R at this value gives maximum force as \(\sqrt{\frac{{4a^3 }}{{27e}}}\) and \(B\sin \omega t < \sqrt{\frac{{4a^3 }}{{27e}}}\). Maximizing the left term (keeping \(B=1\)), \(e < \frac{{4a^3 }}{{27}}\). To get the parameter values, we consider \(a =w \times 2\sigma _0 ^2\), and \(e=z \times \sqrt{\frac{{4a^3 }}{{27}}}\); w and z are weight parameters for a and e. Initially, w is an experimentally chosen constant that later becomes input image standard deviation dependent, while z is a number less than 1 to ensure sub-threshold condition of the signal. In this way, the noise in input image is countered and maximum information from the image is achieved. This enhanced image is used subsequently.

3.2 Contourlet transform

Multiresolution analysis techniques usually utilize the image features for computer vision. There have been several multiresolution analysis techniques such as Wavelet, Ridgelet, Curvelet, and Contourlet transforms for medical image segmentation (AlZubi et al. 2011; Po and Do 2006; Moayedi et al. 2010). We have preferred the Contourlet Transform (CT) in our proposed approach due to some of its advantages (Do and Vetterli 2003). CT being an ideal 2D transform in the discrete domain has other salient features: multiresolution, localization, critical sampling, anisotropy, and multiple directions for different resolutions. In addition, this transformation provides a sparse representation saving a significant amount of memory by offering simple and fast data processing as it requires O(N) operations for an image with N-pixels (Po and Do 2006). These characteristics capture the geometrical smoothness of the 2D contours.
The Pyramidal Directional Filter Bank (PDFB) (Do and Vetterli 2001) combines the Laplacian Pyramid (LP) and Directional Filter Bank (DFB) to extract the desirable fine details of CT. LP (Burt and Adelson 1983) allows the multiresolution representation of an image to capture point singularities by removing the noise. DFB (Bamberger and Smith 1992) decomposes an image into multiple directions to capture high-frequency content as smooth contours and directional edges by formulating the captured point singularity into a linear structure.
Briefly, the Contourlet works as follow: First, the image, \(a_{0}[n]\), is passed to the LP filter to produce two images as output: a coarse/approximated/lowpass image, \(a_{1}[n]\), and a fine/detailed/bandpass image, \(b_{1}[n]\). The latter image (bandpass) is passed to the DFB to produce \(2^{L_{j}}\) bandpass directional images, \(c^{L_{j}}_{j,k}[n]\). Subsequently, the lowpass image, \(a_{1}[n]\), is passed again through LP to repeat the same process until a certain predefined number of decomposition levels, \(L_{j}\), is reached. Figure 3 illustrates the CT process to decompose a \(512 \times 512\) image into two levels, where eight and four directions are applied at each level, respectively.
CT has the adeptness at capturing geometrical smoothness of 2D contours and anisotropy in the discrete domain. In addition, it has a high degree of directionality as it allows to define different directions for different scales, which is not possible in other multiresolution analysis techniques. These advantages help extract the features from images resulting in improved segmentation.
To summarize, the CT takes a 2D image, \(a_{0}[n]\), as an input and decomposes it into a lowpass sub-band, \(a_{j}[n]\), and several bandpass directional sub-bands, \(c^{L_{j}}_{j,k}[n]\), which are called as the Contourlet coefficients. This process can be expressed by:
$$\begin{aligned} a_{j}[n]= & {} f, \theta ^{(L)}_{j,k,n} \longrightarrow \theta ^{(L)}_{j,k,n} = \sum _{n \in Z^{d}} g_{k}[n] \phi _{j,k}(t) \end{aligned}$$
(2)
$$\begin{aligned} c^{L_{j}}_{j,k}[n]= & {} f, \rho ^{(L)}_{j,k,n} \longrightarrow \rho ^{(L)}_{j,k,n} = \sum _{n \in Z^{d}} d_{k}[n] \varphi _{j,n}(t) \end{aligned}$$
(3)
where \(\theta ^{(L)}_{j,k,n}\) is the LP basis function for scale decomposition composed of the lowpass filter, \(g_{k}[n]\), and the scaling function, \(\phi _{j,k}(t)\); and \(\rho ^{(L)}_{j,k,n}\) is the DFB basis function for directional decomposition composed of orthogonal filter, \(d_{k}[n]\), and the directional function, \(\varphi _{j,n}(t)\). The parameters j, k, d, and n, used in Eqs. (2) and (3), are number of levels/scales, number of directions for each level, dimensionality (in our case, it is 2 since we are working in 2D domain), and a scale parameter along the frequency axis, respectively. The overall decomposition process of the CT is provided in Algorithm 1.
The number of the decomposition levels is experimentally selected since it depends on the image size and the amount of the details to be highlighted. For the number of directions, it is recommended to gradually increase them by \(2^k\).

3.3 Hidden Markov random field model

HMRF model is a statistical approach in stochastic domain that provides prior knowledge to simplify the segmentation task.
HMRF model segments the medical images based on the spatial correlation between the neighboring pixels. Some important notions about this model are:
  • Random field: The random variables are the intensity levels in an image. In HMRF model, two random fields exist:
    • Hidden random field: \(X = {(x_{1},x_{2},..,x_{N}) \Vert x_{i} \in L, i \in S}\) is a random field in a finite state space, L, and indexed by a set, S, with respect to a neighboring system of size, N. The state of this field X is unobservable/hidden and every \(x_{i}\) is independent of all other \(x_{j}\). The objective of this assumption is to classify each pixel independently (Marroquin et al. 2003).
    • Observable random field: \(Y = \{y = (y_{1},y_{2},..,y_{N}) \Vert y_{i} \in D, i \in S\}\) is a random field in a finite space, D, and indexed by a set, S, with respect to a neighboring system of size, N. This random field, Y, is observable and can only be defined with respect to X, where \(y_{i}\) follows a conditional probability distribution given any particular configuration \(x_{i} = l\): \(p(y_{i}\Vert l) = \{f(y_{i}; \theta _{l}), \ \forall l \in L \}\), where \(\theta _{l}\) is the set of the parameters that are involved.
  • Parameters: The set of parameters, \(\theta _{l}\), are unknown. Therefore, a model fitting solution is adopted to estimate them. In our context, the parameters are mainly the mean, \(\mu\), and the standard deviation, \(\sigma\).
  • Conditional independence: The two random fields, (XY), are conditionally independent.
  • Neighborhood system: It is a way to define the surrounding pixels for a specific pixel (Chen et al. 2011).
  • Clique: It is a subset of pixels, where every pair of distinct pixels are neighbors. A value is assigned to each clique, c, to define the clique potential \(V_{c}(x)\), where the sum of all of these values results in the energy function, U(x), that we aim to minimize.
    $$\begin{aligned} U(x) = \sum _{c \in C} V_{c}(x) \end{aligned}$$
    (4)
To know \(\theta _{l}\), EM algorithm is preferred. The HMRF-EM framework (Zhang et al. 2001) incorporates the EM algorithm in association with HMRF model to estimate the parameters and segment using iterative updates. The framework starts by initializing both: the segmentation and parameters (means, \(\mu\), and standard deviations, \(\sigma\)). Then, iteratively, it goes through the Expectation Step (E-Step) and Maximization Step (M-Step) to update the parameters and the initial segmentation until no further development is observed.
The E-Step updates the segmentation by assigning to each pixel an estimated class label, \(\hat{x}\), from a set of labels, L. The assignment is performed based on the MAP criterion, which tries to minimize the posterior energy using the current parameters estimate; during the energy maximization, the conditional posterior probability distribution, \(P(Y\Vert X)\), gets maximized. Equation (5) illustrates the energy calculation:
$$\begin{aligned} \hat{x} = \arg \min (\, U(y\Vert x) + U(x) \,) \end{aligned}$$
(5)
where U(x) is the energy function as illustrated in Eq. (4) and \(U(y\Vert x)\) is the likelihood energy illustrated below in Eq. (6).
$$\begin{aligned} U(y\Vert x) = \sum _{i \in S} \, \left[ \, \frac{ (y_{i} - \mu _{x_i})^2 }{ 2\sigma ^2 } + \log (\sigma _{x_i}) \, \right] \end{aligned}$$
(6)
While the M-Step updates the parameters based on ML criterion, which tries to maximize the expected likelihood found in the E-Step. The parameters \(\mu\) and \(\sigma\) are calculated using the Eqs. (7) and (8), respectively.
$$\begin{aligned} \mu= & {} \frac{ \sum _{i \in S} P^{(l)}(l\Vert y_{i})y_{i}}{\sum _{i \in S} P^{(l)}(l\Vert y_{i}) } \end{aligned}$$
(7)
$$\begin{aligned} \sigma= & {} \sqrt{ \frac{ \sum _{i \in S} P^{(l)}(l\Vert y_{i}) (y_{i}-\mu )^{2}}{\sum _{i \in S} P^{(l)}(l\Vert y_{i}) } } \end{aligned}$$
(8)
This framework works well for small data dimensions; the main advantages are, it: is easy to implement, provides an accurate segmentation, and is less sensitive to noise as compared to other segmentation techniques since it also considers the contextual information (Yazdani et al. 2015).

3.4 Modifications to conventional k-means clustering

This is well-known that k-means is a clustering technique maximizing the similarity of intra-class and minimizing the similarity of inter-class; it is computationally fast. In our method, we initialize k automatically based on the image entropy, a statistical measure of randomness that can be used to characterize the texture of the gray-scale image (Liang et al. 2012). It is expressed in Eq. (9):
$$\begin{aligned} k=- \,\sum ^{n-1}_{i=0} \, P(x_{i})\log _{2}P(x_{i}) \end{aligned}$$
(9)
where X is a vector of all intensities of an image and n is the number of pixels.

4 Proposed segmentation algorithm

The proposed CA segmentation algorithm is illustrated in Fig. 4.
The algorithm starts by feeding the input 2D scans after enhancement. The selection of the Region of Interest (ROI) from the entire cerebral vasculature is done manually. Subsequently, the below steps are performed on each 2D image.
During the first step, CT is applied to extract features from the input image by decomposing it into six pyramidal levels and different directions for each level, where the number of directional decomposition at each pyramidal level (from coarse to fine) are: \(2^{2}\), \(2^{2}\), \(4^{2}\), \(4^{2}\), \(8^{2}\), and \(8^{2}\) (Hiremath et al. 2010; Po and Do 2006). As discussed earlier, CT consists of two main filters, LP and DFB. A ladder filter, known as PKVA filter (Phoong et al. 1995), is selected for the first filter. The PKVA filter is effective to localize edge direction as it reduces the inter-direction mutual information (Po and Do 2006). As for the second filter, 9-7 bi-orthogonal Daubechies filter (Cohen and Daubechies 1993) is selected; this filter significantly reduces the inter-scale, inter-location, and inter-direction mutual information of the Contourlet (Po and Do 2006). After the decomposition by CT, the lowpass subband image, \(a_{j}[n]\), is used to perform the rest of the steps.
To apply the second step (HMRF-EM) of the segmentation algorithm, two prior steps need to be performed. First, an initial segmentation is generated using the k-means method, which is known with its under-estimation to complement the HMRF-EM framework (AlZubi et al. 2011). In addition, a Canny edge filter (Gonzalez and Wood 2002) is applied to highlight the image edges. After that, the HMRF-EM algorithm iterates between the E-Step and M-Step to enhance the initial segmentation constrained by the Canny edge detector.
The HMRF-EM method starts after getting the initial segmentation, \(\hat{x}^{(0)}\), and initial parameters, \(\theta ^{(0)}\), obtained by the k-means clustering, the constrained image, \(ce_{j}[n]\), obtained by the Canny edge operator, and the lowpass subband image, \(a_{j}[n]\), obtained by the Contourlet decomposition. During this step, the algorithm iterates between the steps E and M to refine the initial segmented image, constrained by the Canny segmented image, resulting in the final segmented 2D image by minimizing the posterior energy function as explained in Sect. 3.3.
As the last step, Inverse Contourlet Transform (ICT) is applied to reconstruct the image. Here, the lowpass subband image, \(a_{j}[n]\), the coarsest Contourlet coefficients, is replaced by the final segmented image, \(\hat{x}^{(MAP_{\_}itr)}\). The ICT is achieved using the same filters as in the decomposition stage, where the 9-7 and PKVA filters are used for the LP and DFB, respectively.
After completing these two phases, the reconstruction of all the segmented 2D images is performed to get the final segmented 3D volume of the ROI. The pseudocode for the overall proposed CA segmentation algorithm is presented in Algorithm 3.

5 Datasets and results

5.1 Datasets

The aneurysms obtained from the respective 3DRA scanner (Siemens machine) were small (\(\le\) 3 mm). Six 3DRA datasets are provided by Hamad Medical Cooperation (HMC) to validate the proposed segmentation; each dataset consists of 385 2D slices of size \(512 \times 512\) each. In addition, the corresponding ground truth datasets were obtained from the same hospital that were prepared by three neuro-experts.

5.2 Results

For preprocessing/denoisng of the input images, the initial values of \(\Delta {t}\) and z are taken as 0.007 and 0.000027, respectively. The results are provided in Fig. 5. To determine the quality of the denoised image, we have calculated distribution separation measure (DSM) that estimates the degree of image quality. The DSM is defined as [?]: DSM = \(\Vert {\mu _T^E - \mu _B^E}\Vert -\Vert {\mu _T^O - \mu _B^O}\Vert\), where \(\mu _T^E\) and \(\mu _T^O\) are the mean of the selected target regions of the denoised and original images, respectively; \(\mu _B^E\) and \(\mu _B^O\) are the mean of the selected background region of the denoised and original image, respectively. Higher the value of DSM, better is the quality. It is observed that the value of DSM is maximum at iteration 200 and then it starts decreasing, therefore, this iteration is considered as the optimal. These denoised data are used subsequently as the input volumetric data. We have now two scenarios: (1) to segment the region of interest, and (2) the entire input volumetric data. in scenario 1, we manually, select the slices to get the ROI and it is, on average, 86 slices and they are continuous, whereas in scenario 2, we feed all the slices as input to the algorithm. Both the quantitative and qualitative results are obtained by comparing the segmented volumes with the respective ground truth volume, as elaborated in Sects. 5.2.2 and 5.2.3, respectively. Figures 6, 7, 8, and 9 depict each dataset before and after applying the segmentation. To test, if the results are statistically significant, we set the significance level, \(\alpha\) = 0.05 and we obtained the p-values smaller than 0.05, indicating that these segmentation results correctly identified the brain aneurysm.

5.2.1 Registration

We have also considered image registration in our work to enable the comparison between the segmented volume and the corresponding ground truth. In this process, one of the images is defined as the target (or the subject), which we wish to apply a transformation, while the other image is defined as the reference (or the source). In our case, the target image is the segmented ROI volume; while the reference image is the ground truth ROI volume. The target image is transformed by means of the selected mapping functions to align it with the reference image (Zitova and Flusser 2003). We have selected an affine transformation; Fig. 10c illustrates the coordinate systems of the segmented volume and the ground truth data before and after registration.

5.2.2 Objective evaluation

Six performance metrics are used to measure the proposed segmentation quantitatively: Dice Similarity Index (DSI), sensitivity, specificity, accuracy, False Positive Ratio (FPR), and False Negative Ratio (FNR). The value of these metrics ranges between 0 and 1 (but we have determined their percentage value). Table 1 provides the definition and the formula of each metric.
Table 1
The four adopted performance metrics for the quantitative evaluation
Performance
Definition
Equation
Metric
  
Accuracy
Correctness of the overall
\(\displaystyle \frac{ \text {TP} + \text {TN} }{ \text {TP} + \text {TN} + \text {FP} + \text {FN} }\)
Segmentation
DSI
Amount of overlap between
\(\displaystyle \frac{ 2 \times \text {TP} }{ 2 \times \text {TP} + \text {FP} + \text {FN} }\)
The two segmentation
FPR
Number of pixels incorrectly
\(\displaystyle \frac{ \text {FP} }{ \text {FP} + \text {TN} } = 1 - \text {Specificity}\)
Segmented
FNR
Number of pixels incorrectly
\(\displaystyle \frac{ \text {FN} }{ \text {FN} + \text {TP} } = 1 - \text {Sensitivity}\)
rejected
Sensitivity
Number of pixels segmented
\(\displaystyle \frac{ \text {TP} }{ \text {TP} + \text {FN} }\)
Correctly
Specificity
Number of pixels excluded
\(\displaystyle \frac{ \text {TN} }{ \text {TN} + \text {FP} }\)
Correctly
Two measures, True Positive (TP) and True Negative(TN), in Table 1 indicate a correct segmentation, while False Positive (FP) and False Negative (FN) indicate an incorrect segmentation. Figure 11 depicts the meaning of each measure more clearly.
The values of these performance metrics are presented in Table 2 against each dataset.
Table 2
Results of objective evaluation
Measure
Dataset 1
Dataset 2
Dataset 3
Dataset 4
Dataset 5
Dataset 6
Average
Accuracy (%)
99.99
99.92
99.78
99.75
99.28
99.15
\({\text {99.72}}\)
DSC (%)
97.46
93.15
94.13
89.56
94.16
89.59
\({\text {93.52}}\)
FPR (%)
0.01
0.06
0.17
0.08
0.16
0.04
\({\text {0.07}}\)
FNR (%)
1.16
3.58
2.8
13.56
3.50
8.67
\({\text {5.23}}\)
Sensitivity (%)
98.84
96.42
97.20
86.44
97.95
86.99
\({\text {94.77}}\)
Specificity (%)
99.99
99.94
99.83
. 99.92
99.77
99.84
\({\text {99.96}}\)

5.2.3 Subjective evaluation

Each segmentation has been assessed visually by five observers (neurologists). The score, ranging between 0 and 5, is assigned by each observer, where 5 means that the ground truth volume almost completely matches with segmented one and 0 means that the two volumes do not match. Table 3 presents the observations.
Table 3
Results of subjective evaluation
Observation by
Dataset 1
Dataset 2
Dataset 3
Dataset 4
Dataset 5
Dataset 6
Average
Observer 1
5
4
4
3
4
4
4.00
Observer 2
5
4
5
3
5
4
4.41
Observer 3
5
4
5
4
5
3
4.16
Observer 4
5
4
4
4
4
4
4.08
Observer 5
5
4
5
3
5
3
4.08
Average
      
4.14

6 Discussion

Both the qualitative and quantitative results are promising when tested on these 3DRA datasets. In the quantitative evaluation, the average values of accuracy, DSC, FPR, FNR, specificity, and sensitivity, are \(99.72\%\), \(93.52\%\), \(0.07\%\), \(5.23\%\), \(94.77\%\), and \(99.96\%\), respectively; an average of 4.14 over 5 is obtained in the qualitative evaluation.
It may be observed in Tables 2 and 3 that the last dataset has the worst results as compared to the remaining datasets in both the quantitative and qualitative evaluation. This may be due to the fact that the provided ground truth data does not involve the complete brain vessels tree and only a delineated ROI is provided, where some surrounding vessels are excluded. The same may be observed from Fig. 12.
The computational time to segment a CA volume is considerably fast. Table 4 reports the running time of the proposed segmentation algorithm for the ROI in seconds (s) and the whole volume of a subject in minutes (min). Even though the results are acceptable, the computational time can further be reduced by using Field-Programmable Gate Array (FPGAs) or Graphics Processing Unit (GPUs).
Table 4
Time consumption to segment CA using the proposed approach
Region
Dataset 1
Dataset 2
Dataset 3
Dataset 4
Dataset 5
Dataset 6
Average
ROI (s)
32.12
28.56
93.46
57.56
93.61
58.41
68.31
(51 slices)
(51 slices)
(161 slices)
(80 slices)
(155 slices)
(88 slices)
(110 slices)
Volume (min)
4.38
3.21
3.30
5.31
3.90
5.87
4.60
(385 slices)
(385 slices)
(385 slices)
(385 slices)
(385 slices)
(385 slices)
(385 slices)
We have also compared the performance of the proposed method with some similar methods that have been published recently. The results are provided in Table 5. The results show that the performance of the proposed method is better than the others.
Table 5
Comparative performance of segmentation accuracy with similar methods
Method
Accuracy
DSC
FPR
FNR
Sensitivity
Specificity
PCA-based Dakua et al. (2016)
94.27
91.10
0.16
5.52
87.81
93.04
Global optimization Lawonn et al. (2019)
94.98
92.08
0.15
5.45
90.42
94.62
Elastica Chen et al. (2020)
95.07
92.91
0.12
5.33
92.20
97.24
Geometric Rai et al. (2021)
95.11
92.93
0.12
5.31
93.10
97.38
Proposed method
99.72
93.52
0.07
5.23
94.77
99.96
Additionally, we have compared (in Table 6) the proposed method with some popular DL-based methods, including Voxel-Morph (VM) (Balakrishnan et al. 2019), LT-Net (Wang et al. 2020), and Symmetric Normalization (SyN) (Avants et al. 2008). We have selected these methods because they have been regularly preferred by the research fraternity for comparison purpose. Among these, Voxel-Morph is probably the most famous method in recent years. The results show that the proposed method fairly performs as compared to the DL-based methods although the margin is not very significant. Furthermore, we have also compared the computational complexity involved in Table 7.
Table 6
Comparative performance of segmentation accuracy with DL-based techniques
Method
Accuracy
DSC
FPR
FNR
Sensitivity
Specificity
SyN Avants et al. (2008)
92.42
87.86
0.19
6.12
87.12
92.93
LT-Net Wang et al. (2020)
94.57
88.98
0.16
5.76
89.23
94.12
VoxelMorph Balakrishnan et al. (2019)
99.07
92.98
0.11
5.37
94.01
99.14
Dense-PSP-Unet Ansari et al. (2022)
99.09
92.99
0.10
5.35
94.04
99.15
Proposed method
99.72
93.52
0.07
5.23
94.77
99.96
Table 7
Performance comparison of proposed models on GPU using parameter count and model size
Model name
Parameter count
Model size (MB)
SyN Avants et al. (2008)
32,478,291
357.0
LT-Net Wang et al. (2020)
29,478,291
331.2
VoxelMorph Balakrishnan et al. (2019)
31,024,163
349.0
Dense-PSP-Unet Ansari et al. (2022)
13,882,087
143.0
Proposed method
13,289,122
143.0

7 Conclusions

Sentinel headache is often assumed to portend an increased risk of delayed cerebral ischemia and aneurysm rebleeding. High rates of morbidity and mortality are associated with Sub-Arachnoid Hemorrhage (SAH) that is caused by a ruptured CA. Therefore, it is quite imperative to detect and diagnose CAs at an early stage. In this paper, we have presented a semi-automatic CA segmentation method using Contourlet transform and the hidden Markov random field model. Prior to use the method, we have denoised the input data to improve the contrast of the input images and make them suitable for segmentation. We have obtained promising results when tested on 3DRA datasets. Since, it is usually difficult to obtain publicly available original and corresponding ground truth data, we had rely on HMC data. The preparation of ground truth data needs a huge effort and thus, we have used only a small number of datasets. In future, we would like to increase the number of datasets and make the segmentation algorithm further robust. Furthermore, since this segmentation algorithm could be used to improve the cerebral aneurysm treatment planning, we also plan to take this into clinical routine after thorough testing and validation.

Acknowledgements

Not applicable.

Declarations

Conflict of interest

The authors declare that they have no competing interests.
Not applicable.
Not applicable.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Literature
go back to reference Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, Devin M, Ghemawat S, Irving G, Isard M, Kudlur M, Levenberg J, Monga R, Moore S, Murray DG, Steiner B, Tucker P, Vasudevan V, Warden P, Wicke M, Yu Y, Zheng X (2016) Tensorflow: a system for large-scale machine learning. In: Proceedings of the 12th USENIX conference on operating systems design and implementation. OSDI’16. USENIX Association, USA, pp 265–283 Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, Devin M, Ghemawat S, Irving G, Isard M, Kudlur M, Levenberg J, Monga R, Moore S, Murray DG, Steiner B, Tucker P, Vasudevan V, Warden P, Wicke M, Yu Y, Zheng X (2016) Tensorflow: a system for large-scale machine learning. In: Proceedings of the 12th USENIX conference on operating systems design and implementation. OSDI’16. USENIX Association, USA, pp 265–283
go back to reference Bogunovic H, Pozo JM, Villa-Uriol MC, Majoie CB, van den Berg R, Gratama van Andel HA, Macho JM, Blasco J, San Roman L, Frangi AF (2011) Automated segmentation of cerebral vasculature with aneurysms in 3dra and tof-mra using geodesic active regions: an evaluation study. Med Phys 38(1):210–222. https://doi.org/10.1118/1.3515749CrossRef Bogunovic H, Pozo JM, Villa-Uriol MC, Majoie CB, van den Berg R, Gratama van Andel HA, Macho JM, Blasco J, San Roman L, Frangi AF (2011) Automated segmentation of cerebral vasculature with aneurysms in 3dra and tof-mra using geodesic active regions: an evaluation study. Med Phys 38(1):210–222. https://​doi.​org/​10.​1118/​1.​3515749CrossRef
go back to reference Dakua SP, Abinahed J, Zakaria A, Balakrishnan S, Younes G, Navkar N, Al-Ansari A, Zhai X, Bensaali F, Amira A (2019) Moving object tracking in clinical scenarios: application to cardiac surgery and cerebral aneurysm clipping. Int J Comput Assist Radiol Surg 14(12):2165–2176. https://doi.org/10.1007/s11548-019-02030-zCrossRef Dakua SP, Abinahed J, Zakaria A, Balakrishnan S, Younes G, Navkar N, Al-Ansari A, Zhai X, Bensaali F, Amira A (2019) Moving object tracking in clinical scenarios: application to cardiac surgery and cerebral aneurysm clipping. Int J Comput Assist Radiol Surg 14(12):2165–2176. https://​doi.​org/​10.​1007/​s11548-019-02030-zCrossRef
go back to reference Gonzalez RC, Wood RE (2002) Digital image processing, 2nd edn. Prentice-Hall, New Jersey Gonzalez RC, Wood RE (2002) Digital image processing, 2nd edn. Prentice-Hall, New Jersey
go back to reference Higashida RT (2003) What you should know about cerebral aneurysms. Pamphlet, American Heart Association Cardiovascular Council Higashida RT (2003) What you should know about cerebral aneurysms. Pamphlet, American Heart Association Cardiovascular Council
go back to reference Hiremath P, Akkasaligar PT, Badiger S (2010) Speckle reducing contourlet transform for medical ultrasound images. Int J Compt Inf Eng 4(4):284–291 Hiremath P, Akkasaligar PT, Badiger S (2010) Speckle reducing contourlet transform for medical ultrasound images. Int J Compt Inf Eng 4(4):284–291
go back to reference Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, Killeen T, Lin Z, Gimelshein N, Antiga L, Desmaison A, Kopf A, Yang E, DeVito Z, Raison M, Tejani A, Chilamkurthy S, Steiner B, Fang L, Bai J, Chintala S (2019) PyTorch: an imperative style. High-performance deep learning library. Curran Associates Inc., Red Hook, NY, USA Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, Killeen T, Lin Z, Gimelshein N, Antiga L, Desmaison A, Kopf A, Yang E, DeVito Z, Raison M, Tejani A, Chilamkurthy S, Steiner B, Fang L, Bai J, Chintala S (2019) PyTorch: an imperative style. High-performance deep learning library. Curran Associates Inc., Red Hook, NY, USA
go back to reference Wang S, Cao S, Wei D, Wang R, Ma K, Wang L, Meng D, Zheng Y (2020) Lt-net: label transfer by learning reversible voxel-wise correspondence for one-shot medical image segmentation. In: 2020 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 9159–9168. https://doi.org/10.1109/CVPR42600.2020.00918 Wang S, Cao S, Wei D, Wang R, Ma K, Wang L, Meng D, Zheng Y (2020) Lt-net: label transfer by learning reversible voxel-wise correspondence for one-shot medical image segmentation. In: 2020 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 9159–9168. https://​doi.​org/​10.​1109/​CVPR42600.​2020.​00918
Metadata
Title
Development of a cerebral aneurysm segmentation method to prevent sentinel hemorrhage
Authors
Yousra Regaya
Abbes Amira
Sarada Prasad Dakua
Publication date
01-12-2023
Publisher
Springer Vienna
Published in
Network Modeling Analysis in Health Informatics and Bioinformatics / Issue 1/2023
Print ISSN: 2192-6662
Electronic ISSN: 2192-6670
DOI
https://doi.org/10.1007/s13721-023-00412-7

Other articles of this Issue 1/2023

Network Modeling Analysis in Health Informatics and Bioinformatics 1/2023 Go to the issue

Premium Partner