This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

1

Brain Tumor Segmentation Based on Local Independent Projection-based Classification Meiyan Huang, Wei Yang, Yao Wu, Jun Jiang, Wufan Chen, Senior Member, IEEE, and Qianjin Feng*, Member, IEEE  Abstract—Brain tumor segmentation is an important

procedure for early tumor diagnosis and radiotherapy planning. Although numerous brain tumor segmentation methods have been presented, enhancing tumor segmentation methods is still challenging because brain tumor MRI images exhibit complex characteristics, such as high diversity in tumor appearance and ambiguous tumor boundaries. To address this problem, we propose a novel automatic tumor segmentation method for MRI images. This method treats tumor segmentation as a classification problem. Additionally, the local independent projection-based classification (LIPC) method is used to classify each voxel into different classes. A novel classification framework is derived by introducing the local independent projection into the classical classification model. Locality is important in the calculation of local independent projections for LIPC. Locality is also considered in determining whether local anchor embedding is more applicable in solving linear projection weights compared with other coding methods. Moreover, LIPC considers the data distribution of different classes by learning a softmax regression model, which can further improve classification performance. In this study, 80 brain tumor MRI images with ground truth data are used as training data and 40 images without ground truth data are used as testing data. The segmentation results of testing data are evaluated by an online evaluation tool. The average dice similarities of the proposed method for segmenting complete tumor, tumor core, and contrast-enhancing tumor on real patient data are 0.84,

This work was supported by the grants from the National Basic Research Program of China (973 Program) (No. 2010CB732505), National Science & Technology Pillar Program of China (No. 2012BAI14B02), National Natural Science Funds of China (NSFC, No. 81101109 and No. 31371009), National High Technology Research and Development Program of China (863 Program) (No. 2012AA02A616), and Program of Pearl River Young Talents of Science and Technology in Guangzhou (No. 2013J2200065). Asterisk indicates corresponding author. Meiyan Huang, Wei Yang, Yao Wu, Jun Jiang, and Wufan Chen are with School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China (e-mail: [email protected], [email protected], [email protected], [email protected], [email protected]). *Qianjin Feng is with School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China (e-mail: [email protected]). (phone: +86-20-61648286; fax: +86-20-61648286). Copyright (c) 2013 IEEE. Personal use of this material is permitted. However, permission to use this material for any other purposes must be obtained from the IEEE by sending an email to [email protected].

0.685, and 0.585, respectively. These results are comparable to other state-of-the-art methods. Index Terms—Brain tumor segmentation, local independent

projection-based classification, local anchor embedding, softmax regression I. INTRODUCTION

B

RAIN tumor segmentation is one of the crucial procedures in surgical and treatment planning. However, at present, brain tumor segmentation in brain tumor images is mostly performed manually in clinical practice. Apart from being time consuming, manual brain tumor delineation is difficult and depends on the individual operator. Currently, multimodal MRI images are used simultaneously by radiologists in segmenting brain tumor images because multimodal MRI images can provide various data on tumors. In glioma, the tumor area is usually divided into necrosis, contrast-enhancing tumor, non-enhancing tumor, and edema [1]. Different MRI image modalities can reveal different parts in the tumor area. For instance, T1C (T1-weighted images with contrast enhancement) highlights contrast-enhaning regions, whereas T2 highlights edema regions (Fig. 1). Although multimodal MRI images can provide complementary information in the tumor area, brain tumor segmentation is still a challenging and difficult task. Brain tumors can have various sizes and shapes and may appear at different locations. In addition to tumor heterogeneity, tumor edges can be complex and visually vague (Fig. 1). Moreover, some tumors may deform surrounding structures in the brain because of the mass effect or edema (Fig. 1). Additionally, artifacts and noise in brain tumor images increase the difficulty when segmenting tumors. Thus, designing of a semi-automatic or automatic brain tumor segmentation approach is necessary to provide an acceptable performance. Numerous algorithms have been developed to perform brain tumor detection and segmentation. These methods include thresholding and morphological techniques [2-4], watershed method [5], region growing approach [4, 6], asymmetry analysis [7-10], atlas-based method [11-14], contour/surface evolution method [15-17], interactive algorithm [18-20], and supervised [21-29] and unsupervised [6, 30-35] learning methods. A. Intensity-Based Method Low-level operations, such as thresholding, edge detection,

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

2

Fig. 1. Different modalities reveal different parts in the tumor area. The edge of the tumor area is visually vague. In addition, the brain structure is deformed because of the occurrence of edema. (a) T1C-weighted brain tumor MRI image. (b) T2-weighted brain tumor MRI image. (c) Contour of the actual brain tumor. “t” represents the combination of contrast-enhancing and necrotic parts, and “e” represents the edema part.

and morphological techniques [2-4], are fast and can be easily adjusted. However, the tumor segmentation performance of these methods highly depends on evident difference in the intensities between tumor and non-tumor regions. Watershed [5] and region growing [4, 6] approaches are simple and consistently produce complete boundaries. However, these two methods are sensitive to noise, which is a common problem in the intensity-based method. Moreover, most intensity-based methods tend to oversegment tumors because of the weak and diffused edges caused by edema [15]. B. Asymmetry Analysis The healthy human brain is largely symmetric across the mid-sagittal plane. The asymmetric analysis method [7-10] for tumor segmentation is based on the principle that tumors, which appear in one of the cerebral hemispheres, can cause asymmetry between the left and right cerebral hemispheres. This asymmetry can be detected, and tumors can be roughly located in the corresponding cerebral hemisphere. The asymmetric analysis method can hasten the tumor detection and segmentation process because tumor segmentation is implemented in one of the cerebral hemispheres. However, accurately finding the mid-sagittal plane is a challenging and time-consuming task [7]. More importantly, asymmetry analysis may not be useful when a tumor is located across the mid-sagittal plane [9]. C. Atlas-Based Method Atlas-based segmentation methods have been extensively investigated [36, 37]. Brain atlases can provide important data prior to tumor segmentation enhancement by measuring the difference between abnormal and normal brains. However, the deformable registration of the brain atlas to brain images with tumor is an extremely challenging task because of the intensity variations around the tumor caused by edema and the deformations of healthy tissue morphology caused by the tumor mass effect [38]. In a previous study [11], affine registration is used to align the atlas to the tumor image data. When a large brain structure deformation appears, the misalignment issues are noted on the aligned atlas, which may significantly decrease segmentation accuracy [11].

D. Contour/Surface Evolution Method The contour/surface evolution method has been widely used for the tumor segmentation of 2D/3D data [15-17]. This method can be represented implicitly as a level set function or explicitly as an active contour model/snake function. Compared with the parametric active contour model, the level set method can represent contours with complex topology and handle topological changes, such as splitting and merging in a natural and efficient way [19]. Furthermore, the extension of the level set method to 3D is straightforward and does not require additional machinery [16]. However, the contour/surface evolution method does not easily determine the initializations and tune the parameters even when 3D level set surfaces are used. E. Interactive Algorithm Graph-based seeded segmentation framework is one of the popular methods among interactive algorithms. Graph-based seeded segmentation is a global optimization approach, which showed outstanding performance for tumor segmentation in our previous studies [19, 20]. However, this method needs manual seed selection in different tissues, and distinguishing different tissues in the tumor is difficult during the selection of initial seeds for different tissues. In a previous study [18], a cellular automata-based seeded method, called tumor-cut, has been presented for brain tumor segmentation. This method only requires the user to draw a line over the largest visible tumor diameter. Although this initial seed selection strategy can reduce manual interaction and decrease the sensitivity of the method to initialization, this procedure may not include all tumor areas within the volume of interest along the depth direction [18], thus leading to tumor undersegmentation. F. Supervised and Unsupervised Learning Methods Unsupervised learning method, such as k-means and fuzzy clustering, has become popular for brain tumor segmentation in recent years [6, 30-35]. The fuzzy method considers that medical images are inherently fuzzy, so it is a very strong tool for medical image processing. Furthermore, the fuzzy method can capture pixel proximity in the same objective region without a training step. However, most fuzzy methods work well only for tumors that present hyper-intensity and exhibit poor performance on segmenting non-enhanced tumors [7]. These conditions are due to the fact that these fuzzy methods typically use intensity-based method, such as thresholding and morphological operations, as pre- or post-processing. Supervised classification learning method is widely used in tumor segmentation [21-29]. Well-trained classifiers can extract discriminative information from the training data and estimate the label of each voxel in a testing volume. However, the traditional classification methods classify each voxel into different classes without considering the spatial correlation between current and nearby voxels. This method may not obtain a global optimized result. To address this problem, a classification method is generally combined with a regularization step. The regularization step can be implemented by modeling the boundary [25] or by applying a variant of a

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

3 random field spatial prior (MRF/CRF) [23, 24, 26, 29]. In the previous studies [21, 22], context-aware spatial features and the probabilities obtained by tissue-specific Gaussian mixture models are used as inputs for classifiers, and satisfied segmentation results are achieved without using post-hoc regularization. Similar to the previously described methods [21, 22], we propose a novel classification method, named local independent projection-based classification (LIPC), for brain tumor segmentation without using explicit regularization. The contribution of this study can be summarized as follows. First, an LIPC-based method is introduced to solve the tumor segmentation problem. Based on recent studies on sparse representation [39-41], the sparse representation-based classification (SRC) method [42] was adopted in medical image segmentation, thereby producing robust results. In the SRC method, a sample can be sparsely represented in a specific dictionary, which contains samples from all classes. Contrary to SRC, the proposed LIPC assumes that the training samples from different classes are located on different submanifolds. Then, the training samples are divided into different groups and subsequently used to construct different dictionaries. The testing sample is independently projected on these different dictionaries using the local anchor embedding (LAE) method [43]. Second, a softmax model is used to determine the relationship between data distribution and reconstruction error norm. When the data distribution is uniform and the noise is low, classification may be performed well. However, data distribution is complex in brain tumor MRI images. In addition, the data distribution of different classes (i.e., tumor, edema, and brain tissue) may widely vary. Therefore, the data distribution of each class should be considered when segmenting brain tumors. Our evaluation on synthetic data and public available brain tumor image data demonstrates that considering the data distribution of different classes can further improve the segmentation performance. This study is organized as follows. Section II introduces methodology of the proposed algorithm. Section III presents the experimental analysis and results. Section IV provides several discussions and conclusions on the proposed method. A preliminary work of this study appeared in another study [44]. II. MATERIALS AND METHODS A. Materials The brain tumor image data used in this study were provided by the MICCAI 2012 and 2013 Challenges on Multimodal Brain Tumor Segmentation (http://www.imm.dtu.dk/projects/BRATS2012 and http://martinos.org/qtim/miccai2013/data.html). The challenge database contains fully anonymized images from ETH Zurich, University of Bern, University of Debrecen, and University of Utah [45]. The brain tumor image data contain information on 120 subjects with gliomas; 55 images are real patient data, and 65 images are synthetic data (Table 1). A total of 80 images out of 120 images with ground truth data are treated as training data.

TABLE I DETAILS OF THE BRAIN TUMOR IMAGE DATA Real patient data Synthetic data High-grade

Low-grade

High-grade

Low-grade

20

10

25

25

Training data Testing

2012

11

4

10

5

data

2013

10

-

-

-

Total

55

65

Fig. 2. Flowchart of the proposed method

In the ground truth data, the complete tumor region is subdivided into necrosis, contrast-enhancing tumor, non-enhancing tumor, and edema parts for real patient data, whereas the complete tumor region is subdivided into tumor core and edema parts for synthetic data. The remaining 40 images are testing data with publicly unavailable ground truth data. T1, T2, FLAIR, and post Gadolinium T1 MRI images are available for each subject. All volumes are linearly co-registered to the T1 contrast image, skull stripped, and interpolated to 1 mm isotropic resolution. In addition to the brain tumor image data, a set of interleaving spirals was also used in elucidating the mechanism of the proposed LIPC method. B. Overview of the Proposed Method The proposed method consists of four major steps, i.e., preprocessing, feature extraction, tumor segmentation using the LIPC method, and postprocessing. To reduce computational costs, we embedded the proposed method in a multi-resolution framework. The flowchart of the proposed method is illustrated in Fig. 2. C. The basic principle of LIPC Brain tumor segmentation can be considered as a multiclass classification problem. To solve this problem, a one-versus-all (OvA) strategy can be used. In the OvA approach, a classifier is trained per class to distinguish a class from all other classes. Therefore, N classifiers f   fi i 1 have to be determined in this N

study, where N represents the number of classes. Given a testing sample x 

M

, N real classification scores y   yi i 1 N

are computed using the learned classifiers f  x  ; where the sample x stands for the image feature in the current study

(details in Section II. F) and yi   0,1 stands for the probabilities that the sample belongs to the ith class. The label of sample x can be defined as follows

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

4

l  arg max f i  x   arg max yi . i1, , N

(1)

i1, , N

Before the proposed LIPC was introduced, the following assumption was considered as the base for LIPC: Assumption I: Samples from different classes are located on different non-linear submanifolds, and a sample can be approximately represented as a linear combination of several nearest neighbors from its corresponding submanifold. For N-class classification, this assumption indicates that the samples are found on a manifold Mi i1 , which consists of N N

submanifolds; M i represents the submanifold associated with the ith class. For a dictionary D  Di i1 , which consists of N N

subdictionaries, Di  di1 , di2 , , diNi  consists of Ni typical samples from the ith submanifold. Considering Assumption I, a testing sample x can be projected into each submanifold using the following linear representation:

x  Di ai   i   j i 1 aij di j   i N

(2)

 i   , di j   ix  k  , aij  0

s.t.

i where  x  k  denotes the set of k nearest neighbors of x in

Di

dictionary

i

,

is

the

reconstruction

error,

ai   ai1 , ai2 , , aiNi  is the weight coefficient vector of the T

linear combination, and  is a small positive real number that ensures the reconstruction accuracy. Many (at least Ni-k) elements of ai are zeroes for an arbitrary sample x, which indicates that the described linear representation is sparse. The L2-norm of  i noted as  i

2

can be used in measuring the

proximity between x and its projection on Mi. Furthermore, a smaller  i

is associated with increased probability that x

2

belongs to the ith class. Therefore, the classification scores of x have a negative correlation with  i

 a d     g    g  x   a d 

yi  f i  x   f i

Ni

j

j 1

i

i

i

and can be formulated as

j

i

Ni

i

2

j 1 i

i

j

(3)

j

i

(3), the label prediction in Eq. (1) can be rewritten as l  arg max f i  x   arg max yi  arg max gi   i  . i1, , N

a  arg min x   j 1 a j d j N

where g i    is an inverse proportional function. Based on Eq.

i1, , N

D. LIPC Implementation We performed dictionary construction, locally linear representation, and classification score computation to implement LIPC. 1) Dictionary construction The manually labeled original samples in a training set are used to construct D [41]. However, numerous original training samples possibly produce a large D, which dramatically increases computational and memory costs. In the present study, more than half a million samples for each class are available for training. Thus, subsequent processes are impractical when conducted traditionally. Applying a dictionary learning method is necessary in learning a compact representation of the original training samples. The k-means [46] method can obtain the typical structures of the original sample space; and thus, this method is used in the current study to learn a compact representation of the original training samples of each class. 2) Locally linear representation Several methods, such as sparse coding (SC) [41], locality-constrained linear coding (LLC) [47], and LAE, have been proposed to approximately represent a sample that is linearly based on training samples. Sparse representation emphasizes the sparsity of the representation and attempts to use the smallest number of training samples in reconstructing the testing sample with minimal error. LLC and LAE belong to the special cases of sparse representation. These two approaches focus on locality rather than sparsity by limiting the linear coding within a local neighborhood. In particular, LAE forces the weight coefficients to maintain non-negativity and a sum that is equal to one, which ensures that the reconstructed sample is a convex combination of its closest neighbors. Therefore, the reconstructed sample is located in a convex region on a hyperplane spanned by its closest neighbors. Actually, this condition is a tough constraint of the locality. The locality of the representation is highly preferable for classification tasks [43, 47, 48]. Thus, LAE is appropriate in solving the linear representation in the current study. For completeness, we briefly introduce LAE. Considering the concrete task in this paper, we formally re-formulize the cost function of LAE as

i1, , N

(4)

The proposed LIPC is different from SRC [42]. Although the reconstruction errors are both used as the classification measure in these two methods, the testing sample is projected onto each submanifold independently in the proposed method instead of the whole manifold in SRC. For each submanifold, the projection calculation is independent. Therefore, if Assumption I is considered, the proposed method is more applicable than SRC.

2

a

s.t.  d j   x  k  , a j  0,  j a j  1, a j  0

(5)

N

Three steps were performed to obtain the solution of LAE. First, the k nearest neighbors of the testing sample x was

selected from D and  x  k  was constructed. Second, for djs that do not belong to  x  k  , the associated ajs were set to 0. Third, for the remaining djs that belong to  x  k  , their corresponding ajs was calculated using the projected gradient method [43]. The updating rule in the projected gradient method can formally be expressed as the following iterative formula:

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

5

 



ait 1  P ait   t h ait 

(6)

where t is the number of iterations, t  0 is the step size,

h  a  is the gradient of function h at a, and P  a  is the simplex projection operator on any a 

p

. The simplex

projection operator can be formulated as

P  a   arg min a  a

(7)

a 

As introduced in [43], Eq. (6) uses the simplex projection

operator in Eq. (7), is implemented efficiently in O  s log s  , and can be accelerated using Nesterov’s method [49]. 3) Classification score computation As described in Section II. C, the classification scores of x have a negative correlation with  i

. Furthermore, to

2

implement the classification within a multi-resolution framework (details in Section II. G), the classification score can be easily represented as a probability form: yi  g   i   exp    i

2

 

N i 1

exp    i

2



(8)

However, Eq. (8) considers only the relationship between the reconstruction error norm and the classification score, which ignores the relationship between the reconstruction error norm and the data distribution on each submanifold. In general, the data distributions on different submanifolds are different. Therefore, the reconstruction error norms of the samples cover a wide range and may sometimes violate the negative correlation as listed in Eq. (8) [Fig. 3 (a)]. To address this problem, we propose to establish a softmax regression model using the reconstruction error norms of the training data to achieve the classification score [Fig. 3 (b)]. Thus, the classification score can be defined as yi  exp  wiT 

2

 

N k 1

exp  wkT 

2



(9)

Given the original training set T = Ti i 1  N

x , l   j ni

j

i

i

N

j 1 i 1

and

the learned dictionary D = Di i 1 , the weight coefficient N

 j

vector ai that corresponds to each sample in T can be calculated using Eq. (5). Then, the reconstruction error norm



2



  1 

 j 2

,

, N

 j  T 2





n

of all training samples in T can j 1

be computed using Eq. (2). With 

2

, the cost function of the

softmax regression model can be defined as  j exp( wiT  2 )  N 1 n  j  (10) J  w      j 1 i 11 l  i log N  j T n  exp( w  )  k 2 k 1   is an indicator function. Given that no closed-form where 1





technique can be used to solve for the minimum of J  w  , we employ the gradient descent algorithm to iteratively optimize Eq. (10).

Fig. 3. (a) An example that shows how a testing sample may be wrongly classified if the data distributions on different submanifolds are not considered.  2 2 is smaller than  1 2 , so the testing sample will be assigned to Class 2 based on Eq. (8). However, assigning the testing sample to be Class 1 is reasonable after considering the distribution of all training data on different submanifolds. (b) Learning a softmax regression model using the reconstruction error norms to separate the data into different classes.

E. Summary of LIPC We provide the pseudo-code of the proposed method in Algorithm I. F. Feature Extraction Before extracting image features to represent a sample, image inhomogeneity correction and intensity normalization should be performed because the image intensities in MRI images do not have a fixed meaning and widely vary within or between subjects. In this study, all MRI image modalities are processed as follows. First, the N3 algorithm [50] is applied to remove the bias field artifacts from the images. Second, intensity values at the 1% and 99% quantiles are computed for the brain region (including tumors, edema, and brain tissues), and then these two values are used to linearly scale the voxel intensities to the range [0,100]. In the present paper, a patch-based technique was used in extracting the image feature. The intensity values in a patch around a voxel v were obtained and rearranged as a feature vector. Four modalities of MRI images were used, so the feature of v that used a cubic patch with a size of w  w  w was (w3 × 4)-dimensional. G. Multi-resolution Framework To improve robustness and reduce computational cost, a multi-resolution framework was adopted in the proposed method. For a multi-resolution framework with P levels, classification originated from the coarsest Lp 1 to the finest level L0 (original resolution) to classify the voxels into different classes. The classification scores for all voxels in coarser levels were up-sampled to initialize the classification for a finer level using the trilinear interpolation method. At each level, a confidence threshold  was defined to determine the specific voxels that can be directly labeled using the propagated initial classification scores and the specific voxels that required further processing. In this study, the voxels with initial classification scores yi  1   were labeled as the ith class. The remaining voxels with yi  0,1    were fed into the proposed classifier for an accurate classification. This procedure was repeated until the resolution of the original level

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

6 ALGORITHM I



Input: Training set T = Ti i 1  xi , li N

j





N

j ni

j 1 i 1

; A testing sample x.

Output: The classification scores y   yi i 1 and the label l of x. N

Stage 1: Construction of sub dictionary for each class Partition  xij , li j  j 1 into Ni sub-sets and calculate the cluster centers denoted as Di  di1, di2 , , diNi  using k-means methods. Stage 2: Calculation of locally linear representation coefficients ni

 

Reconstruct each x j  in T based on dictionary D using the LAE method and calculate the corresponding coefficient vector ai j 

N

i 1

.

Stage 3: Softmax regression model determination

 

Calculate reconstruction error vector  i of all training samples for each class based on dictionary D = Di i 1 and coefficient vectors a  j  N

n j 1

.

Use the calculated reconstruction error norms to determine a softmax regression model by minimizing Eq. (10). Stage 4: Core of the proposed method Reconstruct the input sample x based on dictionary D using the LAE method and calculate the coefficient vectors ai i 1 . N

Calculate the reconstruction error vector  i of x for each class according to Eq. (2). Calculate the classification scores y   yi i 1 of x based on Eq. (9). N

Achieve the label l of x based on Eq. (1).

L0 was reached. At level L0, we directly used Eq. (1) to achieve the final label. Three levels (P = 3) were used with corresponding voxel sizes of 1 mm × 1 mm × 1 mm, 2 mm × 2 mm × 2 mm, and 4 mm × 4 mm × 4 mm. The voxel number for each class in each training image was set to 8,000 to construct the training set at each level, and the threshold value   was empirically set to 0.2 at L1 and L2. In the coarser levels, the voxels in the tumor or edema regions were very few. To increase the number of training samples, we randomly transformed the MRI images by smoothing and adding noise to the images at each level. H. Post-processing with Spatial Constraint In this study, we used the assumption that each edema region is located near tumor core regions to post-process the classified edema region as introduced in [11]. According to this assumption, each classified edema region must have a voxel near the classified tumor regions within some small distance. Therefore, the connected component algorithm and mathematical morphology can be used to refine the classified edema regions. First, a binary image representing the classified edema regions is formed. Second, the binary image is used as an input for the connected component algorithm, and then some individual edema regions are generated. Third, each individual edema region is dilated with a small structuring element and compared against the classified tumor regions. Finally, the dilated edema regions that share at least a voxel with the classified tumor regions are considered valid. The edema regions from these valid regions are retained as the final edema classification results, whereas the other edema regions are discarded. III. EXPERIMENTAL RESULTS A. Experiments on Synthetic Data A two-spiral structure was generated to evaluate the

Fig. 4. Error rates of different methods on different data. LIPC and LIPC-learn represent the proposed method without and with the learned softmax regression model, respectively.

classification performance of the proposed LIPC method. A comparison was also made between the proposed method and two other popular classification methods [SRC and support vector machine (SVM) [51]] using the generated two-spiral structures. As described in Section II. C, SRC uses the whole dictionary to represent the testing sample and classifies x based on the performance of the coefficients associated with the training samples. In the proposed LIPC method, each subdictionary was used to represent x, and the closest subspace was determined to accomplish the classification task. Moreover, a softmax model was learned in LIPC to capture the information of data distribution of different classes. SVM attempts to minimize the empirical classification error and maximize the geometric margin simultaneously in the training set, which leads to high generalization ability in the new samples. In this section, the parameters in the two-spiral structures

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

7 TABLE II SUMMARY OF THE PARAMETER SETTINGS USED IN THE PROPOSED METHOD Parameter Description Setting k

Number of the nearest neighbors of LAE

10

N

Dictionary size

40,000

w

Patch size (w × w × w)

5

P α

Levels of Multi-resolution

3

Confidence threshold described in Section

0.2 at levels

2.5

L1 and L2

Fig. 5. (a) The number of training samples in each of the two classes was 500, and the noise level was 2.9. The decision curves learned by LIPC (b) without and (c) with the learned softmax regression model on the testing samples. (d) The number of training samples in the two classes was 50 and 500, and the noise level was 2.9. The decision curves learned by LIPC (e) without and (f) with the learned softmax regression model on the testing samples.

were designed as follows. The number of training samples on one spiral was fixed to 500. Then, the number of training samples on other spiral was set to 50, 250, and 500. The number of testing samples on each spiral was set to 1000. Noises with different levels (0.5 to 3) were also added to the generated interleaving spirals. The coordinates of each point in the spirals were the features to be fed to different classification methods. For LIPC and SRC, the training samples were used as dictionary because the sample number was small. To evaluate the LIPC effectiveness, both SRC and LIPC used LAE as the coding method. To make a fair comparison, we carefully designed and selected the parameters with optimal performance in each classification method. The classification results of different methods on different data are shown in Fig. 4. When the number of training samples on the two spirals was the same (e.g. 500 and 500) and the noise was small, the difference in the classification accuracy of the different methods was nonsignificant. However, the classification accuracy of the different methods varied significantly when the training samples on one spiral were much sparser than the training samples on the other spiral (e.g., 50 and 500). A higher noise level led to a large variety of classification accuracy under the different methods. In general, the classification accuracy decreased when the data distribution was non-uniform and the noise level was high. However, the proposed method with a learned softmax regression model achieved the lowest error rate when a serious non-uniformity of data distribution and high noise were observed in the training samples. The higher classification accuracy of the proposed method contributes in considering the data distribution of different classes. The classification performance of the proposed method without the learned softmax regression model was comparable to that of SVM. The proposed method with or without the learned softmax regression model also outperformed the SRC method, which indicates that the proposed LIPC was more effective than

Fig. 6. Effect of different numbers of k nearest neighbors on the reconstruction error

SRC in this classification task. The decision curves of the proposed method are shown in Fig. 5. As shown in Figs. 5 (b) and (c), the decision curves of the proposed method with and without the learned softmax regression model were similar when the non-uniformity of the data distribution was not obvious. However, the decision curve of the proposed method without the learned softmax regression model was worse than that of the proposed method with the learned softmax regression model when the training samples on one spiral were much sparser than those on the other spiral [Figs. 5 (e) and (f)]. B. Experiments on Brain Tumor Image Data A series of experiments were performed on the training and testing brain tumor image data. For the training data, a complete tumor is subdivided into tumor core (necrosis, contrast-enhancing tumor, and non-enhancing tumor in real patient data) and edema parts. The proposed method was evaluated using a five-fold cross-validation fashion. All experiments were repeated five times, and the final results were reported as the mean and standard deviation of the results from the individual runs. For each run, a total of 64 images were used in training and 16 images were used for testing. Meanwhile, no joint set existed between the training and testing datasets in all experiments. To evaluate the proposed method for the testing data, the training data was used to train classifiers for different tissues. The segmentation results of the testing data were

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

8

Fig. 7. Classification performance of the proposed method with different values of N on different data groups.

Fig. 9. Qualitative results of the proposed method for different data groups. Each row shows a typical result for each data group. (a) Slice of T2 images. (b) Contours of the edema classification results (yellow) and ground truth (red). (c) Contours of the tumor core classification results (yellow) and ground truth (red). (d) Ground truth of the tumor core (red voxels represent the necrosis, green voxels represent the non-enhancing tumor, and blue voxels represent the contrast-enhancing tumor). (e) Classification results of the tumor core.

Fig. 8. Classification performance of the proposed method with different values of w on different data groups.

uploaded to the online evaluation platform and were evaluated automatically using the online evaluation tool (http://virtualskeleton.ch/). 1) Quantitative evaluation metrics The dice similarity (DS) and jaccard similarity (JS) were used in this study to quantitatively assess the performance of the proposed method. These two similarities are defined as follows: DS  2 A  B  A  B  , JS  A  B A  B (10) where A and B are the voxel sets of the result and ground truth, respectively. The false positive rate (FPR) and false negative rate (FNR) were also adopted to quantify the over-segmentation and under-segmentation. These two metrics are computed as follows:

FPR  A  B A  B , FNR  A  B A  B

(11)

2) Parameter optimization The parameter settings of the proposed method were carefully considered to obtain the optimum performance during our experiments. A summary of the parameter settings used in the proposed method is listed in Table II. In this study, we employed LAE method to achieve the linear representation coefficients to linearly reconstruct a testing

Fig. 10. Typical results using the proposed method on high-grade and low-grade real patient data. (a) Slice of T2 images. (b) Ground truth of the complete tumor (yellow voxels represent the edema, red voxels represent the necrosis, green voxels represent the non-enhancing tumor, and blue voxels represent the contrast-enhancing tumor). (c) Classification results of the complete tumor.

sample and ensure the minimum reconstruction error. To obtain the solution of LAE, we first selected the k nearest neighbors of the testing sample from the dictionary. In this experiment, we varied k from 5 to 100 to evaluate the effects of the different numbers of k nearest neighbors on the reconstruction error. We first randomly selected 10,000 samples from the training data for each class at each level. For each class, we then computed the reconstruction error norms of the selected samples using the corresponding dictionary. We fixed the dictionary and patch sizes of each class at each level to be 40,000 and 5, respectively. The effect of the varied numbers of k nearest neighbors on the reconstruction error is shown in Fig. 6. We found that the lowest value for the reconstruction error was always obtained when k was equal to 10 for each class at each level. This result is significant in rapidly obtaining the solution of LAE. The dictionary size N was a crucial parameter in the proposed LIPC. Therefore, we performed the proposed method

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

9 TABLE III EFFECT OF LEARNING A SOFTMAX REGRESSION MODEL USING THE RECONSTRUCTION ERROR NORM. “WITH” AND “WITHOUT” REPRESENT THE PROPOSED METHOD WITH AND WITHOUT USING THE LEARNED SOFTMAX REGRESSION MODEL, RESPECTIVELY. High-grade (real) Low-grade (real) High-grade (synth) Low-grade (synth) Edema Tumor Edema Tumor Edema Tumor Edema Tumor DS Without 71.9 ± 11.7 79.8 ± 17.0 50.5 ± 8.5 72.8 ± 22.4 65.8 ± 17.0 89.5 ± 10.2 66.8 ± 13.2 90.0 ± 1.2 (%)

With

76.6 ± 8.5

86.9 ± 4.9

63.9 ± 14.7

83.7 ± 6.7

66.5 ± 20.8

90.1 ± 2.2

67.4 ± 13.9

90.4 ± 2.3

JS

Without

58.6 ± 16.6

70.1 ± 18.5

38.8 ± 11.2

64.3 ± 23.3

52.6 ± 21.1

80.8 ± 14.6

51.1 ± 16.5

80.9 ± 2.1

(%)

With

62.8 ± 10.7

77.2 ± 7.6

48.5 ± 15.8

72.5 ± 9.3

53.0 ± 21.4

82.1 ± 3.6

52.2 ± 14.3

82.5 ± 3.8

FPR

Without

32.8 ± 26.2

11.8 ± 10.1

50.9 ± 26.9

28.9 ± 24.8

9.1 ± 1.1

1.0 ± 0.94

3.9 ± 0.9

0.2 ± 0.03

(%)

With

2.8 ± 1.7

0.90 ± 0.1

2.9 ± 2.8

0.90 ± 0.84

3.9 ± 2.6

1.07 ± 0.1

1.9 ± 1.5

0.6 ± 0.1

FNR

Without

8.6 ± 5.2

18.1 ± 7.9

10.3 ± 7.6

6.8 ± 4.3

38.3 ± 27.4

18.2 ± 14.6

45.0 ± 18.9

18.9 ± 3.6

(%)

With

34.4 ± 11.1

21.9 ± 7.6

48.6 ± 15.3

26.6 ± 9.5

43.1 ± 20.6

16.8 ± 4.1

45.9 ± 13.7

16.9 ± 4.2

on the brain tumor image data to compare the effect of different dictionary sizes Ni on the classification performance. We set Ni for different classes to be of the same values at different levels, and the dictionary size was denoted as N. We varied N from 5,000 to 40,000. k was also set to 10 for each class at each level for different values of N. The patch size for each class at each level was fixed to 5. As shown in Fig. 7, increasing N improved the classification accuracy on the different image data groups. Meanwhile, the standard deviation decreased with increasing N. in general, large dictionaries contain numerous discriminative data, which result in a more accurate classification compared with small dictionaries. However, to obtain the trade-off between memory and computational costs and accuracy, we selected N = 40,000 for the following experiments. The patch size w is also an important parameter that should be carefully determined. We set the same w for the different classes at different levels and fixed w to 3, 5, and 7. Parameters k and N were set to 10 and 40,000, respectively, for each class at each level. The classification accuracy was improved by increasing w from 3 to 5 (except for the tumor classification accuracy on real patient data with high-grade gliomas and synthetic data with high-grade gliomas), in which the average DS increased from 0.742 to 0.782 (Fig. 8). The average DS of w = 5 (0.782) was also higher than that of w = 7 (0.756). Therefore, we selected w = 5 for the following experiments. 3) Qualitative results The qualitative results of the proposed method with the learned softmax regression model on different data groups are shown in Fig. 9. Parameters k, N, and w were set to 10, 40,000, and 5, respectively, for each class at each level. The tumor boundaries of the real patient data were more blurry than those of the synthetic data [Fig. 9 (a)]. Therefore, the tumor classification performance was better in the synthetic data than that in the real patient data [Fig. 9 (c)]. The edema boundaries of both real patient data and synthetic data were quite blurry [Fig. 9 (a)], which led to more inaccurately classified voxels in the edema regions than those in the tumor regions [Fig. 9 (b)]. Two other typical segmentation results using the proposed method are shown in Fig. 10. Fig. 10 shows that the tumor boundary of low-grade real patient data was more blurry than that of high-grade real patient data. The edema region was also

Fig. 11. Tumor segmentation results of LLC, SC, and LAE.

Fig. 12. Comparison of LIPC and SRC

blurry in low-grade real patient data. Thus, the classification accuracy of the tumor and edema in the low-grade patient data was lower than those in the high-grade patient data. 4) Effectiveness of LAE method Numerous coding methods can be used to solve the coefficients of the dictionary. In this section, we show the comparative results of the tumor segmentation with several coding methods. The SC, LLC, and LAE coding methods were used for comparison. SC focuses on the sparse property, and LLC concentrates on the local property. In this study, LAE was used as the coding method, which emphasizes the local and

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

10 TABLE IV DS OF DIFFERENT METHODS ON THE TESTING DATA. THE RESULTS OF THE PROPOSED METHOD ARE SHOWN IN BOLD FORM. 2012 2013 Real patient data

Synthetic data

Real patient data

Position

complete

core

enhancing

complete

core

complete

core

enhancing

1

0.75

0.70

0.43

0.92

0.88

0.88

0.83

0.72

2

0.82

0.63

0.44

0.91

0.86

0.87

0.78

0.74

3

0.82

0.66

0.36

0.87

0.81

0.86

0.74

0.73

4

0.78

0.58

0.40

0.88

0.76

0.85

0.81

0.70

5

0.75

0.54

0.37

0.83

0.62

0.85

0.73

0.73

non-negative properties. To obtain a fair comparison, these three coding methods were performed in the LIPC framework, and the classification scores were computed using Eq. (8). Parameters N and w were set to 40,000, and 5, respectively, for each class at each level. For LAE, parameter k was set to 10 for each class at each level. For LLC and SC, we first conducted an experiment, which was similar to the experiment described in Section III. B. 2), to determine the number of nonzero values. We found that when the numbers of nonzero values were 90 and 15 for LLC and SC, respectively, the corresponding reconstruction error of the randomly selected samples reached the minimum value. The tumor segmentation results with different coding methods are shown in Fig. 11. LAE exhibited better robustness compared with SC and LLC. The mean DS of LAE was 2.6% and 3.1% higher than that of LLC and SC, respectively. This result indicates that LAE was more applicable in addressing difficulties in brain tumor segmentation compared with SC and LLC. 5) Comparison between LIPC and SRC To evaluate the effectiveness of LIPC, both SRC and LIPC used LAE as the coding method. Moreover, the classification scores were computed using Eq. (8). For LIPC, parameters k, N, and w were set to 10, 40,000, and 5, respectively, for each class at each level. For SRC, a dictionary containing samples from three classes was constructed for each level. This dictionary consisted of three subdictionaries and each subdictionary corresponded to a class. For a fair comparison, the size of each subdictionary was set to 40,000. Therefore, the size of the dictionary for SRC was 120,000 at each level. The number of nonzero values in LAE for SRC was determined as follows. We first randomly selected 10,000 samples from the training data for each class at each level and computed the reconstruction error norms of all the selected samples using the dictionary. The number of nonzero values in LAE was varied from 5 to 1,200. Finally, the minimum reconstruction error was found when the number of nonzero values was set to 1000. After we investigated the results of different data groups, the mean DS of LIPC was 5.3% higher than that of SRC. The classification results with LIPC and SRC on different data groups are displayed in Fig. 12, which shows that the proposed LIPC could be effectively used in tumor segmentation. 6) Effectiveness of softmax regression model learning To assess the effect of learning a softmax regression model using the reconstruction error norm on the classification performance, we compared the classification results obtained

from Eqs. (8) and (9). Parameters k, N, and w were set to 10, 40,000, and 5, respectively, for each class at each level. As listed in Table III, the accuracy of the proposed method with the learned softmax regression model was higher than that of the proposed method without using the learned softmax regression model. The case without the learned softmax regression model had a higher FPR than that with the learned softmax regression model, which indicates the former tended to retain extra background. 7) Comparison with other methods The segmentation results of the testing data were evaluated by the online evaluation tool in the evaluation platform. This evaluation platform contains an archive of all uploaded results, which enables segmentation methods to be objectively benchmarked and compared with each other. The tumor region in the real patient data was subdivided into four classes (necrosis, contrast-enhancing tumor, non-enhancing tumor, and edema), and the evaluation was performed for three different tumor subregions: complete tumor (necrosis, contrast-enhancing tumor, non-enhancing tumor, and edema), tumor core (necrosis, contrast-enhancing tumor, and non-enhancing tumor), and contrast-enhancing tumor. The tumor region in the synthetic data was subdivided into two classes (tumor core and edema). For the real patient data, five dictionaries and five classifiers were learned from the real patient training data. For the synthetic data, three dictionaries and three classifiers were learned from the synthetic training data. For all testing data, parameters N, w, and k were set to 40,000, 5, and 10 for each class at each level. Table IV shows the tumor segmentation results with different methods evaluated by the online evaluation tool. The top five results are listed. At the time of writing of this study, the proposed method was in the second and third places in the 2012 and 2013 real patient data, respectively. Moreover, the proposed method is in the first place in 2012 synthetic data. These results indicate that the proposed method is comparable to other state-of-the-art methods. 8) Computation time In this study, all experiments were implemented on a standard PC using a single thread on an Intel Core i5-2400 processor at 3.10 GHz. In the testing step, the processing time for tumor segmentation of a subject with N = 40,000 was approximately 26 min, at which 2 min was used for intensity normalization and 24 min was used for segmentation. In the training step, the construction of a dictionary with N = 40,000

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

11 from a training dataset with 64 images took 5 h. IV. DISCUSSION AND CONCLUSION An automatic method is proposed for brain tumor segmentation in MRI images. An LIPC-based method was introduced to solve the tumor segmentation problem. The proposed LIPC used local independent projection into the classical classification model, and a novel classification framework was derived. Compared with other coding approaches, the LAE method was more suitable in solving the linear reconstruction weights under the locality constraint. The data distribution in each submanifold was important for the classification, and we used a softmax model to learn the relationship between the data distribution and reconstruction error norm. We evaluated the proposed method using both synthetic data and public available brain tumor image data. In both problems, our method outperformed competing methods. A. Comparison with SRC Method The proposed LIPC method applied each subdictionary to represent the testing sample instead of using the whole dictionary to represent the testing sample in the SRC method. The classification strategy used in LIPC was more applicable than that used in SRC, especially when the data distribution is nonuniform and the noise level is high (Fig. 4). When the data distribution of one class is sparser than that of other classes, SRC may select few nearest neighbors in the sparse class. This selection strategy may result in the wrong classification if the testing sample belongs to the sparse class. To prevent this problem, SRC tended to select more nearest neighbors than LIPC (1000 neighbors in SRC vs. 10 neighbors in LIPC as mentioned in Section III. B. 5)), which leads to increased computation costs. B. Comparison with the State-of-the-Art in Brain Tumor Segmentation The brain tumor segmentation problem has been widely investigated; putting our results in context helps to reveal their significance. The proposed method was compared with other methods by uploading the segmentation results to the online evaluation tool. In the proceedings of the BRATS2012, Zikic et al. [21] and Bauer et al. [23] proposed an automatic classification method, respectively, and their methods performed best on the image data. Zikic et al. [21] achieved average DS values of 0.75, 0.47, and 0.41 for complete tumor, tumor core, and contrast-enhancing tumor for real patient testing data, respectively, and 0.91 and 0.86 for complete tumor and tumor core for synthetic testing data, respectively. Bauer et al. [23] achieved average DS values of 0.75, 0.54, and 0.37 for complete tumor, tumor core, and contrast-enhancing tumor for real patient testing data, respectively, and 0.87 and 0.81 for complete tumor and tumor core for synthetic testing data, respectively. Our results shown average DS values of 0.82, 0.63, and 0.44 for complete tumor, tumor core, and contrast-enhancing tumor for real patient testing data, respectively, and 0.92 and 0.88 for complete tumor and tumor core for synthetic testing data, respectively, the average of

Fig. 13. Intensity distribution of different classes in different data groups using T2 MRI images. The voxel number of the intensity value of each class was normalized and summed to one. “Background” represents the brain tissue region without tumor and edema.

which was 6% higher than those obtained by Zikic et al. [21] and Bauer et al. [23]. In the proceedings of BRATS2013, Cordier et al. [52] proposed a tumor segmentation method by choosing similar patches from the training data and combining labels of chosen patches to be the labels of testing data. The patches selection step in [52] is similar to the locally linear representation step in our proposed method. However, our proposed method focuses not only on similarity and sparsity, but also on locality. The average DS values of [52] reported in [45] were 0.84, 0.68, and 0.65 for complete tumor, tumor core, and contrast-enhancing tumor, respectively. Our results shown the average DS values of 0.86, 0.74, and 0.73 for complete tumor, tumor core, and contrast-enhancing tumor, respectively, the average of which was 5.4% higher than those obtained by Cordier et al. [52]. C. Computational Complexity Compared with the classical classification model, our proposed method requires an additional step of constructing a dictionary for each class at each level in the training stage. This dictionary construction step is an off-line procedure; thus this step is performed only once and is used for all testing images. In the testing stage, the k nearest neighbor searching and coding algorithms, which were the most time-consuming steps, were run in a single thread. However, the voxel-based implementation in this study can be parallelized and implemented to exploit the multi-core CPUs, thereby significantly decreasing processing time. D. Estimation of the Proposed Method In this study, we learned a softmax regression model using the reconstruction error norm to achieve the classification scores. The experimental results displayed an improvement on the classification performance using the learned softmax regression model (Figs. 4 and 5 and Table III). As shown in Table III, the classification accuracy of the proposed method with the learned softmax regression model was much higher than that of the proposed method without the learned softmax regression model for the real data groups (p < 0.002). However,

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

12 the classification accuracy of the proposed method with the learned softmax regression model was similar to that of the proposed method without the learned softmax regression model for the synthetic data groups (p > 0.6). This finding may be attributed to the more complex intensity distribution in the real data groups than that in the synthetic data groups (Fig. 13). The intensities of the three classes could be easily separated in synthetic data with high-grade gliomas, whereas the intensity distribution of the three classes largely overlapped with one another in the real data groups (Fig. 13). We used a patch-based feature, so the image intensity distribution had a strong correlation with the sample distribution. The distribution of the training data in each submanifold is an important clue for the classification task and can bring discriminative information when classifying a testing sample. Thus, the proposed method with a learned softmax regression model is more applicable for data with complex distribution than data without the learned softmax regression model. The proposed classification method requires no explicit regularization because the patch feature contains the contextual information of a voxel in the image. The proposed method leads a natural smoothness to the segmentation results without explicit regularization by using this contextual information. However, the patch feature may be insufficient to discriminate the brain tumor segmentation task because of the complex characteristics of brain MRI images. Other contextual features may be added in future studies to further improve the classification accuracy obtained in this study. REFERENCES [1]

[2]

[3]

[4]

[5]

[6]

[7]

[8]

[9]

D. A. Gutman, L. A. Cooper, S. N. Hwang, C. A. Holder, J. Gao, T. D. Aurora, W. D. Dunn, Jr., L. Scarpace, T. Mikkelsen, R. Jain, et al., "MR imaging predictors of molecular profile and survival: multi-institutional study of the TCGA glioblastoma data set," Radiology, vol. 267, no. 2, pp. 560-9, May 2013. D. Bhattacharyya and T. H. Kim, "Brain tumor detection using MRI image analysis," Communications in Computer and Information Science, vol. 151, pp. 307-314, 2011. C. L. Biji, D. Selvathi, and A. Panicker, "Tumor detection in brain magnetic resonance images using modified thresholding techniques," Communications in Computer and Information Sicence, vol. 4, pp. 300-308, 2011. P. Gibbs, D. L. Buckley, S. J. Blackband, and A. Horsman, "Tumour volume determination from MR images by morphological segmentation," Physics in Medicine and Biology, vol. 41, no. 11, pp. 2437-46, Nov 1996. R. Ratan, S. Sharma, and S. K. Sharma, "Brain tumor detection based on multi-parameter MRI image analysis," International Journal on Graphics, Vision and Image Processing, vol. 9, no. 3, pp. 9-11, 2009. T. M. Hsieh, Y. M. Liu, C. C. Liao, F. Xiao, I. J. Chiang, and J. M. Wong, "Automatic segmentation of meningioma from non-contrasted brain MRI integrating fuzzy clustering and region growing," BMC Medical Informatics and Decision Making, vol. 11, p. 54, 2011. B. N. Saha, N. Ray, R. Greiner, A. Murtha, and H. Zhang, "Quick detection of brain tumors and edemas: A bounding box method using symmetry," Computerized Medical Imaging and Graphics, vol. 36, no. 2, pp. 95-107, 2012. Z. Iscan, Z. Dokur, and T. Olmez, "Tumor detection by using Zernike moments on segmented magnetic resonance brain images," Expert Systems with Applications, vol. 37, no. 3, pp. 2540-2549, 2010. H. Khotanlou, O. Colliot, J. Atif, and I. Bloch, "3D brain tumor segmentation in MRI using fuzzy classification, symmetry analysis and spatially constrained deformable models," Fuzzy Sets and Systems, vol. 160, no. 10, pp. 1457-1473, 2009.

[10] [10] Z. J. Wang, Q. M. Hu, K. F. Loe, A. Aziz, and W. L. Nowinski, "Rapid and automatic detection of brain tumors in MR images," in Proc. of SPIE, 2004. [11] M. Prastawa, E. Bullitt, S. Ho, and G. Gerig, "A brain tumor segmentation framework based on outlier detection," Med Image Anal, vol. 8, no. 3, pp. 275-83, Sep 2004. [12] M. B. Cuadra, C. Pollo, A. Bardera, O. Cuisenaire, J. G. Villemure, and J. P. Thiran, "Atlas-based segmentation of pathological MR brain images using a model of lesion growth," IEEE Transactions on Medical Imaging, vol. 23, no. 10, pp. 1301-1314, 2004. [13] N. Moon, E. Bullitt, K. V. Leemput, and G. Gerig, "Model-based brain and tumor segmentation," in 16th International Conference on Pattern Recognition, 2002, pp. 528-531. [14] M. R. Kaus, S. K. Warfield, A. Nabavi, P. M. Black, F. A. Jolesz, and R. Kikinis, "Automated segmentation of MR images of brain tumors," Radiology, vol. 218, no. 2, pp. 586-91, Feb 2001. [15] J. Sachdeva, V. Kumar, I. Gupta, N. Khandelwal, and C. K. Ahuja, "A novel content-based active contour model for brain tumor segmentation," Magnetic Resonance Imaging, vol. 30, no. 5, pp. 694-715, Jun 2012. [16] S. Taheri, S. H. Ong, and V. F. H. Chong, "Level-set segmentation of brain tumors using a threshold-based speed function," Image and Vision Computing, vol. 28, no. 1, pp. 26-37, 2010. [17] T. Wang, I. Cheng, and A. Basu, "Fluid vector flow and applications in brain tumor segmentation," IEEE Transactions on Bio-medical Engineering, vol. 56, no. 3, pp. 781-9, Mar 2009. [18] A. Hamamci, N. Kucuk, K. Karaman, K. Engin, and G. Unal, "Tumor-Cut: segmentation of brain tumors on contrast enhanced MR images for radiosurgery applications," IEEE Transactions on Medical Imaging, vol. 31, no. 3, pp. 790-804, Mar 2012. [19] J. Jiang, Y. Wu, M. Huang, W. Yang, W. F. Chen, and Q. J. Feng, "3D brain tumor segmentation in multimodal MR images based on learning population- and patient-specific feature sets," Computerized Medical Imaging and Graphics, vol. 37, no. 7-8, pp. 512-521, Jun 28 2013. [20] Y. Wu, W. Yang, J. Jiang, S. Q. Li, Q. J. Feng, and W. F. Chen, "Semi-automatic segmentation of brain tumors using population and individual information," Journal of Digital Imaging, vol. 26, no. 4, pp. 786-796, 2013. [21] D. Zikic, B. Glocker, E. Konukoglu, J. Shotton, A. Criminisi, D. H. Ye, C. Demiralp, O. M. Thomas, T. Das, R. Jena, et al., "Context-sensitive classification forests for segmentation of brain tumor tissues," in Proceedings of MICCAI-BRATS 2012, France, 2012. [22] D. Zikic, B. Glocker, E. Konukoglu, A. Criminisi, C. Demiralp, J. Shotton, O. M. Thomas, T. Das, R. Jena, and S. J. Price, "Decision forests for tissue-specific segmentation of high-grade gliomas in multi-channel MR," presented at the Med Image Comput Comput Assist Interv, 2012. [23] S. Bauer, T. Fejes, J. Slotboom, R. Wiest, L.-P. Nolte, and M. Reyes, "Segmentation of brain tumor images based on integrated hierarchical classification and regularization," in Proceedings of MICCAI-BRATS 2012, France, 2012. [24] S. Bauer, L. P. Nolte, and M. Reyes, "Fully automatic segmentation of brain tumor images using support vector machine classification in combination with hierarchical conditional random field regularization," presented at the Med Image Comput Comput Assist Interv, 2011. [25] M. Wels, G. Carneiro, A. Aplas, M. Huber, J. Hornegger, and D. Comaniciu, "A discriminative model-constrained graph cuts approach to fully automated pediatric brain tumor segmentation in 3-D MRI," presented at the Med Image Comput Comput Assist Interv, 2008. [26] C. H. Lee, S. Wang, A. Murtha, M. R. Brown, and R. Greiner, "Segmenting brain tumors using pseudo-conditional random fields," presented at the Med Image Comput Comput Assist Interv, 2008. [27] J. J. Corso, E. Sharon, S. Dube, S. El-Saden, U. Sinha, and A. Yuille, "Efficient multilevel brain tumor segmentation with integrated bayesian model classification," IEEE Transactions on Medical Imaging, vol. 27, no. 5, pp. 629-40, May 2008. [28] G. P. Beyer, R. P. Velthuizen, F. R. Murtagh, and J. L. Pearlman, "Technical aspects and evaluation methodology for the application of two automated brain MRI tumor segmentation methods in radiation therapy planning," Magnetic Resonance Imaging, vol. 24, no. 9, pp. 1167-78, Nov 2006. [29] C. H. Lee, M. Schmidt, A. Murtha, A. Bistritz, J. Sander, and R. Greiner, "Segmenting brain tumors with conditional random fields and support vector machines," in Proceedings of the First international conference on Computer Vision for Biomedical Image Applications, 2005, pp. 469-478. [30] V. Harati, R. Khayati, and A. Farzan, "Fully automated tumor segmentation based on improved fuzzy connectedness algorithm in brain

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TBME.2014.2325410, IEEE Transactions on Biomedical Engineering

13

[31]

[32]

[33]

[34]

[35]

[36]

[37]

[38]

[39]

[40]

[41]

[42]

[43] [44]

[45]

[46]

[47]

[48] [49] [50]

[51] [52]

MR images," Computers in Biology and Medicine, vol. 41, no. 7, pp. 483-492, Jul 2011. T. Logeswari and M. Karnan, "An enhanced implementation of brain tumor detection using segmentation based on soft computing," presented at the International Conference on Signal Acquisition and Processing, Bangalore, 2010. M. M. Ahmed and D. B. Mohammad, "Segmentation of brain MR images for tumor extraction by combining kmeans clustering and perona-malik anisotropic diffusion model," International Journal of Image Processing, vol. 2, no. 1, pp. 27-34, 2008. W. B. Dou, S. Ruan, Y. P. Chen, D. Bloyet, and J. M. Constans, "A framework of fuzzy information fusion for the segmentation of brain tumor tissues on MR images," Image and Vision Computing, vol. 25, no. 2, pp. 164-171, 2007. J. Liu, J. K. Udupa, D. Odhner, D. Hackney, and G. Moonis, "A system for brain tumor volume estimation via MR imaging and fuzzy connectedness," Computerized Medical Imaging and Graphics, vol. 29, no. 1, pp. 21-34, Jan 2005. L. M. Fletcher-Heath, L. O. Hall, D. B. Goldgof, and F. R. Murtagh, "Automatic segmentation of non-enhancing brain tumors in magnetic resonance images," Artif Intell Med, vol. 21, no. 1-3, pp. 43-63, Jan-Mar 2001. J. Eugenio Iglesias, M. Rory Sabuncu, and K. Van Leemput, "A unified framework for cross-modality multi-atlas segmentation of brain MRI," Medical Image Analysis, vol. 17, no. 8, pp. 1181-1191, Dec 2013. H. Wang, J. W. Suh, S. R. Das, J. Pluta, C. Craige, and P. A. Yushkevich, "Multi-Atlas Segmentation with Joint Label Fusion," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 3, pp. 611-623, Jun 22 2013. A. Gooya, G. Biros, and C. Davatzikos, "Deformable registration of glioma images using em algorithm and diffusion reaction modeling," IEEE Trans Med Imaging, vol. 30, no. 2, pp. 375-390, 2011. S. Zhang, Y. Zhan, and D. N. Metaxas, "Deformable segmentation via sparse representation and dictionary learning," Medical Image Analysis, vol. 16, no. 7, pp. 1385-96, Oct 2012. H. Lee, D. S. Lee, H. Kang, B. N. Kim, and M. K. Chung, "Sparse brain network recovery under compressed sensing," IEEE Transactions on Medical Imaging, vol. 30, no. 5, pp. 1154-1165, May 2011. J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, "Robust face recognition via sparse representation," IEEE Trans Pattern Anal Mach Intell, vol. 31, no. 2, pp. 210-227, Feb 2009. Y. Gao, S. Liao, and D. Shen, "Prostate segmentation by sparse representation based classification," Medical Physics, vol. 39, no. 10, pp. 6372-6387, Oct 2012. W. Liu, J. F. He, and S.-F. Chang, "Large graph construction for scalable semi-supervised learning," in ICML, 2010. Y. Wu, G. Liu, M. Huang, J. Jiang, W. Yang, W. Chen, and Q. Feng, "Prostate Segmentation based on Variant Scale Patch and Local Independent Projection," IEEE Transactions on Medical Imaging, 2014. (Online) B. H. Menze, A. Jakab, S. Bauer, J. Kalpathy-Cramer, K. Farahani, J. Kirby, Y. Burren, N. Porz, J. Slotboom, R. Wiest, et al., "The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS)," IEEE Transactions on Medical Imaging, 2014. (Submitted) J. Macqueen, "Some Methods for Classification and Analysis of Multivariate Observations," in Fifth Berkeley Symp. on Math. Statist. and Prob., 1967, pp. 281-297. J. J. Wang, J. C. Yang, K. Yu, F. J. Lv, T. Huang, and Y. H. Gong, "Locality-constrained linear coding for image classification," presented at the Computer Vision and Pattern Recognition, 2010. K. Yu, T. Zhang, and Y. H. Gong, "Nonlinear learning using local coordinate coding," presented at the NIPS, 2009. Y. Nesterov, Introductory lectures on convex optimization: A basic course: Kluwer Academic Publishers, 2003. J. G. Sled, A. P. Zijdenbos, and A. C. Evans, "A nonparametric method for automatic correction of intensity nonuniformity in MRI data," IEEE Transactions on Medical Imaging, vol. 17, no. 1, pp. 87-97, Feb 1998. C. Cortes and V. Vapnik, "Support-Vector Networks," Machine Learning, vol. 20, no. 3, pp. 273-297, 1995. N. Cordier, B. Menze, H. Delingette, and N. Ayache, "Patch-based Segmentation of Brain Tissues," in Proceedings of MICCAI-BRATS 2013, 2013.

0018-9294 (c) 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Brain tumor segmentation based on local independent projection-based classification.

Brain tumor segmentation is an important procedure for early tumor diagnosis and radiotherapy planning. Although numerous brain tumor segmentation met...
1MB Sizes 0 Downloads 4 Views