Next Article in Journal
Classification of Clouds in Satellite Imagery Using Adaptive Fuzzy Sparse Representation
Next Article in Special Issue
Hyperspectral Imagery Super-Resolution by Adaptive POCS and Blur Metric
Previous Article in Journal
MB-OFDM-UWB Based Wireless Multimedia Sensor Networks for Underground Coalmine: A Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data Field Modeling and Spectral-Spatial Feature Fusion for Hyperspectral Data Classification

School of Electronic, Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Sensors 2016, 16(12), 2146; https://doi.org/10.3390/s16122146
Submission received: 13 November 2016 / Revised: 6 December 2016 / Accepted: 12 December 2016 / Published: 16 December 2016
(This article belongs to the Special Issue Multi-Sensor Integration and Fusion)

Abstract

:
Classification is a significant subject in hyperspectral remote sensing image processing. This study proposes a spectral-spatial feature fusion algorithm for the classification of hyperspectral images (HSI). Unlike existing spectral-spatial classification methods, the influences and interactions of the surroundings on each measured pixel were taken into consideration in this paper. Data field theory was employed as the mathematical realization of the field theory concept in physics, and both the spectral and spatial domains of HSI were considered as data fields. Therefore, the inherent dependency of interacting pixels was modeled. Using data field modeling, spatial and spectral features were transformed into a unified radiation form and further fused into a new feature by using a linear model. In contrast to the current spectral-spatial classification methods, which usually simply stack spectral and spatial features together, the proposed method builds the inner connection between the spectral and spatial features, and explores the hidden information that contributed to classification. Therefore, new information is included for classification. The final classification result was obtained using a random forest (RF) classifier. The proposed method was tested with the University of Pavia and Indian Pines, two well-known standard hyperspectral datasets. The experimental results demonstrate that the proposed method has higher classification accuracies than those obtained by the traditional approaches.

1. Introduction

With the development of imaging instruments in the past few years, hyperspectral data processing has become increasingly more important in many fields [1,2,3,4,5]. As a data tool with high spectral resolution, hyperspectral sensors usually utilized hundreds of spectral channels to describe spectral signatures. Generally, the primary purpose of hyperspectral images (HSI) processing is to analyze and recognize spectral data acquired by hyperspectral sensors. It is established that, different materials have distinct reflectance spectral signatures. Thus, reflectance spectra are always used for material recognition and image analysis [6].
However, while the high dimensionality of HSI supports accurate descriptions for spectral signatures, they lead to some theoretical and practical problems, particularly the curse of dimensionality problem. In classification problems, classification accuracies are not positively correlated to the dimensionality of input data. Usually, classification is most accurate with a particular feature number, as has been demonstrated in References [7,8,9]. Hence, feature extraction and dimensionality reduction techniques are important and indispensable in high-dimensional data classification and analysis. Based on known information, feature extraction (FE) techniques are generally categorized into unsupervised and supervised methods. Unsupervised FE techniques, e.g., principle component analysis (PCA) [10], are always used for data description and representation. Supervised FE focuses on reducing the dimensionality of data to achieve better classification performance and avoid Hughes phenomena [7]. Many supervised feature extraction algorithms have been proposed and widely used in hyperspectral image processing, such as the discriminant analysis feature extraction (DAFE) algorithm [11], the decision boundary feature extraction (DBFE) approach [12], and the nonparametric weighted feature extraction (NWFE) method [13], etc.
Conventionally, HSI are treated by classifiers as spectral data cubes and a set of spectral measurements without spatial structure [14]. Hence, the spatial structure features in HSI are discarded. However, with the development of sensors, HSI usually provides both detailed spatial structural and spectral information. Crisp and adaptive neighborhood systems are commonly used to characterize spatial structural features [15]. The crisp system generally analyzes the spatial structure based on a tough neighborhood. The crisp system is widely used in spatial information extraction. However, it has the following limitations: (1) the classifier effectiveness may be influenced by the predefined neighborhood system without enough samples; and (2) a large neighborhood system usually results in computation problems [16]. For this reason, adaptive neighborhood systems are also taken into account. Based on the morphology theory [17], which has been widely used in image processing, a set of methods for spatial information extraction using adaptive neighborhood systems [18,19,20,21,22,23,24] have been proposed.
Morphological profiles (MPs) [18] have demonstrated their usefulness in spatial structure description. The sizes of different structures in an image can be determined by using geodesic opening/closing through reconstruction [19,20]. For any given size of a structuring element (SE), the structures that are smaller than the SE are removed, while larger structures are preserved. The spatial information of the image is extracted by applying such operators with an SE range of different sizes. This concept is usually called granulometry [21]. The attribute profiles (APs) technique [22] is a further development of MPs based on attribute filters, which allow for the modeling of geometrical characteristics. Compared with MPs, APs allow more precise modeling of spatial information. This is because an input image can be processed based on multiple attributes, by which different aspects of spatial structures can be described with great flexibility. When dealing with vectorial images, typically HSI, the application of morphological filters has been extended based on the concept of the vectorial image profile. Extended morphological profiles (EMPs) [23] and extended attribute profiles (EAPs) [21] were proposed to extract the spectral and spatial features of the hyperspectral data. In References [21,23], PCA was first implanted in original hyperspectral data, and the first principle components that contained particular cumulative variance were selected as the baseline images. Then MPs and APs were performed on all the selected PCs. EMP and EAP were composed by these MPs and APs, respectively. In later studies, an extended multi-attribute profile (EMAP) was proposed in References [15,24]. EMAP, which utilizes multi-attributes, is a more advanced version of EAP. Additionally, Reference [25] proposed a supervised feature selection approach in attribute profiles on the basis of a genetic algorithm (GA). By introducing the GA technique, the EMAPs with the highest importance are preserved for classification. In References [15,25], supervised FE techniques were used to create better profiles and extract more discriminate spatial features. In Reference [26], a state-of-the-art hyperspectral classification based on sparse representation and EMAPs was proposed. Based on the fact that the extracted EMAPs with high dimensionality should have particular class-dependent manifold structures, this classification approach exploits the inherent characteristics of EMAPs embedded in high-dimensional feature space. This method, called SUnSAL in Reference [26], combines the benefits of sparse representation and the rich spatial structural information obtained by EMAPs.
In order to consider both the spectral features and spatial features, spectral-spatial classifiers have become increasingly important in HSI classification. A few studies, such as References [27,28], have proposed several spectral-spatial FE methods based on supervised FE techniques and morphological filters for HSI classification. In References [27,28], the spectral and spatial features were extracted using supervised FE approaches and morphological filters, and then the extracted spectral features and spatial features were fused via vector stacking. Thus, both spectral and spatial information were utilized in classification. Reference [29] extracted the local image structures by employing local binary patterns (LBP). LBP features were extracted on all selected spectral bands. Next, the local image patterns and spectral features were fused both at the feature and decision level for classification. In Reference [30], a spectral-spatial method using multi-hypothesis (MH) prediction for noise-robust HSI classification was proposed. By using a weighted regularization, the MH prediction finds the best linear hypothesis combination and achieves spectral-spatial classification. Inspired by the deep leaning idea, a deep feature extraction algorithm based on convolution neural networks was presented in Reference [31]. However, it is important to note that the inner relationship between spatial and spectral features has received little attention. In order to further improve the classification accuracy, new information must be introduced and explored, particularly the information hidden in the relationship between the spectral and spatial information. References [32,33,34] have demonstrated that spatial neighbors always contribute to the measured signal through adjacency effects. Hence, spectral and spatial features are not independent due to data interaction in HSI.
In this study, a supervised spectral-spatial classification algorithm based on data field theory is proposed. This algorithm improves classification accuracy by further processing the extracted spectral and spatial information. Unlike the current classification approaches, the proposed method aims to further improve classification performance by exploring the inner relationship between spectral and spatial information. The main motivation of the proposed method is that data influences and interactions should be taken into consideration, which is often neglected in spatial-spectral classification tasks. By considering the mutual influences and interactions between pixels, we attempt to build the connection between the spectral and spatial domains. So, more useful information hidden in the relationship between spectral and spatial information, or, for simplicity, the adjacency effects, can be explored and included for classification. In our study, spectral information was extracted by supervised FE techniques and spatial information was generated by EMAP, as performed in Reference [27]. Next, data field modeling was applied to both the spectral and spatial domains. Based on data field modeling, the spectral and spatial information are unified. So, the unified radiation features containing both spectral and spatial information can be fused into a new radiation feature by using a linear model. Another advantage of data field modeling in both spectral and spatial domains is that the problem of the extracted spectral and spatial features having different scales can be avoided. An random forest (RF) classifier provides a final classification map [35]. The novelty of the proposed algorithm lies in its use of data field theory to explore the relationship between the spatial and spectral information. To measure the efficacy of the presented method, we tested it by using two standard hyperspectral datasets. In the remainder of this paper, Section 2 covers the detailed presentation of the proposed algorithm. Section 3 presents a series of experiments with two standard HSI test datasets. In Section 3, the experimental results of different test cases are analyzed and key parameters used in the proposed method are discussed. The advantages of the proposed approach and proposed subjects for future investigation are drawn in Section 4, followed by the conclusions in Section 5.

2. The Proposed Method

2.1. Data Field Modeling

Data fields are the mathematical expression of field theory in physics. Data fields establish models in which data can be seen as a whole by studying the interactions of data. To describing the relationship between data, data are treated as radiation sources within the data field. Thus, the radiation effect can be used to mathematically describe the data interaction. Employing this approach, the property of a vector point is determined not only by its location in the data space, but also by the other surrounding data in the data field owing to the radiation effect. In this paper, both the spectral and spatial domains of an HSI are considered as data fields. Thus, the recognition and identification of a pixel in a HSI do not depend only on its position in the spectral space—on its spectral signature for simplicity—but also take into account its interactions with the other pixels in the HSI.
In this paper, we define the radiation intensity as a function that depends on a distance measurement. The function is called the radiation function, and is mathematically expressed as:
E ( ρ , d ) = E 0 exp { ρ d 2 }
where d denotes the distance to the radiation source, E is the radiation intensity at d , ρ is a radiation factor and E 0 indicates the initial energy. Both Mahalanobis and Euclidean distances are employed as the distance measurements in this paper. We term the Mahalanobis distance d M and the Euclidean distance d E . Apparently, while d is small, the points in a data pair interact with each other intensively. In contrast, the e ρ d 2 term tends toward zero and the interaction is negligible when d is large. The radiation function allows us to establish the connections between the data in data fields and to describe the interactions between the data pairs as radiation intensities.
Suppose x = [ x φ , x ω ] T is a feature vector that corresponds to a pixel in HSI; here, x φ represents the spectral feature extracted by supervised FE techniques, and x ω denotes the spatial structural feature. In the following description, the symbols related to spectral space are denoted by suffix φ and those related to spatial space are denoted by ω . Thus, a pixel in HSI corresponds to a feature vector x φ in the spectral feature space R φ , and a feature vector x ω in the spatial feature space R ω . In this paper, both R φ and R ω are considered data fields. Thus, a data point receives radiations in both R φ and R ω . Furthermore, we suppose that all the data have a unit initial radiation energy, i.e., E 0 = 1 , when data field modeling in both the spectral and spatial domains.
Suppose a training sample set { ( x i , u i ) } i = 1 N , where x i = [ x i , φ , x i , ω ] T denotes an input pattern, u i { 1 , ... , L } denotes its class label, and N and L are the numbers of the training samples and classes, respectively. For training sample ( x i , l ) , according to the label u i = l , two subsets of the training set are defined. The first subset contains all the training samples that have the same label as x i , and we term this subset the Same Class Subset. The other subset contains all the training samples with class labels different from x i , and is called the Different Class Subset. We suppose that a given training sample x i receives radiation from its k-nearest training samples in every class. For example, { ( s j , v j ) } j = 1 k × L denotes the set of nearest neighbors (NNs) with respect to ( x i , l ) , and s j is the jth nearest neighbor (NN) with a class label v j , where j = 1 , ... , k × L , v j = 1 , ... , L . We have x i = [ x i , φ , x i , ω ] T and s j = [ s j , φ , s j , ω ] T . Then, radiations from s j to x i in R φ and R ω are respectively defined as:
{ e j , ω = exp { ρ ω l d M ( x i , ω , s j , ω ) 2 } e j , φ = exp { ρ φ l d M ( x i , φ , s j , φ ) 2 }
Here, ρ φ l and ρ ω l are the radiation factors of ( x i , l ) in the spectral domain data field and spatial domain data field, respectively, and d M ( ) denotes the Mahalanobis distance. It should be noted that we used different radiation factors in different spaces and classes. The radiation factors can be determined by the training samples, and will be discussed in the following section. Consequently, x i is projected as:
x i x i e = [ e 1 , φ 1 , ... , e N c , φ k , e 1 , ω 1 , ... , e N c , ω k ] T m
where m denotes an m-dimensional space, m = 2 × N k , and N k = k × N c is the number of the NNs in all classes. We term x i e the data field radiation feature (DFRF).
In this paper, we define the total radiation as a weighted addition of the radiations in both R φ and R ω , i.e., the total radiation x i received from s j can be defined as:
e j = α e j , φ + ( 1 α ) e j , ω ,   0 < α < 1
where 0 < α < 1 is a weight coefficient. In Equation (4), the first term on the right indicates the spectral information of the input pattern describing the radiation received in the spectral feature space. The second term represents the radiation in the spatial feature space that can be seen as the spatial information. The weight coefficient α is used to describe the inner connection between spectral and spatial features. Hence, Equation (4) contains the spatial features, spectral features and the spectral-spatial information relationship. The left term in Equation (4) fuses the radiation in the spectral and spatial spaces into a total radiation feature. In essence, the data modeling operation is a feature-unifying and fusion procedure. Following the data field modeling approach presented here, the data radiation interactions are built in both the spatial and spectral domains. Hence, the spectral and spatial feature domains are unified and correlated. Consequently, the spatial and spectral information are unified and fused through the data field modeling, and x i is transferred to y i which is termed as the fused data field radiation feature (FDFRF):
x i y i = [ e 1 1 , ... , e 1 k , ... , e N c 1 , ... , e N c k ] T N k

2.2. Weight Coefficient Training

The weight coefficient α describes the inner connection between spectral and spatial features. We discuss the method of determining the value of the coefficient in this section. For a given training sample ( x i , u i ) , the k 1 -nearest-neighbors ( k 1 -NN) in its Same Class Subset are selected and denoted by x i 1 , , x i k 1 , while k 2 -NN in the Different Class Subset are selected and denoted by x i 1 , , x i k 2 . Then, the data patch of x i can be built as:
X i = [ x i , x i 1 , ... , x i k 1 , x i 1 , ... , x i k 2 ]
The corresponding DFRF and FDFRF are, respectively:
X i e = [ x i e , x i 1 e , ... , x i k 1 e , x i 1 e , ... , x i k 2 e ]   and   Y i = [ y i , y i 1 , ... , y i k 1 , y i 1 , ... , y i k 2 ]
It is easy to derive that Y i = ( α A + B ) X i e , where, A = [ I | I ] , B = [ 0 | I ] and I is an identity matrix. For the FDFRFs in each patch, we want the distances between y i and y i 1 , , y i k 1 to be as small as possible. Meanwhile, the distances between y i and y i 1 , , y i k 2 are as large as possible. So, we have:
arg min y i   ( j = 1 k 1 d E ( y i , y i j ) 2 β p = 1 k 2 d E ( y i , y i p ) 2 )
where β [ 0 , 1 ] is a scaling factor, and d E ( ) represents the Euclidean distance. Local information is introduced to train the weight coefficient. The local information specifies the subspaces in which the boundary regions are embedded and deemphasizes those samples far from the boundaries. We define the λ i as:
λ i = d E ( x i e , M i e ) / [ d E ( x i e , M i e ) + min 1 j k 2 d E ( x i e , x i j e ) ]
where λ i [ 0 , 1 ] describes the local information, M i e is the center of x i 1 e , x i k 1 e , and min 1 j k 2 d E ( x i e , x i j e ) denotes the minimum distance from x i e to x i 1 e , , x i k 2 e . Then Equation (8) can be changed into:
arg min y i   λ i ( j = 1 k 1 d E ( y i , y i j ) 2 β p = 1 k 2 d E ( y i , y i p ) 2 )
Furthermore, we define a coefficients vector and patch matrix:
ω i = [ 1 , ... , 1 k 1 , β , ... , β k 2 ] T   and   L i = [ k = 1 k 1 + k 2 ( ω i ) j ω i T ω i d i a g ( ω i ) ]
where diag ( ) is the diagonalization operation. Then Equation (10) can be reduced to:
arg min Y i   t r ( λ i Y i L i Y i T )
where tr ( ) is the trace operator. Furthermore, all the Y i are taken into account, and then
arg min Y i = 1 N t r ( λ i Y i L i Y i T ) = arg min α [ t r ( A G A T ) α 2 + t r ( A G B T + B G A T ) α + t r ( B G B T ) ]
where G = i = 1 N ( λ i X i e L i X i e T ) and weight coefficient α can be uniquely determined. Hence, the spectral-spatial relationship is described and hidden information is explored. It can be seen from Equation (13) that the weight coefficient training is actually an additional information extraction operation. In other words, the most discriminative features in the spectral and spatial features are extracted by introducing α in this procedure.
The implementation scheme of the proposed algorithm for hyperspectral imagery classification is shown in Figure 1. As shown the data field modeling operation is implemented in both the spectral space and image spatial domain. Based on the prior information provided by the training set, which consists of spectral information, local spatial information, and label information, the spectral features can be obtained by supervised FE techniques. The spatial structural features can be extracted by the spatial feature extraction algorithms, such as EMP, EAP, and EMAP. The data field modeling operation is carried out in the two spaces, and then the DFRF is built. The feature fusion with local information is then performed. This process fuses the spectral and spatial features into an FDFRF, and learns the fusing weight coefficient. For an unlabeled test pixel, we extract the spectral and spatial features. Then, the extracted features are fused into an FDFRF based on data field modeling. Finally, the classification is implemented by classifiers.

3. Experiments and Results

Two standard datasets, the Reflective Optics Systems Imaging Spectrometer (ROSIS-03) University of Pavia dataset and the Airborne Visible Infrared Imaging Spectrometer (AVIRIS) Indian Pines dataset, which are frequently used in research, were used in this study.
The first test dataset is a hyperspectral dataset collected from the University of Pavia, Italy, by the ROSIS-03 airborne instrument. In this dataset, nine classes of interest were considered in the image scene. This dataset, which is composed of 103 bands of 610 × 340 pixels, provides a high spatial resolution of 1.3 m/pixel. The training and test sets were composed of 3909 and 42,788 samples, respectively. The number of training and test samples is shown in Table 1.
The Indian Pines dataset is a standard test dataset acquired in 1992 using the AVIRIS sensor. The data consists of 145 × 145 pixels with a medium spatial resolution of about 20 m/pixel. In this test case, the spectral channels in the atmosphere absorption bands were removed, so 200 data channels were used. Sixteen classes of interest were considered. For this dataset, a total of 695 pixels and 9671 pixels were used to make up the training and test sets, respectively. The number of available test and training samples is displayed in Table 2.
The details of the training and test sets of the two datasets are given in References [27,36]. To maintain consistency with previous results, we used the same size training and test sets adopted by other state-of-the-art approaches. We also adopted samples with precisely the same spatial locations as in the previous studies. Each method was executed only once because the samples that we used were identical to those used in the previous studies. False color images of the two datasets are presented in Figure 2.

3.1. Experimental Setup

In all the experimental datasets, the spectral-spatial classification method η n which was proposed in Reference [27], known as AUTOMATIC, was employed for comparison. The FE approach used is denoted by n. Here, the HSI data were first transformed by the FE approach. The spectral feature x φ was the output of this step. Next, the spatial feature x ω was obtained by EMAP and the FE approach. Finally, x φ and x ω were stacked together for classification. DAFE and DBFE were employed for supervised FE. DAFE is often applied to dimension reduction and feature extraction in a pattern recognition field. The class centers and covariance matrix of each class are calculated by training samples in DAFE. As a parametric method, DAFE achieves a satisfactory performance if the data approximately follow a normal distribution. DBFE extracts both discriminately informative and redundant features from the decision boundary. Using the decision boundary feature matrix, the decision boundary is described and features are extracted. For example, η D A denotes that the raw data were first transformed by DAFE. Then the EMAP was performed on the baseline images obtained by DAFE. Finally, the spectral features extracted by DAFE and the spatial features obtained by EMAP were stacked together.
It should be emphasized that the features containing more than 99% of cumulative eigenvalues were selected when DAFE and DBFE were employed in the following experiments. The classification results obtained by using the spectral information were reported only for comparison. We use DA and DB to indicate the spectral information extracted by DAFE and DBFE, respectively. The EMAP methods were also employed to demonstrate the superiority of the proposed algorithm. The DAp and DBp denote the EMAPs that were generated based on the features extracted by DAFE and DBFE, respectively. The EMAP-based classification methods proposed in References [25,26], which were respectively denoted by GA and SUnSAL, were employed. The recent state-of-the-art spectral-spatial classification approaches, including MH [29] and LBP [30], were used for comparisons. For the MH approach, the hypotheses for prediction were generated using the manually selected spectral-band partitions as suggested in [29]. In the LBP method, the criterion of linear prediction error (LPE) [37] was used for spectral band selection, and LBP features were extracted on these selected bands. Then, the LBP features and selected spectral bands were fused at the feature level, and processed by the classifier. To make our methods fully comparable with the reference techniques, the thresholds and values used for this experimental setup were selected from References [15,27].
The term n signifies our proposed method. The FE approach is denoted by n. The spectral feature x φ and spatial feature x ω were fused into FDFRF in our proposed method. In the experiments, we set k = 5 , i.e., five NNs in each class were considered in the data field modeling. The features extracted by all the methods were analyzed by an RF classifier. In all the experiments, the number of trees was set to 200, as suggested in References [15,35,36], in order to achieve a trade-off between the classification performance and time cost for the learning phase. The method performances were evaluated by three measurements: the overall accuracy (OA), the average accuracy (AA), and the Kappa coefficient ( κ ). However, in order to avoid unnecessary redundancy in the following, the experimental results and comparison will only be analyzed based on OA.

3.2. Results

As shown in Table 3 and Table 4, the results of our experiments with the two datasets show that feature fusion based on the data field theory can improve classification accuracy compared to the reference methods. The classification results acquired by the proposed method on the two datasets by the proposed method are shown in detail in Figure 3 and Figure 4.
For the University of Pavia dataset, the data field feature fusion resulted in significantly improved classification accuracy. As can be observed from Table 3, DB outperformed the other methods with an OA of 99.4%. DA achieved 19%, 4.1%, and 13.4% improvement in OA over DA, DAp and η D A , respectively. Compared with the corresponding reference DB, DBp and η DB methods, DB improved the OA by 20.5%, 3.4%, and 2.6%, respectively. It is also important to emphasize that η DB exhibited excellent classification performances with an OA of 96.8%. In comparison, DA and DB achieved small improvements in OA of 2.1% and 2.6%, respectively. Although the improvements in classification accuracy are not remarkable in the manner of OA, more than 65.6% and 81.2% of test samples misclassified by η D B were corrected by DA and DB , respectively. We can therefore conclude that the proposed method effectively improved the classification performance.
Compared with the results reported in Table 3, it is easy to deduce that DBFE outperforms DAFE. The primary reason may be that DAFE is not full rank, so that some discriminative spectral information was lost. It should be noted that the classification performances of AUTOMATIC, which stacked the spectral and spatial features together, were affected by different FE approaches. The OA resulting from η DB is 11.3% more than that of η DA . Compared to the EMAP approaches, AUTOMATIC improved the classification accuracy when DBFE was employed. However, AUTOMATIC classification decreased when DAFE was performed. The proposed method is much more robust with respect to the choice of the FE technique. Classification results always remained at a high level when different FE approaches were used. This is because our method further fused the extracted spectral and spatial features. As a result, the useful information that lies in the spectral-spatial relationship and can contribute to the classification was included.
Compared with the employed state-of-the-art HSI classification methods, the proposed method additionally achieved competitive classification performance in this test case. D B achieved the best classification results in terms of OA, AA and the κ value. As can be observed from the classfication results, DB achieved approximately 3.3%, 1.3%, 0.6% and 0.2% improvements in OA over GA, SUnSAL, LBP and MH, respectively. Though the OA improvements are seemingly very small, almost 84.6%, 68.4%, 50% and 25% of misclassified samples in these methods were corrected, respectively. Moreover, DA also produced a satisfactory classification performance with an OA of 98.9%. Although the MH approach reported a higher classification accuracy with an OA of 99.2%, DA is competitive because it perfomed better than all the other reference methods.
In contrast to the University of Pavia dataset, the low spatial resolution, which leads to more mixed pixels, makes the classification task more complex in the Indian Pines dataset. For this test case, the HSI classification results, obtained by further feature fusion, were generally better than the corresponding compared methods. For example, DA achieved 33.3%, 5.7%, and 3.5% improvements in OA over DA, DAp, and η DA , respectively. DB improved the OA of DB, DBp, and η DB by 31.9%, 5.4%, and 11.9%, respectively. The best accuracies were obtained by using DA which achieved an OA of 96.8%. It should be noted that reference methods exhibited acceptable performances in terms of classification accuracies. In contrast, the DA achieved the best performances in 11 classes and DB performed better than all the reference methods in 11 classes. As the results represented in Table 4 show, DAFE performs better than DBFE in terms of OA, AA, and the Kappa coefficient. A possible reason may be that the presence of the pixels with mixed spectra leads to the features number extracted by DBFE being insufficient to discriminate the samples in different classes.
In the Indian Pines dataset, the results also indicate that the AUTOMATIC approach is affected by different FE methods. Our method avoided this problem by using data field modeling and further feature fusion. As can be observed from the classification results reported in Table 4, the state-of-the-art spectral-spatial methods improved the classification more significantly than the spectral-based methods DA and DB in this test case. This may be because the spectral information is less dominant in this test case and introducing spatial information effectively contributes to the classification problem. As with the Pavia University dataset, our method obtained competitive results for this dataset in comparison to the other state-of-the-art methods. The best classification result was obtained by DA with an OA of 96.8%, and the missclassified rates decreased approximately 48.4%, 46.6%, 52.2% and 25.6% compared to GA, SUnSAL, LBP and MH, respectively. Moreover, DB also performed competitively with better classification accuracies than the other reference methods, except for MH.
As Equation (5) shows, the feature number (i.e., the dimensionality of the FDFRFs) in our method is determined by the number of the classes and NNs used in the data field modeling. The feature numbers of our method were 45 and 80 using the Pavia University dataset and Indian Pines dataset, respectively. The proposed method can be seen as an advancement of the AUTOMATIC approach. Accordingly, the feature numbers of the proposed method and AUTOMATIC are listed in Table 5. It can be seen from Table 5 that the proposed method achieved better classification results with acceptable feature numbers. Compared to the EMAP reference methods, our methods effectively reduced the feature numbers and improved classification accuracy. Moreover, DB (consisting of 45 features) performed better than η DB , which consisted of 59 features in the Pavia University dataset. In the Indian Pines dataset, the proposed method also showed superior classification performance over AUTOMATIC approaches with an acceptable feature number.
Finally, we compare the computational complexity of the classification methods. As an example, the processing times (in seconds) of the methods with the Indian Pines dataset are shown in Table 6. All experiments were implemented using MATLAB on an Intel Core i5 CPU with 3.2 GHz and 4 GB of RAM. As can be seen in Table 6, the DAFE-based methods have an obvious advantage in computational time compared to DBFE-based approaches because DAFE is faster than DBFE. The computational costs of data field-based methods are higher than those of the corresponding AUTOMATIC approaches owing to the burden of building FDFRFs. Compared to the other methods, our method achieved superior classification performances at the cost of greater computational complexity and time consumption. However, the speed of our method could be improved by using time-efficient feature extraction approaches and parallel computing techniques.

3.3. Parameters

In this section, two important parameters used in the presented algorithm are discussed. First, the radiation factors used in the radiation function are analyzed, and an adaptive method for determining the radiation factor is put forward. Secondly, the relationship between the algorithm performance and k , the number of NNs used in data field modeling, is discussed.
As shown in Equation (1), the radiation intensity is jointly determined by the distance measurement d and the radiation factor. Radiation factors determine the character of the radiation effects in data fields or, for simplicity, the range of the data radiation domain. The distance measurement can lose meaning when ρ is extremely small or large. The data interact strongly when ρ is very small, whereas, the interactions between data are negligible if ρ is very large. Additionally, as before, we use different radiation factors in different spaces and classes when calculating radiation intensities. In this study, the values of the radiation factors were determined by the training samples. For a given training sample ( x i , l ) , the training set can be divided into two parts, as mentioned in Section 2. The vector mean value of the Same Class Subset spectral features is denoted by x ¯ i , φ , which can be considered as the center of the class l in the spectral feature space. It is desirable for the training samples in the same class and different classes to have as strong and weak radiations as possible, respectively, i.e.,:
ρ φ l * = arg max ρ ( e ρ d + , φ 2 e ρ d , φ 2 ) ρ φ l * = 2 ( ln d + , φ ln d , φ ) / ( d + , φ 2 d , φ 2 )
where d + , φ is the mean value of the distances between x ¯ i , φ and the samples in the Same Class Subset, and d , φ is the mean value of the distances from x ¯ i , φ to the samples in the Different Class Subset. Therefore, ρ φ l (i.e., the radiation factor of ( x i , l ) in the spectral domain data field) can be adaptively determined by the training samples. The radiation factor of ( x i , l ) in the spatial domain data field, which is denoted by ρ ω l , can be determined in the same way.
In our proposed method, the number of NNs k is the most important parameter in determining the data field modeling accuracy and classification performance. The influence of k on the algorithm performance, measured by OA, can be observed in Figure 5. Note that OA increases with k. However, the classification performance decreases when k > 5 and k > 10 in the Indian Pines dataset and Pavia University dataset, respectively. This is because a large k may lead to a higher dimension of FDFRF, which may cause the Hughes phenomenon. Moreover, a large k also brings a higher computation cost. Based on our experimental results, it is reasonable to set k = 5, which avoids the Hughes phenomenon and achieves a good trade-off between the classification performance and computation cost.

3.4. Experiments Using Reduced Training Samples

As can be observed in Table 1 and Table 2, a large training set with 3909 training samples was used in the University of Pavia test case and a relatively small training set was employed in the Indian Pines dataset with 15 or 50 training pixels per class. In order to further validate the classification performance using a small training sample size, an additional experiment was performed using the Pavia University dataset with a reduced number of training samples. In this experiment, 30 training samples per class were randomly selected from the provided 3909 training samples to form the small training sample set. Table 6 reports the classification OA, AA, κ value, and individual class accuracies achieved by different approaches. The classification maps acquired by our proposed method using the small training sample size are shown in Figure 6. As can be observed in Table 7, DB and LBP achieved the best classification performance in terms of OA, with an OA of approximately 96.6%. However, LBP performed better in terms of the AA and κ value and obtained the smallest degradation in OA. The reason might be because the LBP approach can extract the detailed local image characteristics, such as corners, edges and knots. Hence, it is more efficient and robust in describing spatial features than EMAP-based methods, particularly in the small training sample size case. DA also demonstrated a competitive performance under the small training sample size. Compared with all the reference methods except LBP and DB , DA obtained higher classification accuracies. Therefore, it can be concluded that our proposed technique can achieve satisfactory classification results with limited training data.

4. Discussion

The experimental results demonstrate that feature fusion can further promote accurate classification performance. Compared to the reference methods, which simply fused the extracted features via vector stacking, the proposed method further fused the spectral and spatial information through the introduction of data field theory. A relationship between the spectral and spatial features was built and previously hidden information was explored. It can be concluded from our results that our method fused the spectral and spatial features in a reasonable and effective way. Furthermore, the proposed method is robust to the FE approaches, which is also desirable.
Two standard hyperspectral data sets were employed to measure the efficacy of our proposed method. The two test cases represent two typical types of classification problems. The Pavia University dataset covers an urban area with both high spectral and spatial resolution. It is a typical urban classification problem. The Indian Pines dataset, with relatively low spatial resolution, represents agriculture land-cover problems. The experimental results obtained on both datasets demonstrate that our proposed method is generally applicable to different classification problems.
A subject for future investigation is the optimization of data field modeling based on an imaging mechanism. The fusing model used in this paper is a linear weighted addition model. A more reasonable and effective model will be studied in future research. Another subject that deserves further research is the adaptive selection of the number of NNs used in the data field modeling.

5. Conclusions

In this study, a feature fusion method based on data field theory was proposed to carry out the supervised classification of HSI. As a mathematical realization of field theory concepts in physics, data field theory was employed to establish data field modeling in HSI. Both the spectral features and spatial space were considered data fields. The fusion weight coefficient was trained based on the data field modeling. Thus, a relationship between the spectral and spatial feature was constructed, and the two features were fused into a discriminative FDFRF. The weight coefficient training procedure was a further feature extraction process. The relationship between the spectral and spatial information was explored and the method was shown to achieve improved classification performance.

Acknowledgments

This work was jointly supported by the National Natural Science Foundation (61673265); The 973 Project (6133190302); the Shanghai Aerospace Science and Technology Innovation Fund (SAST201448); the Aeronautical Science Foundation of China (20140157001); 2015 Industry-university-research cooperation project of AVIC.

Author Contributions

Da Liu and Jianxun Li conceived and designed the study. Da Liu performed the experiments and wrote the paper. Jianxun Li reviewed and edited the manuscript. All authors read and approved the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Z.; Xu, D.; Guo, X. Remote Sensing of Ecosystem Health: Opportunities, Challenges, and Future Perspectives. Sensors 2014, 14, 21117–21139. [Google Scholar] [CrossRef] [PubMed]
  2. Andrew, M.E.; Ustin, S.L. The role of environmental context in mapping invasive plants with hyperspectral image data. Remote Sens. Environ. 2008, 112, 4301–4317. [Google Scholar] [CrossRef]
  3. Bannari, A.; Pacheco, A.; Staenz, K.; McNairn, H.; Omari, K. Estimating and mapping crop residues cover on agricultural lands using hyperspectral and IKONOS data. Remote Sens. Environ. 2006, 104, 447–459. [Google Scholar] [CrossRef]
  4. Eismann, M.T.; Meola, J.L.; Stocker, A.D.; Beaven, S.G.; Schaum, A.P. Airborne hyperspectral detection of small changes. Appl. Opt. 2008, 47, F27–F45. [Google Scholar] [CrossRef] [PubMed]
  5. Wu, J.-L.; Ho, C.-R.; Huang, C.-C.; Srivastav, A.L.; Tzeng, J.-H.; Lin, Y.-T. Hyperspectral Sensing for Turbid Water Quality Monitoring in Freshwater Rivers: Empirical Relationship between Reflectance and Turbidity and Total Solids. Sensors 2014, 14, 22670–22688. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Chang, C.I. Hyperspectral Imaging: Techniques for Spectral Detection and Classification; Plenum Publishing Co.: New York, NY, USA, 2003. [Google Scholar]
  7. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 14, 55–63. [Google Scholar] [CrossRef]
  8. Landgrebe, D.A. Signal Theory Methods in Multispectral Remote Sensing; Wiley: Hoboken, NJ, USA, 2003. [Google Scholar]
  9. Ghamisi, P.; Couceiro, M.S.; Benediktsson, J.A. Classification of hyperspectral images with binary fractional order Darwinian PSO and random forests. In SPIE Remote Sensing; International Society for Optics and Photonics: Bellingham, WA, USA, 2013; p. 88920S-88920S-8. [Google Scholar]
  10. Jolliffe, I.T. Principal Component Analysis; Springer: Berlin, Germany, 1986; Volume 87, pp. 41–64. [Google Scholar]
  11. Kuo, B.C.; Landgrebe, D.A. Nonparametric Weighted Feature Extraction for Classification. In Proceedings of the Ai 2005: Advances in Artificial Intelligence, Australian Joint Conference on Artificial Intelligence, Sydney, Australia, 5–9 December 2005; pp. 1096–1105.
  12. Lee, C.; Landgrebe, D.A. Feature Extraction Based on Decision Boundaries. IEEE Trans. Pattern Anal. Mach. Intell. 1993, 15, 388–400. [Google Scholar] [CrossRef]
  13. Fukunaga, K. Introduction to Statistical Pattern Recognition; Academic Press: New York, NY, USA, 1974. [Google Scholar]
  14. Tadjudin, S.; Landgrebe, D. Classification of High Dimensional Data with Limited Training Samples. Available online: http://docs.lib.purdue.edu/ecetr/56/ (accessed on 30 October 2016).
  15. Ghamisi, P.; Dalla Mura, M.; Benediktsson, J.A. A Survey on Spectral-Spatial Classification Techniques Based on Attribute Profiles. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2335–2353. [Google Scholar] [CrossRef]
  16. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in Spectral-Spatial Classification of Hyperspectral Images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef]
  17. Serra, J. Image Analysis and Mathematical Morphology. Image Analysis and Mathematical Morphology; Academic Press: New York, NY, USA, 1982. [Google Scholar]
  18. Pesaresi, M.; Benediktsson, J.A. A new approach for the morphological segmentation of high-resolution satellite imagery. IEEE Trans. Geosci. Remote Sens. 2001, 39, 309–320. [Google Scholar] [CrossRef]
  19. Crespo, J.; Serra, J.; Schafer, R.W. Theoretical aspects of morphological filters by reconstruction. IEEE Trans. Image Process. 1995, 47, 201–225. [Google Scholar] [CrossRef]
  20. Serra, J.C.; Salembier, P. Connected operators and pyramids. Proc. SPIE 1993, 2030, 65–76. [Google Scholar]
  21. Soille, P. Morphological Image Analysis-Principles and Applications; Springer: New York, NY, USA, 2007; pp. 800–801. [Google Scholar]
  22. Dalla Mura, M.; Atli Benediktsson, J.; Waske, B.; Bruzzone, L. Extended profiles with morphological attribute filters for the analysis of hyperspectral data. Int. J. Remote Sens. 2010, 31, 5975–5991. [Google Scholar] [CrossRef]
  23. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  24. Dalla Mura, M.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of hyperspectral images by using extended morphological attribute profiles and independent component analysis. IEEE Geosci. Remote Sens. Lett. 2011, 8, 542–546. [Google Scholar] [CrossRef]
  25. Pedergnana, M.; Marpu, P.R.; Dalla Mura, M.; Benediktsson, J.A.; Bruzzone, L. A Novel Technique for Optimal Feature Selection in Attribute Profiles Based on Genetic Algorithms. IEEE Trans. Geosci. Remote Sens. 2013, 51, 3514–3528. [Google Scholar] [CrossRef]
  26. Song, B.; Li, J.; Dalla Mura, M.; Li, P.; Plaza, A.; Bioucas-Dias, J.M.; Chanussot, J. Remotely Sensed Image Classification Using Sparse Representations of Morphological Attribute Profiles. IEEE Trans. Geosci. Remote Sens. 2014, 52, 5122–5136. [Google Scholar] [CrossRef]
  27. Ghamisi, P.; Benediktsson, J.A.; Cavallaro, G.; Plaza, A. Automatic Framework for Spectral–Spatial Classification Based on Supervised Feature Extraction and Morphological Attribute Profiles. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2147–2160. [Google Scholar] [CrossRef]
  28. Fauvel, M.; Benediktsson, J.A.; Chanussot, J.; Sveinsson, J.R. Spectral and Spatial Classification of Hyperspectral Data Using SVMs and Morphological Profiles. IEEE Trans. Geosci. Remote Sens. 2007, 46, 3804–3814. [Google Scholar] [CrossRef]
  29. Li, W.; Chen, C.; Su, H.; Du, Q. Local Binary Patterns and Extreme Learning Machine for Hyperspectral Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3681–3693. [Google Scholar] [CrossRef]
  30. Chen, C.; Li, W.; Tramel, E.W.; Cui, M.; Prasad, S.; Fowler, J.E. Spectral–Spatial Preprocessing Using Multihypothesis Prediction for Noise-Robust Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1047–1059. [Google Scholar] [CrossRef]
  31. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  32. Otterman, J.; Fraser, R.S. Adjacency effects on imaging by surface reflection and atmospheric scattering: Cross radiance to zenith. Appl. Opt. 1979, 18, 2852–2860. [Google Scholar] [CrossRef] [PubMed]
  33. Tanré, D.; Deschamps, P.Y.; Duhaut, P.; Herman, M. Adjacency effect produced by the atmospheric scattering in thematic mapper data. J. Geophys. Res. 1987, 92, 12000–12006. [Google Scholar] [CrossRef]
  34. Shaw, G.A.; Burke, H.H.K. Spectral Imaging for Remote Sensing. Linc. Lab. J. 2003, 14, 3–28. [Google Scholar]
  35. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  36. Marpu, P.R.; Pedergnana, M.; Mura, M.D.; Peeters, S.; Benediktsson, J.A.; Bruzzone, L. Classification of hyperspectral data using extended attribute profiles based on supervised and unsupervised feature extraction techniques. Int. J. Image Data Fusion 2012, 3, 269–298. [Google Scholar] [CrossRef]
  37. Du, Q.; Yang, H. Similarity-based unsupervised band selection for hyperspectral image analysis. IEEE Geosci. Remote Sens. Lett. 2008, 5, 564–568. [Google Scholar] [CrossRef]
Figure 1. The implementation scheme of the proposed algorithm.
Figure 1. The implementation scheme of the proposed algorithm.
Sensors 16 02146 g001
Figure 2. False color representation and corresponding ground truth of (a,b) ROSIS-03 University of Pavia dataset, Sensors 16 02146 i001 Trees, Sensors 16 02146 i002 Asphalt, Sensors 16 02146 i003 Bitumen, Sensors 16 02146 i004 Gravel, Sensors 16 02146 i005 Metal sheets, Sensors 16 02146 i006 Shadows, Sensors 16 02146 i007 Meadows, Sensors 16 02146 i008 Bricks, Sensors 16 02146 i009 Bare soil; (c,d) AVIRIS Indian Pines dataset, Sensors 16 02146 i010 Alfalfa, Sensors 16 02146 i011 Corn-notil, Sensors 16 02146 i012 Corn-mintill, Sensors 16 02146 i013 Corn, Sensors 16 02146 i014 Grass-pasture, Sensors 16 02146 i015 Grass-trees, Sensors 16 02146 i016 Grass-pasture-mowed, Sensors 16 02146 i017 Hay-windrowed, Sensors 16 02146 i018 Oats, Sensors 16 02146 i019 Soybean-notill, Sensors 16 02146 i020 Soybean-mintill, Sensors 16 02146 i021 Soybean-clean, Sensors 16 02146 i022 Wheat, Sensors 16 02146 i023 Woods, Sensors 16 02146 i024 Bldg-grass-tree-drives, Sensors 16 02146 i025 Stone-steel-towers.
Figure 2. False color representation and corresponding ground truth of (a,b) ROSIS-03 University of Pavia dataset, Sensors 16 02146 i001 Trees, Sensors 16 02146 i002 Asphalt, Sensors 16 02146 i003 Bitumen, Sensors 16 02146 i004 Gravel, Sensors 16 02146 i005 Metal sheets, Sensors 16 02146 i006 Shadows, Sensors 16 02146 i007 Meadows, Sensors 16 02146 i008 Bricks, Sensors 16 02146 i009 Bare soil; (c,d) AVIRIS Indian Pines dataset, Sensors 16 02146 i010 Alfalfa, Sensors 16 02146 i011 Corn-notil, Sensors 16 02146 i012 Corn-mintill, Sensors 16 02146 i013 Corn, Sensors 16 02146 i014 Grass-pasture, Sensors 16 02146 i015 Grass-trees, Sensors 16 02146 i016 Grass-pasture-mowed, Sensors 16 02146 i017 Hay-windrowed, Sensors 16 02146 i018 Oats, Sensors 16 02146 i019 Soybean-notill, Sensors 16 02146 i020 Soybean-mintill, Sensors 16 02146 i021 Soybean-clean, Sensors 16 02146 i022 Wheat, Sensors 16 02146 i023 Woods, Sensors 16 02146 i024 Bldg-grass-tree-drives, Sensors 16 02146 i025 Stone-steel-towers.
Sensors 16 02146 g002
Figure 3. ROSIS-03 University of Pavia dataset. The classification maps of the proposed method obtain by the random forest classifier. (a) DA ; (b) DB . Sensors 16 02146 i001 Trees, Sensors 16 02146 i002 Asphalt, Sensors 16 02146 i003 Bitumen, Sensors 16 02146 i004 Gravel, Sensors 16 02146 i005 Metal sheets, Sensors 16 02146 i006 Shadows, Sensors 16 02146 i007 Meadows, Sensors 16 02146 i008 Bricks, Sensors 16 02146 i009 Bare soil.
Figure 3. ROSIS-03 University of Pavia dataset. The classification maps of the proposed method obtain by the random forest classifier. (a) DA ; (b) DB . Sensors 16 02146 i001 Trees, Sensors 16 02146 i002 Asphalt, Sensors 16 02146 i003 Bitumen, Sensors 16 02146 i004 Gravel, Sensors 16 02146 i005 Metal sheets, Sensors 16 02146 i006 Shadows, Sensors 16 02146 i007 Meadows, Sensors 16 02146 i008 Bricks, Sensors 16 02146 i009 Bare soil.
Sensors 16 02146 g003
Figure 4. AVIRIS Indian Pines dataset. The classification maps of the proposed method obtain by the random forest classifier, (a) DA ; (b) DB . Sensors 16 02146 i010 Alfalfa, Sensors 16 02146 i011 Corn-notil, Sensors 16 02146 i012 Corn-mintill, Sensors 16 02146 i013 Corn, Sensors 16 02146 i014 Grass-pasture, Sensors 16 02146 i015 Grass-trees, Sensors 16 02146 i016 Grass-pasture-mowed, Sensors 16 02146 i017 Hay-windrowed, Sensors 16 02146 i018 Oats, Sensors 16 02146 i019 Soybean-notill, Sensors 16 02146 i020 Soybean-mintill, Sensors 16 02146 i021 Soybean-clean, Sensors 16 02146 i022 Wheat, Sensors 16 02146 i023 Woods, Sensors 16 02146 i024 Bldg-grass-tree-drives, Sensors 16 02146 i025 Stone-steel-towers.
Figure 4. AVIRIS Indian Pines dataset. The classification maps of the proposed method obtain by the random forest classifier, (a) DA ; (b) DB . Sensors 16 02146 i010 Alfalfa, Sensors 16 02146 i011 Corn-notil, Sensors 16 02146 i012 Corn-mintill, Sensors 16 02146 i013 Corn, Sensors 16 02146 i014 Grass-pasture, Sensors 16 02146 i015 Grass-trees, Sensors 16 02146 i016 Grass-pasture-mowed, Sensors 16 02146 i017 Hay-windrowed, Sensors 16 02146 i018 Oats, Sensors 16 02146 i019 Soybean-notill, Sensors 16 02146 i020 Soybean-mintill, Sensors 16 02146 i021 Soybean-clean, Sensors 16 02146 i022 Wheat, Sensors 16 02146 i023 Woods, Sensors 16 02146 i024 Bldg-grass-tree-drives, Sensors 16 02146 i025 Stone-steel-towers.
Sensors 16 02146 g004
Figure 5. Overall accuracy with different number of NNs. (a) Pavia University dataset; (b) Indian Pines dataset.
Figure 5. Overall accuracy with different number of NNs. (a) Pavia University dataset; (b) Indian Pines dataset.
Sensors 16 02146 g005
Figure 6. ROSIS-03 Pavia University dataset. The classification maps of the proposed method with 30 training pixels per class. (a) DA ; (b) D B . Sensors 16 02146 i001 Trees, Sensors 16 02146 i002 Asphalt, Sensors 16 02146 i003 Bitumen, Sensors 16 02146 i004 Gravel, Sensors 16 02146 i005 Metal sheets, Sensors 16 02146 i006 Shadows, Sensors 16 02146 i007 Meadows, Sensors 16 02146 i008 Bricks, Sensors 16 02146 i009 Bare soil.
Figure 6. ROSIS-03 Pavia University dataset. The classification maps of the proposed method with 30 training pixels per class. (a) DA ; (b) D B . Sensors 16 02146 i001 Trees, Sensors 16 02146 i002 Asphalt, Sensors 16 02146 i003 Bitumen, Sensors 16 02146 i004 Gravel, Sensors 16 02146 i005 Metal sheets, Sensors 16 02146 i006 Shadows, Sensors 16 02146 i007 Meadows, Sensors 16 02146 i008 Bricks, Sensors 16 02146 i009 Bare soil.
Sensors 16 02146 g006
Table 1. University of Pavia dataset: number of training and test samples.
Table 1. University of Pavia dataset: number of training and test samples.
ClassNumber of Samples
LabelsNameTrainingTest
1Asphalt5486631
2Meadow54018,649
3Gravel3922099
4Trees5243064
5Metal Sheets2561345
6Bare Soil5325029
7Bitumen3751330
8Bricks5143682
9Shadows231947
Table 2. Indian Pines dataset: number of training and test samples.
Table 2. Indian Pines dataset: number of training and test samples.
ClassNumber of Samples
LabelsNameTrainingTest
1Alfalfa1539
2Corn-notil501384
3Corn-mintill50784
4Corn50184
5Grass-pasture50447
6Grass-trees50697
7Grass-pasture-mowed1511
8Hay-windrowed50439
9Oats155
10Soybean-notill50918
11Soybean-mintill502418
12Soybean-clean50564
13Wheat50162
14Woods501244
15Bldg-grass-tree-drives50330
16Stone-steel-towers5045
Table 3. University of Pavia dataset: OA, AA and κ value of the classification results. The best classification accuracies are marked in bold.
Table 3. University of Pavia dataset: OA, AA and κ value of the classification results. The best classification accuracies are marked in bold.
Class LabelsDADBDApDBp η DA η DB GASUnSALLBPMH DA DB
182.884.998.098.198.396.795.597.596.899.199.499.5
272.166.992.694.469.095.895.797.899.399.499.599.7
371.967.581.098.091.687.076.899.197.297.899.599.5
492.893.997.887.399.599.396.797.798.999.395.798.9
510099.999.899.610099.899.6100100100100100
689.193.898.610099.799.999.599.410099.298.599.2
783.885.510010099.799.910099.299.699.597.298.9
882.887.596.198.199.499.499.597.098.098.998.098.9
998.098.294.597.192.491.897.310010010096.8100
OA(%)79.978.994.896.085.596.896.198.198.899.298.999.4
AA(%)85.986.595.197.094.496.796.098.698.999.298.399.5
κ 0.750.740.930.950.820.960.950.970.970.980.980.99
Table 4. Indian Pines dataset: OA, AA and κ value of the classification results. The best classification accuracies are marked in bold.
Table 4. Indian Pines dataset: OA, AA and κ value of the classification results. The best classification accuracies are marked in bold.
Class LabelsDADBDApDBp η DA η DB GASUnSALLBPMH DA DB
153.948.797.497.494.897.497.4100100100100100
253.049.182.779.988.573.782.783.791.495.795.288.0
352.248.096.096.495.190.497.294.197.192.299.194.0
478.370.192.988.598.994.010092.910099.597.999.5
584.179.693.793.594.693.293.193.398.493.598.895.5
688.489.296.199.097.198.599.499.798.599.699.6100
710081.810010010010010010010010010092.3
898.298.699.799.398.699.099.810010097.7100100
940.020.010010010080.0100100100100100100
1060.562.391.687.286.677.889.192.390.195.691.791.9
1139.042.785.182.091.370.394.392.486.893.194.293.4
1266.068.187.784.989.777.690.898.583.794.198.998.2
1397.598.899.310099.310098.210010099.499.399.4
1484.885.199.399.699.493.199.610099.998.610099.8
1582.170.099.098.799.398.497.697.410099.410099.7
1697.893.310010010010010097.710097.810097.8
OA(%)63.563.191.189.693.383.193.894.093.395.796.895.0
AA(%)73.569.195.194.295.990.396.296.496.697.398.496.8
κ 59.058.50.900.880.920.810.930.9292.80.950.960.93
Table 5. The number of features used for classification in the proposed method and the corresponding AUTOMATIC approaches in the two test cases.
Table 5. The number of features used for classification in the proposed method and the corresponding AUTOMATIC approaches in the two test cases.
AlgorithmsFeatures
University of PaviaIndian Pines
DA613
DB2916
DAp306663
DBp1479816
η DA 1426
η DB 5959
D A 4580
D B 4580
Table 6. Processing time (in seconds) for Indian Pines dataset with 695 training samples and 9671 test samples.
Table 6. Processing time (in seconds) for Indian Pines dataset with 695 training samples and 9671 test samples.
AlgorithmsTime (s)
DA2
DB39
DAp13
DBp45
η DA 19
η DB 133
GA14
SUnSAL19
LBP15
MH253
DA 67
DB 512
Table 7. Accuracies in percentage for classification of the Pavia image with 30 training pixels per class. The best results in terms of accuracy are marked in bold.
Table 7. Accuracies in percentage for classification of the Pavia image with 30 training pixels per class. The best results in terms of accuracy are marked in bold.
Class LabelsDADBDApDBp η DA η DB GA SUnSALLBPMH DA DB
174.076.491.989.189.891.388.792.891.393.696.896.9
268.970.881.687.569.790.890.788.498.495.494.496.3
370.967.172.363.296.498.980.596.297.096.597.295.6
490.689.187.187.688.195.693.493.996.592.991.994.1
599.910010099.910099.999.399.7100100100100
676.382.187.090.587.894.692.688.896.395.396.398.4
787.888.686.198.695.599.694.999.798.798.397.995.8
873.882.171.987.070.692.092.594.594.595.296.196.6
999.999.799.799.499.899.996.598.410097.695.397.1
OA(%)74.977.284.087.980.092.991.090.996.695.295.496.6
AA(%)82.584.086.489.280.695.892.193.697.096.196.296.7
K0.68.0.710.790.840.750.910.880.880.960.930.940.96

Share and Cite

MDPI and ACS Style

Liu, D.; Li, J. Data Field Modeling and Spectral-Spatial Feature Fusion for Hyperspectral Data Classification. Sensors 2016, 16, 2146. https://doi.org/10.3390/s16122146

AMA Style

Liu D, Li J. Data Field Modeling and Spectral-Spatial Feature Fusion for Hyperspectral Data Classification. Sensors. 2016; 16(12):2146. https://doi.org/10.3390/s16122146

Chicago/Turabian Style

Liu, Da, and Jianxun Li. 2016. "Data Field Modeling and Spectral-Spatial Feature Fusion for Hyperspectral Data Classification" Sensors 16, no. 12: 2146. https://doi.org/10.3390/s16122146

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop