Land Cover Classification Using Hidden Markov Models

The schemes of incorporating 2D spatial information of remotely sensed imagery into a onedimensional linear HMM have been proposed and demonstrated in terms of visual quality through unsupervised classifications. In this paper the three proposed methods are applied to the Landsat imagery which is (30*30)meters of resolution which is the only available imagery for this work , if these methods are applied to higher resolution remotely sensed imagery the accuracy and the visual quality will raises to a better results.

pdf8 trang | Chia sẻ: huongthu9 | Lượt xem: 650 | Lượt tải: 0download
Bạn đang xem nội dung tài liệu Land Cover Classification Using Hidden Markov Models, để tải tài liệu về máy bạn click vào nút DOWNLOAD ở trên
N C S C International Journal of Computer Networks and Communications Security VOL. 1, NO. 4, SEPTEMBER 2013, 165–172 Available online at: www.ijcncs.org ISSN 2308-9830 Land Cover Classification Using Hidden Markov Models Dr. GHAYDA A. AL-TALIB1 and EKHLAS Z. AHMED2 1Assist. Prof. Computer Science Department, College of Computer Science and Mathematics, Mosul University, Mosul, Iraq 2M. Sc. Student, Computer Science Department, College of Computer Science and Mathematics, Mosul University, Mosul, Iraq E-mail: 1ghaydatalib@yahoo.com, 2ekh_aa2007@yahoo.com ABSTRACT This paper, proposed a classification approach that utilizes the high recognition ability of Hidden Markov Models (HMM s) to perform high accuracy of classification by exploiting the spatial inter pixels dependencies ( i.e. the context ) as well as the spectral information. Applying unsupervised classification to remote sensing images can provide benefits in converting the raw image data into useful information which achieves high classification accuracy. It is known that other clustering schemes as traditional k-means does not take into account the spatial inter-pixels dependencies. Experiments work has been conducted on a set of 10 multispectral satellite images. Proposed algorithm is verified to simulate images and applied to a selected satellite image processing in the MATLAB environment. Keywords: Hidden Markov Models (HMM), Land cover, Multispectral Satellite Images, Clustering, Unsupervised classification. 1 INTRODUCTION In this paper the Hidden Markov Models (HMM s) for unsupervised satellite image classification has been used. An HMMs were extensively and successfully used for texture modeling and segmentation (i.e., classification), this is majorly due to their ability to model contextual dependencies and noise absorption [1]. The land cover is an important geospatial variable for studying human and physical environments and is increasingly used as input data in spatially explicit ecological and environmental models, ranging from global climate change to detailed studies of soil erosion [2]. During the last decades, new sensors and easily accessible data archives have increased the amount of available remote sensing data. This development necessitates robust, transferable and automated methods and processing services for derivation of accurate information that can be produced within a short period of time [3]. Image classification refers to the computer-assisted interpretation of remotely sensed images. Mainly, there are two ways to do the remote sensing image classification. One is visual interpretation, and the other is computer automatic interpretation [4]. The classification is an important process, which translates the raw image data into meaningful and understandable information. The aim of image classification is to assign each pixel of the image to a class with regard to a feature space. These features can be consider the basic image properties as intensity, amplitude, or some more advanced abstract image descriptors as textures which can also be exploited as feature[5]. In this work the intensity property of the satellite images has been used to classify the land cover. The computer automatic classification of remotely sensed imagery has two general approaches, supervised and unsupervised classification. The supervised classification approach usually carries out the classification of land objects in the image by establishing study samples of typical land objects, such as building, water area, vegetation, roads, etc. Although supervised classification consumes shorter time [4]. While unsupervised classification approach automatically cluster the image into a number of groups according to specific predefined criterion [6]. Cases of unsupervised classification, some of the statistical properties of the different classes are unknown and have to be estimated with iterative methods such as estimation-maximization 166 Dr. G. A. Al-Talib et al. / International Journal of Computer Networks and Communications Security, 1 (4), September 2013 (EM)[7]. In the first step of this work the model parameters are set randomly and then will estimated the optimal values with Baum-Welch algorithm which is the most widely adopted methodology for model parameters estimation [8]. After the model parameters are well estimated the clustering and classification process starts with Viterbi algorithm. Finally the displaying the classified image. 2 HIDDEN MARKOV MODELS An HMM is distinguished from a general Markov model in that the states in an HMM cannot be observed directly (i.e. hidden) and can only be estimated through a sequence of observations generated along a time series. Assume the total number of states being N, and let qt and ot denote the system state and the observation at time t. HMM can be formally characterized by λ= (A, B, π), where A is a matrix of probability transition between states, B is a matrix of observation probability densities relating to states, and π is a matrix of initial state probabilities, respectively. Specifically, A, B, and π are each further represented as follows [6]: A=[aij], aij = P(qt+1=j | qt =i ), 1≤ i,j ≤ N (1) Fig.1. HMM parameters A, B and π in the case of t=1. Where aij≥0,∑ = 1,for i=1,2,,N (2) B=[bj(ot)], bj(ot)=P(ot | qt=j), 1≤ ,j ≤ N (3) π= [πi], πi=P(q1=i), 1≤ ,i ≤ N (4) where ∑Ni=1 πi =1 (5) For illustration purposes, an HMM model and related parameters A, B, and π are shown in Fig. 1. Given HMM, λ and observation sequence O={o1, o2,, oT}, one may estimate the best state sequence Q*={q1, q2,,qT} based on a dynamic programming approach so as to maximize P(Q*|O, l), [ 5]. In order to make Q* meaningful, one has to well set up the model parameters A, B and π. The Baum-Welch algorithm is the most widely adopted methodology for model parameters estimation. The model parameters pi, aij, are each characterized as: π = γ1 (i) (6) )( ),( 1 1 1 1 i ji a T t t T t ij         i=1,2,,N (7)         1 1 1 1 )( )( )( T t t T O t t j j j kb kVt   i=1,2,.,N (8) Where γt(i) denotes the conditional probability of being state i at time t, given the observations, and ξt(i, j) is the conditional probability of a transition from state i at time t to state j at time t + 1, given the observations. Both γt(i) and ξt(i, j) can be solved in terms of a well-known forward-backward algorithm [9]. Define the forward probability αt(i) as the joint probability of observing the first t observation sequence O1 to t={o1, o2,,ot} and being in state i at time t. The αt(i) can be solved inductively by the following formula: α1(i) = πi bi(o1) , 1 ≤ i ≤ N (9) αt+1(i) =b(ot+1) ∑ N j=1[αt(i) αij], For 1≤ t ≤ T, For 1≤ i ≤ N (10) Let the backward probability bt(i) be the conditional probability of observing the observation sequence Ot to T={ot+1, ot+2,,oT} after time t given that the state at time t is i. As with the forward probability, the bt(i) can be solved inductively as: T = 1 , 1 ≤ I ≤ N (11) 167 Dr. G. A. Al-Talib et al. / International Journal of Computer Networks and Communications Security, 1 (4), September 2013 βt(i)=∑ ij bj(ot+1)βt+1(j), t=T-1 , T-2,...,1, 1≤ i ≤ N (12) The probabilities γt(i) and ξt(i, j) are then solved by: γt (i) = ()() ∑ ()() (13) ξt(i,j)= () ()() ∑ ∑ () ()() (14) As a result, if both the observation density bi(ot) and observation sequence O={o1,o2,,oT} are well managed, then the hidden state sequence will be closer to the ideal situation. Moreover, the Viterbi algorithm is usually employed to perform global decoding which found the states of each observation separately. Using the Viterbi algorithm is aimed to find the most likely sequence of latent states corresponding to the observed sequence of data[10]. Also the Viterbi algorithm can be used to find the single best state sequence of an observation sequence. The Viterbi algorithm is another trellis algorithm which is very similar to the forward algorithm, except that the transition probabilities are maximized at each step, instead of summed [11].First define: δt(i) = max P(q1 q2···qt = si, o1, o2··· ot | λ) (15) q1,q2,···,qt-1 As the probability of the most probable state path of the partial observation sequence. The Viterbi algorithm steps can be stated as: 1. Initialization δ1(i) = πi bi(o1), 1 ≤ i ≤ N, (16) 2. Recursion: N δt(j)= max [δt-1(i) aij ] bj(ot), 2 ≤ t ≤ T,1 ≤ j ≤ N (17) i=1 N Ψt(j) = arg max [δt-1(i) aij ] , 2 ≤ t ≤ T, 1 ≤ j ≤ N (18) i=1 3. Termination: N P*= max [δT(i)] (19) i=1 N q*T = arg max [δT(i)] (20) i=1 When implementing HMM for unsupervised image classification, the pixel values (or vectors) correspond to the observations, and after the estimation for the model parameter is completed, the hidden state then corresponds to the cluster to which the pixel belongs. So the first step is set the model parameters and the second step is applying the Viterbi algorithm to segmenting the pixels of the image to several clusters and the last step is for mapping these clusters to a number of classes which is determined by the user. 3 METHODOLOGY Usually, in a traditional k-means, the clustering scheme uses a spectral data alone in classifying satellite images, while in this work the spectral and spatial information have been used to classify satellite imagery and the method used is an observation density based method for sequencing the pixels of the satellite images which take blocks of pixels (2*2) , (3*3), and cross neighbours so that this method can be extend the scope of the observation in terms of combining the pixel with its neighbouring pixels to form a new observation vector. So each block of size (2*2), (3*3) and cross neighbours will represent an observation. Following such a consideration, to build up an HMM, the pixel fitting direction aligned to the row direction (i.e. strip-like) has been maintain, but for each pixel of interest, the vertical and horizontal neighbouring pixel(s) are added in order to take the vertical and horizontal spatial dependencies into account. In this method (4 pixels ), (9 pixels) and (5 pixels) will contribute to produce a single observation and as a result made the classification more accurate by maximizing the probabilities of the observations (pixels) to which class is belong. For illustration we use a sample image of ( 5*5 ) pixels to show the method as in Fig. 2. 168 Dr. G. A. Al-Talib et al. / International Journal of Computer Networks and Communications Security, 1 (4), September 2013 (b) (c) (d) Fig. 2. (a) is a (5*5) sample image, (b) an observations represented with a block of size (2*2) neighbourhood pixels, (c)an observations represented with a block of size (3*3) neighbourhood pixels, and (d) a cross neighbourhood pixels observations. 4 THE TEST AREA Two test areas have been considered, one is located in the north east of Kirkuk and the other is located in the south Dohuk in Iraq. Both areas contains the same classes of land cover which are (water, vegetation, bare soil, and rocky areas). The Landsat satellite imagery captured at 2012, the resolution of the image is (30×30) meter which means that each pixel represents a(30×30) meter on the earth. Satellite imagery can provide a convenient means of monitoring the evolution of the area. For experimental purposes, a test area with (300 × 300) pixels was extracted from the Landsat multispectral imagery to facilitate the analysis of the classification methodology. Each classified image was then evaluated in terms of its visual quality. The test area is shown in Fig. 3. Fig. 3. (a) Landsat multispectral imagery in terms of false color composite data, in the north of Iraq,(b) Test area (1), (c) Test area (2). 169 Dr. G. A. Al-Talib et al. / International Journal of Computer Networks and Communications Security, 1 (4), September 2013 5 THE PROPOSED ALGORITHM This paper will use unsupervised Hidden Markov Models (HMM) to classifying the land cover from multispectral satellite images and propose a new technique to sequencing the pixels (observations) of the image so as to fit them into the HMM and to get the best classification. Fig. 4 shows the flowchart of this work. At the first step the test areas have been determined and selected best bands. Here, band combinations (band2, band4, band7) for the two test areas have been chosen. After that the image entered to the program. The number of classes in the image was defined by the user, then change the image from the RGB form into gray scale image. In the next step, coding of each pixel will be done depending on the intensity of the pixel. Convert the image from two dimension representation into one dimension (a sequence of pixels or a vector of observations) because the (HMM) original theory assumes a one-dimensional linear dependent structure. This research uses a block of (2*2) neighbours and (3*3) neighbours as well as the cross neighbours method to utilize the dependencies between pixels. In this way, the 2D information can be embedded into HMM, while the original one-dimensional linear structure of an HMM is still remained. Fig. 4. The flowchart showing the steps of the proposed algorithm. After the estimation of the model parameters λ(A , B , π ) and by specifying a threshold to each class, then Viterbi algorithm for unsupervised image classification was done to classify each pixel of the image. The pixel values (or vectors) corresponding to the observations and the hidden state corresponding to the cluster to which the pixel belongs, if the resulted classified image is the best classification then save it and display the last result or return to updating the model parameters and maximizing the probability of the observation sequence and repeat the procedure until the best classification resulted then save and display the resulted image. 6 RESULTS AND DISCUSSION The schemes of using HMMs to pursue higher classification accuracy in unsupervised sense as described earlier have been implemented for the test imagery. HMM was constructed by each method according to the parameter estimation process and then the state for each corresponding pixel was extracted to form classified imagery. For all HMMs, the original model parameters, namely, A, B, and π, are randomly assigned [12]. The iterations of the model parameters estimations are converged within ten or eleven runs. The hidden state (i.e. cluster) sequence Q* is then estimated using Viterbi algorithm (i.e. dynamic programming) so as to maximize the probability of sequence of states given the observation and the model P (Q*|O, λ). The resulting clusters are then mapped to the information classes according to the knowledge of ground data. The resulting images for unsupervised classifications using HMM based on the (2*2), (3*3) and cross neighbourhood observation sequence of the test area (1) which is shown in Fig. 5. (a) to ( c ) and for the test area (2) which is shown in Fig. 6. (a) to ( c ). In both areas, the three results show that (3*3) neighbourhood method gives the lowest accuracy classification and the cross neighbourhood gives the highest accuracy. In the three results of the first test area there are three classes ( water, vegetation, bare soil) the water body is best classified so as the vegetation areas, there was some confusion between bare soil and water in the three results, and the less confusion was in the (2*2) neighbourhood method, while in the three results of the second test area there are six classes ( water, vegetation, bare soil, Rocky lands, High soil moisture, and shades). 170 Dr. G. A. Al-Talib et al. / International Journal of Computer Networks and Communications Security, 1 (4), September 2013 (a) (b) (c) Fig. 5. The result of classification of the test area (1). (a)the result of (3×3)neighbourhood, (b) the result of (2×2)neighbourhood, and ( c ) the result of cross neighbourhood. (a) (b) (c) Fig. 6. The result of classification of the test area (2). (a)the result of (3*3)neighbourhood, (b) the result of (2*2)neighbourhood, and ( c ) the result of cross neighbourhood. 171 Dr. G. A. Al-Talib et al. / International Journal of Computer Networks and Communications Security, 1 (4), September 2013 The Figure 6 (a) and (b) Shows the classes distribution for the two test areas. (a) (b) Fig. 6. Shows the class distribution, (a) class distribution for test area(1), (b) class distribution for test area(2) 172 Dr. G. A. Al-Talib et al. / International Journal of Computer Networks and Communications Security, 1 (4), September 2013 6 CONCLUSION The schemes of incorporating 2D spatial information of remotely sensed imagery into a one- dimensional linear HMM have been proposed and demonstrated in terms of visual quality through unsupervised classifications. In this paper the three proposed methods are applied to the Landsat imagery which is (30*30)meters of resolution which is the only available imagery for this work , if these methods are applied to higher resolution remotely sensed imagery the accuracy and the visual quality will raises to a better results. 7 REFERENCES [1] Mohamed El Yazid Boudaren, and Abdel Belaid, “A New scheme for land cover classification in aerial images: combining extended dependency tree-HMM and unsupervised segmentation,” Lecture Notes in Electronic Engineering - Electronic Engineering and Computing Technology Spriger, inroad-00579704, version 1-24, pp. 471-482, Mar 2011. [2] Weidong Li and Chuanrong Zhang, “A markov chain geostatistical framework for land-cover classification with uncertainty assessment based on expert-interpreted pixel form remotely sensed imagery,” IEEE transactions on geoscience and remote sensing, Vol. 49, No. 8, 2983-2992, August 2011. [3] Juliane Huth, Claudia Kuenzer, Thilo Wehrmann, steffen Gebhardt, Vo Quoc Tuan, and Stefan Dech, “Land cover and land use classification with TWOPAC: towards automated processing for pixel- and object- based image classification,” Remote Sens. 4, doi:10.3390/rs4092530, pp. 2530-2553, 2012. [4] Qian Wang, Jianping Chen, and Yi Tian, “Remote sensing image interpretation study serving urban planning based on GIS,” The international Archives of the Photogrammetry, Remote sensing and Spatial Information Sciences. Vol. XXXVII. Part B4. Beijing 2008. [5] Koray Kayabol, and Josiane Zerubia, “Unsupervised mplitude and texture based classification of SAR images with multinomial latent model,” Institute National de Recharge en Information en Automatique. Hal- 00612491, version 2-2 May 2012. [6] Tso B. , and Olsen R. C., “Combining spectral and information into hidden markov models for unsupervised image classification,” International Journal of Remote Sensing. Vol. 26, No. 10, pp. 2113-2133, 20 May 2005. [7] Roger FjØrtoft, Jean-Marc Boucher, Yves Delignon, Rene Garello, Jean-Marc Le Caillec, Henri Maître, Jean-Marie Nicolas, Wojciech Pieczynski, Marc Sigelle, and Florence Tupin,”Unsupervised classification of radar images based on hidden markov models and generalised mixture estimation,” Proc. SAR Image Analysis Modeling, and Techniques V, Vol.SPIE 4173, Barcelona, Spain, September 2000. [8] Wleed Abdulla, and Nikola Kasabov, “The concept of hidden markov model in speech recognition,” The Information Science Discussion Paper Series, No. 99/09, ISSN 1177-45X, , May 1999. [9] Leonard E. Baum, Ted Petrie, George Soules, and Norman Weiss, “A Maximization technique occurring in the statistical analysis of probabilistic functions of markove chains,” The Annals of Mathematical statistics, Vol. 41, No. 1, pp. 164-171, 1970. [10] Silvia Pandolfi, and Francesco Bartolucci, “A new constant memory recursion for hidden markov models,” FIRB (“Future in ricerca” 2012), Perugia (IT), March 15-16, 2013. [11] Daniel Jurafsky, and Jamse H. Martin, “Speech and language processing: An introduction to natural language processing, computational linguistics and speech recognition,” 2nd Ed., prentice-Hall 2000, ISBN: 0-13-095069-6. [12] Lawrence R. Rabiner, “A Tutorial on hidden markov models and selected applications in speech recognition,” Proceedings of the IEEE, Vol. 77, No. 2,February 1989.

Các file đính kèm theo tài liệu này:

  • pdfland_cover_classification_using_hidden_markov_models.pdf