Tidak ada review pada koleksi ini: 8586
This dissertation describes the synergy use of remote sensing data from multi-temporal and multisensor (optical and radar) for improving our understanding of the land-cover structural phenomena. A tropical country like Indonesia has a high cloud coverage throughout the year with a maximum during the rainy season, and hence the availability of cloud-free optical images is minimal. To solve this problem, radar images have been intensively introduced. The radar images are cloud-free but their use is hampered due to their speckle noise and topographic distortions, and the lack of a suitable radar image classification system. In many cases, the use of optical or radar image alone is not sufficient. Therefore, the main objectives of this research are: (i) to developa framework for mutidate and multisensor (optical and radar) image classification; (ii) to solve the cloud cover problem in optical images; and (iii) to obtain a more consistent image classification using multidate and multisensor images. we have proposed a framework for multidate and multisensor image classification based on a uniform image classification scheme. The term uniform means that the same procedure can be used to classify optical or radar images, low-level mosaic or fused images, single or multiple feature images. To be able conduct a multitemporal and multisensor analysis, we have unified the optical and radar image classification procedure after finding that both optical and radar images consist of homogeneous and textured regions. A region is considered as homogeneous if the local variance of gray level distribution is relatively low, and a region is considered as textured if the local variance is high. We used a multivariate Gaussian distribution to model the homogeneous part and a multinomial distribution to model the gray level co-occurrences of the textured part, and applied a multiple classifier system to improve the classification accuracy. The main advantages of the uniform classification scheme are as follow. First, we can tune the homogeneous-textured threshold value parameter in order to obtain an optimal result by allowing the classfier working as a single (conventional) or multiple classifier system. The classifier can have a better or at least the same classfication accuracy as the conventional one. Second, we can use either single-band or multi-band input images. This will make it possible to classify a radar image based on multimodel texture feature images or to classify multispectral optical images. Third, we can use the same procedure to classify any input images. Compared to the conventional classifiers, the multiple classifier system can improve the classification result from 0% to 20% for radar images and from 0% to 2% for optical images. The proposed framework incorporates the image mosaicking and data fusion at the low-level stage (before the classification process) as well as at the high - level stage (after the classficiation process). For cloud cover remaoval, the image mosaicking at the low-level stage is usually done using multitemporal optical images, whereas mosaicking at the high-level stage is applied to the the classified optical and radar images. To be able to obtain a cloud-free image, we have modified the existing Soofi and Smith algorithm which is using multitemporal optical images to an algorithm using multisensor images. In the high-level data fusion, we have also been able to incorporate a mechanism for cloud cover removal by omitting the information from the optical sensor and using only the information from the radar sensor. According to a case study in our experiment, the cloud cover removal and image classification using the low-level image mosaicking, the high-level image masaicking, and the high-level data fusion gave 80.2%, 76.2% and 80.5% classification accuracy, respectively. The high-leve data fusion combines the decisions from several input images to obtain a consensus of classfied image. We have applied both the maximum joint posterior probabilityand the highest rank method for the decision combination functions. We have utilized two existing data fusion methods and have proposed an alternative data fusion method based on the compound conditional risk. According to the experimental results, the dicision combination function based on the maximum joint posterior probability favors the optical feature image, while the highest rank method favors the radar feature image. The preference of using the maximum joint posterior probability results in the domination of optical features in the fusion result, and the classification accuracy of the fused image can be better 8.5% in average than the individual radar classified image.