Movatterモバイル変換


[0]ホーム

URL:


Next Article in Journal
Coastal Wetland Mapping with Sentinel-2 MSI Imagery Based on Gravitational Optimized Multilayer Perceptron and Morphological Attribute Profiles
Next Article in Special Issue
Spectral Super-Resolution with Optimized Bands
Previous Article in Journal
Terrestrial Structure from Motion Photogrammetry for Deriving Forest Inventory Data
Previous Article in Special Issue
Pansharpening Using Guided Filtering to Improve the Spatial Clarity of VHR Satellite Imagery
 
 
Search for Articles:
Title / Keyword
Author / Affiliation / Email
Journal
Article Type
 
 
Section
Special Issue
Volume
Issue
Number
Page
 
Logical OperatorOperator
Search Text
Search Type
 
add_circle_outline
remove_circle_outline
 
 
Journals
Remote Sensing
Volume 11
Issue 8
10.3390/rs11080951
Font Type:
ArialGeorgiaVerdana
Font Size:
AaAaAa
Line Spacing:
Column Width:
Background:
Article

A Local Feature Descriptor Based on Oriented Structure Maps with Guided Filtering for Multispectral Remote Sensing Image Matching

1
National Key Laboratory of Science and Technology on Multispectral Information Processing, Huazhong University of Science and Technology, Wuhan 430074, China
2
School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China
3
Science and Technology on Complex System Control and Intelligent Agent Cooperation Laboratory, China Aerospace Science and Industry Corporation third Academy, Beijing 100074, China
*
Author to whom correspondence should be addressed.
Remote Sens.2019,11(8), 951;https://doi.org/10.3390/rs11080951
Submission received: 22 March 2019 /Revised: 15 April 2019 /Accepted: 18 April 2019 /Published: 20 April 2019
(This article belongs to the Special IssueMultispectral Image Acquisition, Processing and Analysis)

Abstract

:
Multispectral image matching plays a very important role in remote sensing image processing and can be applied for registering the complementary information captured by different sensors. Due to the nonlinear intensity difference in multispectral images, many classic descriptors designed for images of the same spectrum are unable to work well. To cope with this problem, this paper proposes a new local feature descriptor termed histogram of oriented structure maps (HOSM) for multispectral image matching tasks. This proposed method consists of three steps. First, we propose a new method based on local contrast to construct the structure guidance images from the multispectral images by transferring the significant contours from source images to results, respectively. Second, we calculate oriented structure maps with guided image filtering. In details, we first construct edge maps by the progressive Sobel filters to extract the common structure characteristics from the multispectral images, and then we compute the oriented structure maps by performing the guided filtering on the edge maps with the structure guidance images constructed in the first step. Finally, we build the HOSM descriptor by calculating the histogram of oriented structure maps in a local region of each interest point and normalize the feature vector. The proposed HOSM descriptor was evaluated on three commonly used datasets and was compared with several state-of-the-art methods. The experimental results demonstrate that the HOSM descriptor can be robust to the nonlinear intensity difference in multispectral images and outperforms other methods.

    1. Introduction

    Feature matching is a process that can obtain the correspondence of interest points among two or more images of the same scene with varying degrees of overlap [1]. The correspondence relationships can be used to geometrically align these images [2]. Multispectral images often provide complementary information about the same scene by capturing the different characteristics of different spectral bands. For example, an infrared image captures thermal radiation of the objects in a scene, while a visible image mainly records optical reflection information in a scene. Therefore, feature matching between the multispectral images is at the base of many computer vision and remote sensing applications, such as target recognition [3,4], image registration [5,6], modern military surveillance [7,8], 3D reconstruction [9,10], image fusion [11,12], and medical image processing [13,14]. Due to differences in sensors and spectra between multispectral images, there are significant nonlinear intensity variations, which mean that the grayscales and intensity distributions between multispectral images are nonlinearly correlated. This is because the scene has different reflection characteristics for the electromagnetic spectrum of different bands [15,16,17]. Therefore, obtaining accurate and robust matching between multispectral images is still a very challenging task. Although there are many different feature matching methods, they all have three steps in common: interest point detection, interest point description, and interest point matching [18,19,20]. This study focuses on the interest point description, which is the difficulty and core problem in the matching task.
    In recent years, many excellent feature descriptors have been proposed for different matching applications. Among these methods, the scale invariant feature transform (SIFT) [21] and speeded-up robust feature (SURF) [22,23] are the most widely used algorithms. Due to the robustness of these two descriptors for scale invariance and rotation invariance, they can achieve good performance in many image matching tasks. However, when these two methods are applied to multispectral image matching works, they often fail to obtain correct correspondences. This is because they are robust to illumination and viewpoint changes but sensitive to nonlinear intensity changes [24,25,26].
    Scholars have designed many feature descriptors to deal with the problem of significant nonlinear intensity differences in multispectral images, in which many methods are improvements to the classic descriptors. Saleem and Sablatnig [24] proposed a modification to SIFT termed normalized gradient SIFT (NG-SIFT), which utilizes normalized gradients to compute the feature vector as the description of interest points to achieve robustness against nonlinear intensity changes in multispectral images. A combined method based on SURF, partial intensity invariant feature descriptor (PIIFD), and robust point matching (RPM), called SURF-PIIFD-RPM [25] are presented to improve the matching performance of multimodal retinal images. This method first employs the SURF detector to extract scale invariant and stable feature points. And then it constructs the local feature descriptors based on PIIFD that can keep robust to intensity changes. Finally, the PRM method is used for feature matching and outliers removing.
    Many methods are specially designed to work with multispectral images and can achieve better robustness and efficiency than these progressive SIFT-like or SURF-like algorithms. The paper [26] proposed the edge histogram descriptor (EHD) that has been adopted in the MPEG-7 standard. The EHD employs the spatial distribution of edge points to represent the robust image features, which can keep reliable texture information even when there are significant intensity variations between multispectral images. The paper [27] presented a new feature descriptor called the edge-oriented histogram (EOH), which was used for the matching task between the visible images and the long wave infrared (LWIR) images. The EOH descriptor uses the edge points distribution of four directional edges and one non-directional edge to construct the feature description, which can keep reliable structure information even when there are significant intensity variations between multispectral images. Although the EHD and EOH can describe image contour information, they have difficulty extracting highly similar edges from multispectral images. To solve this deficiency, the paper [28] proposed local descriptor call histograms of directional maps (HoDM), which combined the structure and texture features to descript a keypoint. Different from the EOH descriptor, the Log-Gabor histogram descriptor (LGHD) [29] and multispectral feature descriptor (MFD) [30] use multi-scale and multi-oriented Log-Gabor filters to replace the multi-oriented spatial filters. The LGHD can get richer and more robust feature representation in multispectral images but suffers from high dimensionality and low efficiency. The MFD can significantly reduce the feature dimension while maintaining feature description capabilities.
    Figure 1 shows the comparison between the variations of gradient direction in multispectral images and corresponding structure guidance images. It can be seen that the structure guidance images of visible and infrared (IR) image can retain the common structures although the gradient directions are reversed. Many traditional methods usually use the gradient-oriented information to capture the edges and to build the local feature vectors in multispectral images, which are good at characterizing local shape, but are sensitive to changes of the gradient directions, as shown inFigure 1a1,a2 andFigure 1b1,b2. In order to overcome this deficiency, we propose a new feature descriptor based on oriented Sobel edge maps with guided filtering.
    The main contribution in this paper consists of three aspects.
    (1) We design a new local feature descriptor called histogram of oriented structure maps (HOSM) for multispectral image matching.
    (2) We propose a new guidance image extractor, as shown inFigure 1a3,b3, which can extract highly similar structure information from multispectral images.
    (3) We provide evaluations on our method in three multi-source datasets and compare it with several state-of-the-art methods, the experiments show that HOSM is more robust to nonlinear intensity variations by keeping more structure features with guided filtering.
    The remainder paper is organized as follows.Section 2 introduces the proposed HOSM descriptor. InSection 3, the experimental results and corresponding analyses are exhibited. Finally, we conclude this paper inSection 4.

    2. Materials and Methods

    In this part, we present the proposed HOSM descriptor. This method contains three parts. First, a new local contrast-based operator is proposed to construct the structure guidance image to preserve the significant contour characteristics in source images. Second, the oriented structure maps are computed with guided image filtering to capture the common structure properties in multispectral images. Finally, we calculate the histogram of oriented structure maps to build the feature vector.

    2.1. Construct Guidance Image

    The significant contour information of an image is the key feature to characterize structures that can usually be expressed by the image gradients. Because of the nonlinear intensity variations in multispectral images, the gradient directions are inconsistent, which leads to the failure to obtain similar structures. To overcome this problem, we propose an isotropic extractor to construct the structure guidance image that can capture the significant contours between the multispectral images. The processing flow is shown inFigure 2 and extractor is expressed as follows.
    G(x,y)=i=kkj=kk|I(x+i,y+j)Imean|max(I(x+i,y+j),Imean)
    whereI is source image,G is the structure guidance image,(x,y) represents the pixel location of filter region center,k is the radius of filter window, andImean is the mean of intensities of the local region. Then, to get the robust guidance image, we normalize the magnitude of imageG to the range of 0–255 as follows.
    G(x,y)=255×G(x,y)minGmaxGminG

    2.2. Compute Oriented Structure Maps

    The Sobel operator has good noise-suppression ability and is simple to carry out but can only extract the horizontal and vertical edges. The progressive Sobel filters [26] shown inFigure 3 are able to detect multi-oriented edges in images. They can extract the spatial distribution of one non-directional edge and four oriented edges, including 0°, 45°, 90°, and 135°. This filter bank follows the MPEG-7 standard [31] and is able to represent the main structure information of the image [27]. Nunes and Padua [30] have pointed out that more orientations have little improvement in description capabilities but can significantly increase time consumption. In order to get a good trade-off between performance and efficiency, we use the five-oriented Sobel filters to construct structure maps. As shown inFigure 4a, letfn(x,y),n=1,2,3,4,5 denotes the five Sobel filters of 0°, 45°, 90°, 135°, and no orientation. The multi-oriented Sobel edges can be extracted by follows.
    SEn(x,y)=|I(x,y)*fn(x,y)|
    whereI(x,y) is pixel of source image, andSEn(x,y) represents the Sobel edge of orientationn at the pixel(x,y). “” is the convolution operator, and “||” represents the absolute operation.SEn is normalized to the range of 0–255. The largest value at each pixel in different edge images is used to build the edge maps as follows.
    EMn(x,y)={SEn(x,y),if SEn(x,y)=max{SEk(x,y)}150,others
    whereEMn(x,y) shown inFigure 4b is edge maps. The edge maps maintain the common structure characteristics of multispectral images but suffer from the deficiencies that are sensitive to noise and can cause aliasing effects. To address these problems, the structure maps are computed by guided image filtering [32,33,34] on edge maps as follows.
    SMn=GF(EMn,G,r,ε)
    whereG is the guidance image,SMn represents structure maps as shown inFigure 4c,r is the local window radius, andε is the regularization parameter. In the guided filtering operation, the output image is computed by a linear transformation of the guidance image, which can be represented as follows.
    SMn(i)=amG(i)+bm,iωm
    whereωm is a local window,am andbm are linear coefficients, andi represents the pixel location in the local window. To get the linear coefficients, we have to minimize the following cost function.
    E(am,bm)=iωm((amG(i)+bmEMn(i))2+εam2)
    am andbm can be represented as.
    am=1|ω|iωmG(i)EMn(i)μmEMn¯σm2+ε
    bm=EMn¯amμm
    whereμm andσm2 represent the mean and variance of the guidance image,EMn¯ is the mean of input image, and|ω| is the number of the pixels in the local windowωm. Pixeli can be contained in more than one local windows, in different local windows, the output values are changes. To solve this problem, the mean of all possible coefficientsam andbm are used as the final coefficients, as follows.
    SMn(i)=ai¯G(i)+bi¯
    whereai¯ andbi¯ can be computed as follows.
    ai¯=1|ω|mωiam
    bi¯=1|ω|mωibm

    2.3. Proposed HOSM Descriptor Based on Structure Maps

    The proposed HOSM descriptor employs the multi-oriented histogram to build the feature vector based on the oriented structure maps by guided image filtering.Figure 5 illustrates the main flowchart of the proposed HOSM descriptor, where the green point in the region center denotes an interest point. The main processing flow consists of three steps that can be summarized as follows.
    • First, a new structure extractor is applied to construct structure guidance images in the local region around the interest point, which can capture the significant contour information and keep robust to nonlinear intensity changes in multispectral images.
    • Second, we construct the oriented structure maps by performing guided filtering on multi-oriented Sobel edge maps. The detailed construction process of structure maps is shown inFigure 4. (1) The progressive Sobel filters are applied to construct the multi-oriented edge maps, which can extract the significant structure characteristics of images. (2) We perform guided filtering on edge maps to get the final structure maps. This guided filtering operation can enhance the structures in final structure maps while maintaining the advantages of multiple directions in the edge maps. Therefore, the oriented structure maps are able to remain the common features between multispectral images and reduce aliasing effects with the guided filtering.
    • Finally, we calculate the histogram of oriented structure maps in a local region to build the HOSM descriptor. We first divide the local region around the interest point into 16 sub regions by theN ×N window. We then compute and normalize the 5-orientation histogram by L2 norm in each sub region. At last, all the histograms of the 16 sub regions are conjoined to form a feature vector.

    3. Experiments and Analyses

    3.1. Datasets and Settings

    In order to evaluate the description ability of our proposed HOSM descriptor, we carried out the experiments on three widely used datasets as shown inFigure 6 and compared it with seven state-of-the-art approaches that are SURF, SIFT, NG-SIFT, EOH, LGHD, MFD, and HoDM.Figure 6a is a Potsdam dataset that contains 38 visible (VIS) and near infrared (NIR) aerial image pairs and all image pairs have dimensions of 6000 × 6000. This dataset is generated from a remote sensing image dataset and can be accessed at [35]. The EPFL dataset was proposed in [36] as shown inFigure 6b. It consists of 477 visible and NIR images and all image pairs have dimensions of 1024 × 768. The CVC dataset was proposed in [27] as shown inFigure 6c, which contains 100 visible and long wave infrared (LWIR) image pairs and has dimensions 506 × 408 for all images.
    All the three datasets satisfy homography, so the ground truth of each image pair can be computed precisely by manually selected checkpoints. The match was regarded as correct correspondence if the residual error computed by the ground truth was less than 3 pixels. The local feature points are extracted by the FAST extractor [37]. The number of correct feature point matches has a great influence on the matching performance of the descriptors.
    Euclidean distance is used as a measure of similarity for feature matching. To preserve as many correct matches as possible, the nearest neighbor distance ratio (NNDR) is used as the matching strategy in the experiments. The NNDR can be expressed as follows.
    D(a0,b0)<ηD(a0,b1)
    where(·,·) indicates the Euclidean distance,a andb are the feature vectors in two images, andb1 is the second-closest neighbor toa0 compared with the nearestb0.η is the threshold of the NNDR.
    In the experiment implements, the same region size (80 × 80) with EOH was adopted in our descriptor. The guided image filter parameters were set asr=7, andε=0.3. The thresholdη was set from 0.8 to 1.0 with intervals of 0.05. The parameters of all competing methods were set as the original papers. For all testing images, the same parameters in the proposed method and other competing methods were adopted.
    Three metrics, i.e., precision, recall, and F1-score, are used to validate the matching performance, they are defined as follows.
    precision=NcNc+Nf
    recall=NcNc+Nc¯
    F1score=2×precision×recallprecision+recall
    whereNc indicates the number of correct matches,Nf is the number of false matches, andNc¯ indicates the number of discard correct matches.

    3.2. Qualitative Evaluation on Feature Matching

    In the qualitative evaluation experiments, we conduct the feature matching tests on the three datasets by our proposed HOSM descriptor.Figure 7 andFigure 8 show some samples of the multispectral image feature matching results byη is set to 0.8 on three datasets, i.e., Potsdam inFigure 7a andFigure 8a, EPFL inFigure 7b andFigure 8b, and CVC inFigure 7c andFigure 8c, for subjective visual analysis.
    As shown inFigure 7, it can be seen that our proposed HOSM descriptor can get a good matching performance by preserving a large number of correct correspondences between the multispectral image pairs on all three datasets. In details, our method can get the maximum number of correct correspondences in EPFL dataset, followed by the results in Potsdam dataset, and the third best performance in CVC dataset. InFigure 7a,b, the matching point pairs are extracted in most areas, while the matching point pairs inFigure 7c are mainly preserved for the vegetation areas. This difference is caused by a variety of factors. First, FAST detector tends to get feature point responses in complex structures in the image. This results in the most responses being obtained in the vegetation areas, with almost no feature point responses in the grayscale smooth region, such as the wall portion, of the image. Second, due to the larger spectral differences in the CVC dataset, the number of initial matching points inFigure 7c is less than that inFigure 7a,b. The last reason is that when the mismatch points are removed with the NNDR test, some correctly matched points are also removed. Even so, our method still extracts enough correct matching points to get the correct correspondence between the two images.
    Figure 8 presents another group of multispectral image matching results on these three datasets.Figure 8a,b are visible and NIR image pairs selected from the Potsdam and EPFL datasets, andFigure 8c is a pair of visible and LWIR images selected from the CVC dataset. In general, the number of matches inFigure 8a,b is greater than that inFigure 8c. It can also be seen fromFigure 8c that there are several pairs of obvious mismatches which are marked by red rectangles. There are several factors that cause mismatches. First, the spectral difference between visible light and LWIR is greater than that between visible light and NIR, which leads to more challenging for feature matching. Second, the mismatched points have highly similar textures within the local regions, which results in producing highly similar feature vectors. We can also see that these failed matches have similar correspondence relationships to the correct matches, so they are not removed by the NNDR test. However, enough correct matches between the two images are still preserved.
    The experimental results demonstrate that our proposed HOSM descriptor can be robust to the intensity nonlinear variations between the multispectral images and obtain good matching performance.

    3.3. The Advantages of Guided Filtering

    In order to verify the advantages of the guided filtering in structure maps construction, we conduct the comparison tests between the HOSM descriptor and the descriptor without guided filtering on EPFL (VIS/NIR) dataset and CVC (VIS/LWIR) dataset. The qualitative and quantitative comparisons are presented inFigure 9 andTable 1.
    Figure 9 illustrates some samples of matching results. In general, it can be seen that the HOSM descriptor can achieve better matching performance than the descriptor without guided filtering. The matching results show that the HOSM descriptor has preserved a larger number of correct matching points on the two multispectral image pairs and has fewer false matches. This is because the guidance images have preserved the significant contour information in multispectral images, the guided filtering operation is able to enhance the structures of the structure maps to improve the description ability [38]. In detail, due to the abundant structures in the visible and NIR images, the descriptor without guided filtering can also achieve satisfactory results on the EPFL dataset, as shown inFigure 9a. When performing matching comparisons between visible and LWIR images, the performance of the HOSM descriptor is significantly better than the descriptor without guided filtering, as shown inFigure 9b. This is because the structure information in the LWIR image is weak, the enhancement of the structure by the guided filtering has a great influence on the matching results. Furtherly, the guided filtering operation can reduce aliasing effects to improve the robustness of the descriptor.
    Table 1 presents the quantitative metrics of precision, recall, and F1-score. For all metrics, larger values indicate better results. From these objective metrics, we can see that HOSM outperforms the descriptor without guided filtering. Therefore, the experimental results prove that guided filtering can improve matching performance, especially for images with weak structure information.

    3.4. Robustness Evaluations to Noise

    To evaluate the robustness of the proposed HOSM method to noise, we make qualitative and quantitative tests on Potsdam (VIS/NIR) dataset at different noise levels.Figure 10 andTable 2 present the quantitative comparisons on the average of precision and recall results, and the F1-scores. InFigure 11, some matching results at different noise levels are selected for qualitative comparisons.
    Due to the sensors and imaging principles, noise in IR images usually appears as stationary random white noise, which means that its position and intensity appearing in the image are randomly distributed [39,40]. In detail, the noise in the image usually appears as additive noise, and Gaussian white noise with a mean of 0 is the most common form [41]. For visible images, the intensity of the noise is small due to the short exposure time. Therefore, in the experiments, we only add zero-mean Gaussian white noise with different intensity in the IR images. The varianceσ of the Gaussian function can be used to represent the intensity of noise. In the experiment, we setσ to 0, 0.01, 0.05, 0.1, 0.15 and 0.2 respectively. The average of precision and recall curves are illustrated inFigure 10, and the F1-scores are shown inTable 2. The results show that although the values of precision, recall, and F1-score are declining as the noise increases, our method can still get good results. This is because the guided filtering operation can enhance the structures of the structure maps and reduce aliasing effects to improve the robustness of the descriptor.
    For subjective evaluations, some qualitative comparison results are shown inFigure 11. These results intuitively show that our HOSM descriptors can perform well in cases of intensive noise. Even when there are some obvious mismatches atσ=0.2, our method still obtains a lot of correct matching points. The good matching performances show that our method is efficient and robust to noise.

    3.5. Quantitative Evaluation on Metrics

    Figure 12 illustrates the average values of precision-recall pairs, andTable 3 shows the average values of the F1-score for the eight methods on all the three datasets. For all metrics, including precision, recall, and F1-score, larger numbers indicate better performance. It can be seen that the HOSM descriptor outperforms the other methods in the metrics on all the datasets. In general, the matching metrics of all the methods get the best values on EPFL dataset, the second-best values on Potsdam dataset, and the last results on CVC dataset. Compared to NIR images in the Potsdam and EPFL datasets, the LWIR images in the CVC dataset have lower resolution, while having greater nonlinear intensity differences with corresponding visible images. Therefore, it is most challenging for the CVC dataset on all descriptors to extract the common structure features from multispectral images.
    As shown inFigure 12a–c, the HOSM descriptor achieves the best performance in all precision-recall curves, followed by HoDM. Due to the oriented structure maps preserving the common features and structure guidance images overcoming nonlinear variations in the gradient direction between the multispectral images, the HOSM descriptor can be robust to get high precision and recall values in these datasets. The HoDM can get good matching performance due to the combination of structure and texture information. The MFD and LGHD have similar performance because both descriptors are progressive methods based on the EOH, in which the multi-scale and multi-oriented Log-Gabor filters were used to replace multi-directional spatial filters. Because the Log-Gabor filters are better at retaining the oriented edge characteristics of multispectral images, the MFD and LGHD methods can achieve better matching performance than the original EOH. It can be seen from theFigure 12a,c that the EOH is superior to SIFT, SURF and NG-SIFT, but the results are opposite inFigure 12b. This is because in the EPFL dataset, the spectra of the image pairs are close, SIFT, SURF, and NG-SIFT can obtain stable descriptors to get better matching results than EOH. However, EOH is more robust to image pairs with significant nonlinear intensity variations, as shown inFigure 12a,c.
    From the average of F1-score values inTable 3, where the best scores are marked in bold, we can more accurately compare the matching performance of all methods.Table 3 indicates that the HOSM method can obtain the highest average scores in all cases, followed by HoDM. LGHD can get good performance similar to MFD on all datasets, while EOH can achieve third best scores on Potsdam and CVC datasets. For EPFL dataset, these three SIFT-like methods are able to obtain better performance than EOH. The high F1-score values demonstrate that the HOSM descriptor not only achieves good matching performance, but also achieve a good balance between precision and recall.

    3.6. Quantitative Comparisons on Running Time

    A desktop with 4 GB memory and 2.5 GHz Intel Core i3 CPU is used to carry out the experiments and all these methods are implemented by the MATLAB codes.Figure 13 shows the average computation time of the methods in each feature point on all the three datasets, in details,Figure 13a is the average running time of Potsdam dataset,Figure 13b is the average running time of EPFL dataset, andFigure 13c is the average running time of CVC dataset. It can be seen that the HOSM is faster than other methods and achieves the best time consumption. This is because we have previously established the structure maps with linear guided filtering operation. The values of structure maps are directly used to build the HOSM descriptor, without other preprocessing. From all experimental results, it can be seen that our HOSM descriptor is robust to the nonlinear intensity variations in multispectral images and is superior to other methods.

    4. Conclusions

    In this paper, we propose a new local feature descriptor termed HOSM for multispectral remote sensing image matching task. First, we propose a new local contrast based operator to construct the structure guidance images for keeping the significant contour features in multispectral images. Then, construct the oriented structure maps based on multi-oriented Sobel edges with guided filtering. Guided image filtering can achieve better computational efficiency than traditional time-consuming trilinear interpolation methods. Finally, the histograms of oriented structure maps are computed to build the feature vectors. In order to verify the HOSM descriptor, three widely used datasets named Potsdam, EPFL, and CVC are employed to conduct feature matching tests, and seven state of the art methods are used for comparison. The experimental results show that our method is outstanding to address the nonlinear intensity changes in multispectral images and can achieve better matching performance.
    In the future, we will further study how to enhance the description ability of our method, because the HOSM descriptor is robust to nonlinear intensity changes but sensitive to rotation and scale invariance. We can assign a main orientation to the HOSM to cope with the rotation variance and use a multiscale keypoint detector to keep scale invariance.

    Author Contributions

    Writing—original draft, T.M.; Writing—review & editing, J.M; Software, K.Y.

    Funding

    This research received no external funding.

    Conflicts of Interest

    The authors declare no conflict of interest.

    References

    1. Zhou, W.; Wang, C.; Xiao, B.; Zhang, Z. SLD: A Novel Robust Descriptor for Image Matching.IEEE Signal Process. Lett.2014,21, 339–342. [Google Scholar] [CrossRef]
    2. Ma, J.; Ma, Y.; Zhao, J.; Tian, J. Image Feature Matching via Progressive Vector Field Consensus.IEEE Signal Process. Lett.2015,22, 767–771. [Google Scholar] [CrossRef]
    3. Zhu, B.; Xin, Y. Effective and robust infrared small target detection with the fusion of poly directional first order derivative images under facet model.Infrared Phys. Technol.2015,69, 136–144. [Google Scholar] [CrossRef]
    4. Chen, Y.; Ogata, T.; Ueyama, T.; Takada, T.; Ota, J. Automated Field-of-View, Illumination, and Recognition Algorithm Design of a Vision System for Pick-and-Place Considering Colour Information in Illumination and Images.Sensors2018,18, 1656. [Google Scholar] [CrossRef] [PubMed]
    5. Shi, Q.; Ma, G.; Zhang, F.; Chen, W.; Qin, Q.; Duo, H. Robust Image Registration Using Structure Features.IEEE Geosci. Remote Sens. Lett.2014,11, 2045–2049. [Google Scholar]
    6. Guislain, M.; Digne, J.; Chaine, R.; Monnier, G. Fine scale image registration in large-scale urban LIDAR point sets.Comput. Vis. Image Underst.2017,157, 90–102. [Google Scholar] [CrossRef] [Green Version]
    7. Song, J.; Liu, L.; Huang, W.; Li, Y.; Chen, X.; Zhang, Z. Target detection via HSV color model and edge gradient information in infrared and visible image sequences under complicated background.Opt. Quantum Electron.2018,50, 171–175. [Google Scholar] [CrossRef]
    8. Li, Y.; Tao, C.; Tan, Y.; Shang, K.; Tian, J. Unsupervised Multilayer Feature Learning for Satellite Image Scene Classification.IEEE Geosci. Remote Sens. Lett.2016,13, 157–161. [Google Scholar] [CrossRef]
    9. Jung, J.; Sohn, G.; Bang, K.; Wichmann, A.; Armenakis, C.; Kada, M. Matching Aerial Images to 3D Building Models Using Context-Based Geometric Hashing.Sensors2016,16, 932. [Google Scholar] [CrossRef]
    10. Zhao, C.; Sun, L.; Purkait, P.; Duckett, T.; Stolkin, R. Dense RGB-D Semantic Mapping with Pixel-Voxel Neural Network.Sensors2018,18, 3099. [Google Scholar] [CrossRef] [PubMed]
    11. Ma, T.; Ma, J.; Fang, B.; Hu, F.; Quan, S.; Du, H. Multi-scale decomposition based fusion of infrared and visible image via total variation and saliency analysis.Infrared Phys. Technol.2018,92, 154–162. [Google Scholar] [CrossRef]
    12. Feng, Q.; Hao, Q.; Chen, Y.; Yi, Y.; Wei, Y.; Dai, J. Hybrid Histogram Descriptor: A Fusion Feature Representation for Image Retrieval.Sensors2018,18, 1943. [Google Scholar] [CrossRef] [PubMed]
    13. Nguyen, D.; Baek, N.; Pham, T.; Park, K. Presentation Attack Detection for Iris Recognition System Using NIR Camera Sensor.Sensors2018,18, 1315. [Google Scholar] [CrossRef]
    14. So, R.W.K.; Chung, A.C.S. A novel learning-based dissimilarity metric for rigid and non-rigid medical image registration by using Bhattacharyya Distances.Pattern Recognit.2017,62, 161–174. [Google Scholar] [CrossRef]
    15. Ye, Y.; Shan, J.; Bruzzone, L.; Shen, L. Robust registration of multimodal remote sensing images based on structural similarity.Ieee Trans. Geosci. Remote Sens.2017,55, 2941–2958. [Google Scholar] [CrossRef]
    16. Liu, Y.; Mo, F.; Tao, P. Matching Multi-Source Optical Satellite Imagery Exploiting a Multi-Stage Approach.Remote Sens.2017,9, 1249. [Google Scholar] [CrossRef]
    17. He, C.; Fang, P.; Xiong, D.; Wang, W.; Liao, M. A Point Pattern Chamfer Registration of Optical and SAR Images Based on Mesh Grids.Remote Sens.2018,10, 1837. [Google Scholar] [CrossRef]
    18. Ma, J.; Zhao, J.; Tian, J.; Bai, X.; Tu, Z. Regularized vector field learning with sparse approximation for mismatch removal.Pattern Recognit.2013,46, 3519–3532. [Google Scholar] [CrossRef]
    19. Yu, M.; Deng, K.; Yang, H.; Qin, C. Improved WαSH Feature Matching Based on 2D-DWT for Stereo Remote Sensing Images.Sensors2018,18, 3494. [Google Scholar] [CrossRef]
    20. Wang, G.; Wang, X.; Fan, B.; Pan, C. Feature Extraction by Rotation-Invariant Matrix Representation for Object Detection in Aerial Image.Ieee Geosci. Remote Sens. Lett.2017,14, 851–855. [Google Scholar] [CrossRef]
    21. Lowe, D.G. Distinctive image features from scale-invariant keypoints.Int. J. Comput. Vis.2004,60, 91–110. [Google Scholar] [CrossRef]
    22. Bay, H.; Tuytelaars, T.; Van Gool, L. SURF: Speeded Up Robust Features. In Proceedings of the European Conference on Computer Vision (ECCV), Graz, Austria, 7–13 May 2006; pp. 404–417. [Google Scholar]
    23. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-Up Robust Features (SURF).Comput. Vis. Image Underst.2008,110, 346–359. [Google Scholar] [CrossRef] [Green Version]
    24. Saleem, S.; Sablatnig, R. A Robust SIFT Descriptor for Multispectral Images.IEEE Signal Process. Lett.2014,21, 400–403. [Google Scholar] [CrossRef]
    25. Wang, G.; Wang, Z.; Chen, Y.; Zhao, W. Robust point matching method for multimodal retinal image registration.Biomed. Signal Process. Control.2015,19, 68–76. [Google Scholar] [CrossRef]
    26. Manjunath, B.S.; Ohm, J.; Vasudevan, V.V.; Yamada, A. Color and texture descriptors.IEEE Trans. Circuits Syst. Video Technol.2001,11, 703–715. [Google Scholar] [CrossRef] [Green Version]
    27. Aguilera, C.; Barrera, F.; Lumbreras, F.; Sappa, A.D.; Toledo, R. Multispectral Image Feature Points.Sensors2012,12, 12661–12672. [Google Scholar] [CrossRef] [Green Version]
    28. Fu, Z.; Qin, Q.; Luo, B.; Wu, C.; Sun, H. A Local Feature Descriptor Based on Combination of Structure and Texture Information for Multispectral Image Matching.IEEE Geosci. Remote Sens. Lett.2019,16, 100–104. [Google Scholar] [CrossRef]
    29. Aguilera, C.A.; Sappa, A.D.; Toledo, R. LGHD: A feature descriptor for matching across non-linear intensity variations. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 178–181. [Google Scholar]
    30. Nunes, C.F.G.; Pádua, F.L.C. A Local Feature Descriptor Based on Log-Gabor Filters for Keypoint Matching in Multispectral Images.IEEE Geosci. Remote Sens. Lett.2017,14, 1850–1854. [Google Scholar] [CrossRef]
    31. Sikora, T. The MPEG-7 visual standard for content description―An overview.IEEE Transations Circuits Syst. Video Technol.2001,11, 696–702. [Google Scholar] [CrossRef]
    32. He, K.; Sun, J.; Tang, X. Guided Image Filtering. In Proceedings of the European Conference on Computer Vision (ECCV), Heraklion, Crete, Greece, 5–11 September 2010; pp. 1–14. [Google Scholar]
    33. He, K.; Sun, J.; Tang, X. Guided Image Filtering.IEEE Trans. Pattern Anal. Mach. Intell.2013,35, 1397–1409. [Google Scholar] [CrossRef]
    34. Zhao, H.; Jiang, L.; Jin, X.; Hui, D.; Li, X. Constant time texture filtering.Vis. Comput.2018,34, 83–92. [Google Scholar] [CrossRef]
    35. Potsdam Dataset of Remote Sensing Images, Distributed by the International Society for Photogrammetry and Remote Sensing. Available online:http://www2.isprs.org/commissions/comm3/wg4/2dsem-label-potsdam.html (accessed on 8 September 2018).
    36. Brown, M.; Süsstrunk, S. Multi-spectral SIFT for scene category recognition. In Proceedings of the 2011 IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Fort Collins, CO, USA, 20–25 June 2011; pp. 177–184. [Google Scholar]
    37. Rosten, E.; Porter, R.; Drummond, T. Faster and Better: A Machine Learning Approach to Corner Detection.Ieee Trans. Pattern Anal. Mach. Intell.2010,32, 105–119. [Google Scholar] [CrossRef] [PubMed] [Green Version]
    38. Choi, J.; Park, H.; Seo, D. Pansharpening Using Guided Filtering to Improve the Spatial Clarity of VHR Satellite Imagery.Remote Sens.2019,11, 633. [Google Scholar] [CrossRef]
    39. Kim, Y.S.; Lee, J.H.; Ra, J.B. Multi-sensor image registration based on intensity and edge orientation information.Pattern Recognit.2008,41, 3356–3365. [Google Scholar] [CrossRef]
    40. Liu, X.; Ai, Y.; Zhang, J.; Wang, Z. A Novel Affine and Contrast Invariant Descriptor for Infrared and Visible Image Registration.Remote Sens.2018,10, 658. [Google Scholar] [CrossRef]
    41. Keller, Y.; Averbuch, A. Multisensor image registration via implicit similarity.IEEE Trans. Pattern Anal. Mach. Intell.2006,28, 794–801. [Google Scholar] [CrossRef] [PubMed]
    Remotesensing 11 00951 g001 550
    Figure 1. Gradient direction variations and structure guidance images between multispectral images (a) visible image, and (b) infrared image. From top to bottom: (1) source images, (2) gradient directions, and (3) structure guidance images. The green dot indicates the position of the interest point.
    Figure 1. Gradient direction variations and structure guidance images between multispectral images (a) visible image, and (b) infrared image. From top to bottom: (1) source images, (2) gradient directions, and (3) structure guidance images. The green dot indicates the position of the interest point.
    Remotesensing 11 00951 g001
    Remotesensing 11 00951 g002 550
    Figure 2. Processing flow of structure guidance image construction. The red point indicates the center pixel of the local region.
    Figure 2. Processing flow of structure guidance image construction. The red point indicates the center pixel of the local region.
    Remotesensing 11 00951 g002
    Remotesensing 11 00951 g003 550
    Figure 3. Sobel filters of five orientations at (a) 0°, (b) 45°, (c) 90°, (d) 135°, and (e) no orientation.
    Figure 3. Sobel filters of five orientations at (a) 0°, (b) 45°, (c) 90°, (d) 135°, and (e) no orientation.
    Remotesensing 11 00951 g003
    Remotesensing 11 00951 g004 550
    Figure 4. Illustration of oriented structure maps construction. From top to bottom: (a) Sobel edgesSEn, (b) edge mapsEMn, and (c) structure mapsSMn.
    Figure 4. Illustration of oriented structure maps construction. From top to bottom: (a) Sobel edgesSEn, (b) edge mapsEMn, and (c) structure mapsSMn.
    Remotesensing 11 00951 g004
    Remotesensing 11 00951 g005 550
    Figure 5. Main flowchart of the proposed HOSM descriptor. The green point of the region center indicates an interest point.
    Figure 5. Main flowchart of the proposed HOSM descriptor. The green point of the region center indicates an interest point.
    Remotesensing 11 00951 g005
    Remotesensing 11 00951 g006 550
    Figure 6. Samples of the multispectral image pairs from three datasets. (a) Potsdam (VIS/NIR) dataset, (b) EPFL (VIS/NIR) dataset, and (c) CVC (VIS/LWIR) dataset.
    Figure 6. Samples of the multispectral image pairs from three datasets. (a) Potsdam (VIS/NIR) dataset, (b) EPFL (VIS/NIR) dataset, and (c) CVC (VIS/LWIR) dataset.
    Remotesensing 11 00951 g006
    Remotesensing 11 00951 g007 550
    Figure 7. Sample one of feature matching results by the HOSM descriptor for the multispectral image pairs in three datasets. From top to bottom: matching results on (a) Potsdam dataset, (b) EPFL dataset, and (c) CVC dataset.
    Figure 7. Sample one of feature matching results by the HOSM descriptor for the multispectral image pairs in three datasets. From top to bottom: matching results on (a) Potsdam dataset, (b) EPFL dataset, and (c) CVC dataset.
    Remotesensing 11 00951 g007
    Remotesensing 11 00951 g008 550
    Figure 8. Sample two of feature matching results by the HOSM descriptor for the multispectral image pairs in three datasets. From top to bottom: matching results on (a) Potsdam dataset, (b) EPFL dataset, and (c) CVC dataset.
    Figure 8. Sample two of feature matching results by the HOSM descriptor for the multispectral image pairs in three datasets. From top to bottom: matching results on (a) Potsdam dataset, (b) EPFL dataset, and (c) CVC dataset.
    Remotesensing 11 00951 g008
    Remotesensing 11 00951 g009 550
    Figure 9. Comparisons of matching results between the HOSM descriptor and the descriptor without guided filtering. The left column is the matching results of the HOSM descriptor, and the right column is the matching results of the descriptor without guided filtering. From top to bottom: matching results on (a) EPFL (VIS/NIR) dataset and (b) CVC (VIS/LWIR) dataset.
    Figure 9. Comparisons of matching results between the HOSM descriptor and the descriptor without guided filtering. The left column is the matching results of the HOSM descriptor, and the right column is the matching results of the descriptor without guided filtering. From top to bottom: matching results on (a) EPFL (VIS/NIR) dataset and (b) CVC (VIS/LWIR) dataset.
    Remotesensing 11 00951 g009
    Remotesensing 11 00951 g010 550
    Figure 10. Quantitative comparisons on the average of precision and recall curves for Potsdam (VIS/NIR) dataset at different noise levels. The Gaussian white noise with a mean of zero is adopted. A large varianceσ indicates great noise intensity.
    Figure 10. Quantitative comparisons on the average of precision and recall curves for Potsdam (VIS/NIR) dataset at different noise levels. The Gaussian white noise with a mean of zero is adopted. A large varianceσ indicates great noise intensity.
    Remotesensing 11 00951 g010
    Remotesensing 11 00951 g011 550
    Figure 11. Qualitative comparisons of feature matching on Potsdam (VIS/NIR) dataset at different Gaussian white noise levels.σ is the variance of Gaussian function, and large varianceσ indicates great noise intensity.
    Figure 11. Qualitative comparisons of feature matching on Potsdam (VIS/NIR) dataset at different Gaussian white noise levels.σ is the variance of Gaussian function, and large varianceσ indicates great noise intensity.
    Remotesensing 11 00951 g011
    Remotesensing 11 00951 g012 550
    Figure 12. Quantitative comparisons on precision-recall averages with NNDR thresholds from 0.8 to 1.0 in three datasets. (a) Potsdam dataset, (b) EPFL dataset, and (c) CVC dataset.
    Figure 12. Quantitative comparisons on precision-recall averages with NNDR thresholds from 0.8 to 1.0 in three datasets. (a) Potsdam dataset, (b) EPFL dataset, and (c) CVC dataset.
    Remotesensing 11 00951 g012
    Remotesensing 11 00951 g013 550
    Figure 13. Quantitative comparisons on average computation time for all the descrtiptors. (a) Potsdam dataset, (b) EPFL dataset, and (c) CVC dataset.
    Figure 13. Quantitative comparisons on average computation time for all the descrtiptors. (a) Potsdam dataset, (b) EPFL dataset, and (c) CVC dataset.
    Remotesensing 11 00951 g013
    Table
    Table 1. Precision, recall and F1-score for HOSM and descriptor without guided filtering.
    Table 1. Precision, recall and F1-score for HOSM and descriptor without guided filtering.
    PairsHOSMDescriptor Without Guided Filtering
    PrecisionRecallF1-ScorePrecisionRecallF1-Score
    Figure 9a0.7850.6500.7110.5420.4130.468
    Figure 9b0.3120.2430.2730.0540.0450.049
    Table
    Table 2. Average of F1-score values with different noise levels.
    Table 2. Average of F1-score values with different noise levels.
    MetricsValues
    σ00.010.050.10.150.2
    F1-score0.5120.4430.3970.3650.3350.305
    Table
    Table 3. Average of F1-score values for all eight methods on all three datasets.
    Table 3. Average of F1-score values for all eight methods on all three datasets.
    DatasetsηMethods
    SIFTSURFNG-SIFTEOHLGHDMFDHoDMHOSM
    Potsdam0.80.2580.1710.2830.2750.3960.4170.4520.471
    0.850.2650.1790.2950.2920.4120.4220.4540.480
    0.90.2570.1790.2940.3090.4080.4240.4660.484
    0.950.2610.1750.2840.3150.4120.4180.4620.479
    1.00.2440.1740.2740.3230.4190.4140.4430.482
    EPFL0.80.6710.6110.6680.4710.7050.7270.7360.761
    0.850.6870.6230.6860.5080.7120.7310.7410.767
    0.90.6910.6370.6990.5370.7230.7340.7450.773
    0.950.6980.6460.7140.5570.7310.7420.7510.778
    1.00.7090.6580.7170.5630.7350.7460.7620.780
    CVC0.80.060.0320.0820.1010.1190.1290.1470.151
    0.850.0780.0550.0820.1040.1260.1350.1520.166
    0.90.0750.0580.0820.1070.1260.1340.1560.174
    0.950.0770.0670.0820.1070.1260.1370.1590.184
    1.00.080.0740.0800.1120.1320.1440.1490.191

    © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).

    Share and Cite

    MDPI and ACS Style

    Ma, T.; Ma, J.; Yu, K. A Local Feature Descriptor Based on Oriented Structure Maps with Guided Filtering for Multispectral Remote Sensing Image Matching.Remote Sens.2019,11, 951. https://doi.org/10.3390/rs11080951

    AMA Style

    Ma T, Ma J, Yu K. A Local Feature Descriptor Based on Oriented Structure Maps with Guided Filtering for Multispectral Remote Sensing Image Matching.Remote Sensing. 2019; 11(8):951. https://doi.org/10.3390/rs11080951

    Chicago/Turabian Style

    Ma, Tao, Jie Ma, and Kun Yu. 2019. "A Local Feature Descriptor Based on Oriented Structure Maps with Guided Filtering for Multispectral Remote Sensing Image Matching"Remote Sensing 11, no. 8: 951. https://doi.org/10.3390/rs11080951

    APA Style

    Ma, T., Ma, J., & Yu, K. (2019). A Local Feature Descriptor Based on Oriented Structure Maps with Guided Filtering for Multispectral Remote Sensing Image Matching.Remote Sensing,11(8), 951. https://doi.org/10.3390/rs11080951

    Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further detailshere.

    Article Metrics

    No
    No

    Article Access Statistics

    For more information on the journal statistics, clickhere.
    Multiple requests from the same IP address are counted as one view.
    Remote Sens., EISSN 2072-4292, Published by MDPI
    RSSContent Alert

    Further Information

    Article Processing Charges Pay an Invoice Open Access Policy Contact MDPI Jobs at MDPI

    Guidelines

    For Authors For Reviewers For Editors For Librarians For Publishers For Societies For Conference Organizers

    MDPI Initiatives

    Sciforum MDPI Books Preprints.org Scilit SciProfiles Encyclopedia JAMS Proceedings Series

    Follow MDPI

    LinkedIn Facebook X
    MDPI

    Subscribe to receive issue release notifications and newsletters from MDPI journals

    © 1996-2025 MDPI (Basel, Switzerland) unless otherwise stated
    Terms and Conditions Privacy Policy
    We use cookies on our website to ensure you get the best experience.
    Read more about our cookieshere.
    Accept
    Back to TopTop
    [8]ページ先頭

    ©2009-2025 Movatter.jp