Coherent change detection (CCD) is an effective method to detect subtle scene changes that occur between temporal synthetic aperture radar (SAR) observations. Most coherence estimators are obtained from a Hermitian product based on local statistics. Increasing the number of samples in the local window can improve the estimation bias, but cause the loss of the estimated images spatial resolution. The limitations of these estimators lead to unclear contour of the disturbed region, and even the omission of fine change targets. In this paper, a CCD approach is proposed to detect fine scene changes from multi-temporal and multi-angle SAR image pairs. Multi-angle CCD estimator can improve the contrast between the change target and the background clutter by jointly accumulating single-angle alternative estimator results without further loss of image resolution. The sensitivity of detection performance to image quantity and angle interval is analyzed. Theoretical analysis and experimental results verify the performance of the proposed algorithm.
In the field of satellite imagery, remote sensing image captioning (RSIC) is a hot topic with the challenge of overfitting and difficulty of image and text alignment. To address these issues, this paper proposes a vision-language aligning paradigm for RSIC to jointly represent vision and language. First, a new RSIC dataset DIOR-Captions is built for augmenting object detection in optical remote (DIOR) sensing images dataset with manually annotated Chinese and English contents. Second, a Vision-Language aligning model with Cross-modal Attention (VLCA) is presented to generate accurate and abundant bilingual descriptions for remote sensing images. Third, a cross-modal learning network is introduced to address the problem of visual-lingual alignment. Notably, VLCA is also applied to end-to-end Chinese captions generation by using the pre-training language model of Chinese. The experiments are carried out with various baselines to validate VLCA on the proposed dataset. The results demonstrate that the proposed algorithm is more descriptive and informative than existing algorithms in producing captions.
With the rapidly growing abuse of drones, monitoring and classification of birds and drones have become a crucial safety issue. With similar low radar cross sections (RCSs), velocities, and heights, drones are usually difficult to be distinguished from birds in radar measurements. In this paper, we propose to exploit different periodical motions of birds and drones from high-resolution Doppler spectrum sequences (DSSs) for classification. This paper presents an elaborate feature vector representing the periodic fluctuations of RCS and micro kinematics. Fed by the Doppler spectrum and feature sequence, the long to short-time memory (LSTM) is used to solve the time series classification. Different classification schemes to exploit the Doppler spectrum series are validated and compared by extensive real-data experiments, which confirms the effectiveness and superiorities of the proposed algorithm.
In this paper, we study autonomous landing scene recognition with knowledge transfer for drones. Considering the difficulties in aerial remote sensing, especially that some scenes are extremely similar, or the same scene has different representations in different altitudes, we employ a deep convolutional neural network (CNN) based on knowledge transfer and fine-tuning to solve the problem. Then, LandingScenes-7 dataset is established and divided into seven classes. Moreover, there is still a novelty detection problem in the classifier, and we address this by excluding other landing scenes using the approach of thresholding in the prediction stage. We employ the transfer learning method based on ResNeXt-50 backbone with the adaptive momentum (ADAM) optimization algorithm. We also compare ResNet-50 backbone and the momentum stochastic gradient descent (SGD) optimizer. Experiment results show that ResNeXt-50 based on the ADAM optimization algorithm has better performance. With a pre-trained model and fine-tuning, it can achieve 97.8450% top-1 accuracy on the LandingScenes-7 dataset, paving the way for drones to autonomously learn landing scenes.
The development of image classification is one of the most important research topics in remote sensing. The prediction accuracy depends not only on the appropriate choice of the machine learning method but also on the quality of the training datasets. However, real-world data is not perfect and often suffers from noise. This paper gives an overview of noise filtering methods. Firstly, the types of noise and the consequences of class noise on machine learning are presented. Secondly, class noise handling methods at both the data level and the algorithm level are introduced. Then ensemble-based class noise handling methods including class noise removal, correction, and noise robust ensemble learners are presented. Finally, a summary of existing data-cleaning techniques is given.
Synthetic aperture radar (SAR) is usually sensitive to trajectory deviations that cause serious motion error in the recorded data. In this paper, a coherent range-dependent map-drift (CRDMD) algorithm is developed to accommodate the range-variant motion errors. By utilizing the algorithm as an estimate core, robust motion compensation strategy is proposed for unmanned aerial vehicle (UAV) SAR imagery. CRDMD outperforms the conventional map-drift algorithms in both accuracy and efficiency. Real data experiments show that the proposed approach is appropriate for precise motion compensation for UAV SAR.