Recently, many studies have successfully exploited convolutional neural networks to improve the performance of object detection in remote sensing images. However, detecting tiny objects is still challenging for two main… Click to show full abstract
Recently, many studies have successfully exploited convolutional neural networks to improve the performance of object detection in remote sensing images. However, detecting tiny objects is still challenging for two main neglected problems: i) The features of tiny objects are insufficient and prone to aliasing during the multiresolution aggregation process; ii) Tiny objects are position-sensitive, resulting in poor localization capabilities. In this article, we propose a feature-and-spatial aligned network (FSANet) to alleviate these issues. FSANet is an anchor-free detector that utilizes the alignment mechanism and progressive optimization strategy to obtain more discriminative features and accurate localization results. Concretely, we first present a feature-aware alignment module ( ${\rm FA}^{{2}}\text{M}$ ) to align features with different resolutions during fusion. By learning transformation offsets, ${\rm FA}^{{2}}\text{M}$ reencodes pixel-spatial information between feature maps of adjacent levels and adaptively adjusts the regular feature interpolation. Additionally, a spatial-aware guidance head (SAGH) is introduced to iteratively optimize network predictions in a coarse-to-fine fashion. SAGH first predicts the object shape at each spatial location on feature maps. For more precise predictions, it then captures geometry-aware convolutional features accordingly to update the coarse localization estimations. Extensive experiments are conducted on three tiny object detection datasets, i.e., AI-TOD, GF1-LRSD, and TinyPerson, demonstrating the effectiveness of our approach.
               
Click one of the above tabs to view related content.