The development of salient object detection is crucial in ubiquitous applications. Existing state-of-the-art models tend to have complex designs and a significant number of parameters, prioritizing performance improvement over efficiency.… Click to show full abstract
The development of salient object detection is crucial in ubiquitous applications. Existing state-of-the-art models tend to have complex designs and a significant number of parameters, prioritizing performance improvement over efficiency. Hence, there pose significant challenges to deploying them in edge devices. The intricacy in these models stems from the complicated encoder-decoder that aims to effectively generate and integrate coarse and semantic features. To address this problem, we introduced EC2Net, an efficient attention-based cross-context network for salient object detection. To start with, we introduce the shallow crossed-context aggregation (SCCA) mechanism to enhance and preserve object boundaries for shallow layers. We introduced a deep cross-context aggregation (DCCA) mechanism to enhance semantic features in deep layers. Subsequently, we introduced the dual cross-fusion module (DCFM) to efficiently merge shallow and deep features. The proposed modules complement each other, enabling EC2Net to accurately detect salient objects with reduced computational overhead. Through experiments on five standard datasets, the proposed method demonstrated competitive performance while utilizing fewer parameters, FLOPS, and memory storage than other resource-intensive models.
               
Click one of the above tabs to view related content.