LAUSR.org creates dashboard-style pages of related content for over 1.5 million academic articles. Sign Up to like articles & get recommendations!

Robust sequential subspace clustering via ℓ1-norm temporal graph

Photo from wikipedia

Abstract Subspace clustering (SC) has been widely applied to segment data drawn from multiple subspaces. However, for sequential data, a main challenge in subspace clustering is to exploit temporal information.… Click to show full abstract

Abstract Subspace clustering (SC) has been widely applied to segment data drawn from multiple subspaces. However, for sequential data, a main challenge in subspace clustering is to exploit temporal information. In this paper, we propose a novel robust sequential subspace clustering approach with a l1-norm temporal graph. The l1-norm temporal graph is designed to encode the temporal information underlying in sequential data. By using the l2 norm, it can enforce well temporal similarity of neighboring frames with a sample-dependent weight, and mitigate the effect of noises and outliers on subspace clustering because large errors mixed in the real data can be suppressed. Under assumption of data self-expression, our clustering model is put forward by further integrating the classical Sparse Subspace Clustering and the l1-norm Temporal Graph (SSC-L1TG). To solve the proposed model, we introduce a new efficient proximity algorithm. At each iteration, the sub-problem is solved by proximal minimization with closed-form solution. In contrast to the alternating direction method of multipliers (ADMM) employed in most existing clustering approaches without convergence guarantee, the proposed SSC-L1TG is guaranteed to converge to the desired optimal solution. Experimental results on both synthetic and real data demonstrate the efficacy of our method and its superior performance over the state-of-the-art methods.

Keywords: temporal graph; subspace clustering; norm temporal; robust sequential; subspace

Journal Title: Neurocomputing
Year Published: 2020

Link to full text (if available)


Share on Social Media:                               Sign Up to like & get
recommendations!

Related content

More Information              News              Social Media              Video              Recommended



                Click one of the above tabs to view related content.