Abstract This paper investigates the split complex gradient descent based neuro-fuzzy algorithm with self-adaptive momentum and L2 regularizer for training TSK (Takagi–Sugeno–Kang) fuzzy inference models. The major threat for disposing… Click to show full abstract
Abstract This paper investigates the split complex gradient descent based neuro-fuzzy algorithm with self-adaptive momentum and L2 regularizer for training TSK (Takagi–Sugeno–Kang) fuzzy inference models. The major threat for disposing complex data with fuzzy system is contradiction of boundedness and analyticity in the complex domain, as expressed by Liouville’s theorem. The proposed algorithm operates a couple of real-valued functions and splits the complex variables into real part and imaginary part. Dynamical momentum is included in the learning mechanism to promote learning speed. L2 regularizer is also added to control the magnitude of the weight parameters. Furthermore, a detailed convergence analysis of the proposed algorithm is fully studied. The monotonic decreasing property of the error function and convergence of the weight sequence are guaranteed. Plus a mild condition, strong convergence of the weight sequence is deduced. Finally, the simulation results are also demonstrated to verify the theoretical analysis results.
               
Click one of the above tabs to view related content.