An Online Gradient Method with Smoothing L_0 Regularization for Pi-Sigma Network

Authors

  • Khidir Shaib Mohamed Department of Mathematics Computer, Faculty of Science, Dalanj University, Dalanj, Sudan;
  • Yousif Shoaib Mohammed Department of Physics, College of Science & Art, Qassim University, Oklat Al- Skoor, Saudi Arabia

DOI:

https://doi.org/10.14738/tmlai.66.5838

Keywords:

Convergence, Online gradient method, Pi-Sigma networks, Smoothing L_0 regularization.

Abstract

The description of this study is to make possibility analysis solution of online gradient method with smoothing  regularization for pi-sigma network training. Due to the effectiveness computational and theoretical analysis are a very important issues to improve the generalization performance of networks and the gradient descent algorithm with regularization is widely used method. However,  regularization is reefed to NP-hard nature problems, which has not differentiable objective functional-penalty term. In this paper to avoid this trick, we use a smoothing function to recover the origin  regularization into smoothing  regularization. Under this condition, the resulting obtained as a good decreases solution when compared with others. The monotonically of the error function, weak and strong convergence theorems are proved.

References

(1) Shin, Y., et al., The pi-sigma network: An efficient higher-order neural network for pattern classification and function approximation. In Neural Networks, 1991., IJCNN-91-Seattle International Joint Conference on IEEE, 1991. 1: p. 13-18.

(2) Akram, U., et al., A comprehensive Survey on Pi-Sigma Neural Network for Time Series Prediction. Journal of Telecommunication, Electronic and Computer Engineering (JTEC), 2017. 9(3-3): p.57-62.

(3) Shin, Y., et al., Computationally efficient invariant pattern recognition with higher order Pi-Sigma Networks. The University of Texas at Austin, 1992.

(4) Setiono, R., A penalty-function approach for pruning feedforward neural networks. Neural computation, 1997. 9(1): p.185-204.

(5) Reed, R., Pruning algorithms-a survey, IEEE Trans Neural Network, 1993. 4 (5): p.740–7.

(6) Wu, W., et al., Deterministic convergence of an online gradient method for BP neural networks. IEEE Transactions on Neural Networks, 2005. 16(3): p.533-540.

(7) Hagan, M.T., et al., Training feedforward networks with the Marquardt algorithm. IEEE transactions on Neural Networks, 1994. 5(6): p.989-993.

(8) Akaike, H., A new look at the statistical model identification. IEEE transactions on automatic control, 1974. 19(6): p.716-723.

(9) Donoho, D. L., Neighborly Polytopes and the Sparse Solution of Underdetermined Systems of Linear Equations. To appear. IEEE Trans. Info. Thry (2005).

(10) Tibshirani, R., Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society. Series B (Methodological), 1996. P. 267-288.

(11) Reed, R., Pruning algorithms-a survey, IEEE Trans Neural Network, 1993. 4 (5): p.740–7.

(12) Hinton, G.E., Connectionist learning procedures. Artificial intelligence, 1989. 40(1-3): 185-234.

(13) Zhang, H., et al., Online gradient method with smoothing ‘0regularization for feedforward neural networks. Neurocomputing, 2017. 224: p.1-8.

(14) Fan, Q., et al., Convergence of online gradient method for feedforward neural networks with smoothing L1/2regularization penalty. Neurocomputing, 2014. 131: p.208-216.

(15) Xu, Z.B., et al., When does online BP training converge?. IEEE Transactions on Neural Networks, 2009. 20(10): p.1529-1539.

Downloads

Published

2018-12-31

How to Cite

Mohamed, K. S., & Mohammed, Y. S. (2018). An Online Gradient Method with Smoothing L_0 Regularization for Pi-Sigma Network. Transactions on Engineering and Computing Sciences, 6(6), 96. https://doi.org/10.14738/tmlai.66.5838