Open Access   Article Go Back

Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks

K. Devkota1 , P. Bhattarai2

  1. Department of Electronics and Computer Engineering, Institute of Engineering (Tribhuvan University), Kathmandu, Nepal.
  2. Department of Electronics and Computer Engineering, Institute of Engineering (Tribhuvan University), Kathmandu, Nepal.

Correspondence should be addressed to: kdkapildevkota@gmail.com.

Section:Research Paper, Product Type: Journal Paper
Volume-5 , Issue-10 , Page no. 266-272, Oct-2017

CrossRef-DOI:   https://doi.org/10.26438/ijcse/v5i10.266272

Online published on Oct 30, 2017

Copyright © K. Devkota, P. Bhattarai . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

View this paper at   Google Scholar | DPI Digital Library

How to Cite this Paper

  • IEEE Citation
  • MLA Citation
  • APA Citation
  • BibTex Citation
  • RIS Citation

IEEE Style Citation: K. Devkota, P. Bhattarai, “Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks,” International Journal of Computer Sciences and Engineering, Vol.5, Issue.10, pp.266-272, 2017.

MLA Style Citation: K. Devkota, P. Bhattarai "Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks." International Journal of Computer Sciences and Engineering 5.10 (2017): 266-272.

APA Style Citation: K. Devkota, P. Bhattarai, (2017). Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks. International Journal of Computer Sciences and Engineering, 5(10), 266-272.

BibTex Style Citation:
@article{Devkota_2017,
author = {K. Devkota, P. Bhattarai},
title = {Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks},
journal = {International Journal of Computer Sciences and Engineering},
issue_date = {10 2017},
volume = {5},
Issue = {10},
month = {10},
year = {2017},
issn = {2347-2693},
pages = {266-272},
url = {https://www.ijcseonline.org/full_paper_view.php?paper_id=1511},
doi = {https://doi.org/10.26438/ijcse/v5i10.266272}
publisher = {IJCSE, Indore, INDIA},
}

RIS Style Citation:
TY - JOUR
DO = {https://doi.org/10.26438/ijcse/v5i10.266272}
UR - https://www.ijcseonline.org/full_paper_view.php?paper_id=1511
TI - Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks
T2 - International Journal of Computer Sciences and Engineering
AU - K. Devkota, P. Bhattarai
PY - 2017
DA - 2017/10/30
PB - IJCSE, Indore, INDIA
SP - 266-272
IS - 10
VL - 5
SN - 2347-2693
ER -

VIEWS PDF XML
503 221 downloads 162 downloads
  
  
           

Abstract

The problem of computational efficiency in adaptive algorithms, which is current and pressing, can be solved through their implementation in parallel frameworks, like CUDA, OpenCL, etc. The approach taken to parallelize any complex operation requires its separation into several distinct and independent sub-operations. We employed the same procedure to parallelize the BP (or Backpropagation) network algorithm. The function breakdown of the BP network involved breaking its overall operation into Feed-forward and Back-Propagate sub-operations, which was further divided into smaller independent execution groups. We applied parallel constructs on those independent execution groups and used the MNIST dataset to compare the algorithm’s performance with respect to the sequential algorithm. Comparing their performances, we found that the efficiency of the algorithm depended on the size of the BP network. In the large network with massive number of weight connections, we saw a significant improvement in the convergence time. This makes our algorithm preferable in feedforward networks having large number of hidden layers, neurons and weight connections.

Key-Words / Index Term

Backpropagation, Supervised Learning, CUDA, parallel

References

[1] G. S. Almasi, A. Gottlieb, “Highly Parallel Computing”, Benjamin-Cummings Publishing Co., Inc., USA, 1989.
[2] D. E. Rumelhart, G. E. Hinton, R. J. Williams, “Learning representations by back-propagating errors,” Nature, vol. 323, no. 6088, pp. 533–536, 1986.
[3] C. Li, C. Yu, “Performance evaluation of public non-profit hospitals using a BP artificial neural network: The case of Hubei province in china,” International Journal of Environmental Research and Public Health, Aug 2013.
[4] Y. Li, Y. Fu, H. Li, and S. W. Zhang, “The improved training algorithm of back propagation neural network with self-adaptive learning rate,” in 2009 International Conference on Computational Intelligence and Natural Computing, pp. 73–76, 2009.
[5] J. Zhu, P. Sutton, “FPGA implementations of neural networks–a survey of a decade of progress,” Field Programmable Logic and Application, pp. 1062-1066, 2003.
[6] I. B. D. Steinkraus, P.Y. Simard, “Using GPUs for machine learning algorithms,” Document Analysis and Recognition, pp. 1115-1120, 2009.
[7] C. M. Bishop, “Pattern Recognition and Machine Learning”, Springer-Verlag New York, USA, 2006.
[8] Y. Yuan, “Step-sizes for the gradient method,” AMS/IP Studies in Advanced Mathematics, 1999.
[9] R. A. Jacobs, “Increased rate of convergence through learning rate adaptation,” Neural Networks, vol. 1, 1988.
[10] M. J. Flynn, “Some computer organizations and their effectiveness,” IEEE Transactions on Computers, vol. C-21, no. 9, pp. 948–960, 1972.
[11] E. Kussul, T. Baidyk, “Improved method of handwritten digit recognition tested on MNIST database,” Image and Vision Computing, vol. 22, no. 12, pp. 971 – 981, 2004.
[12] J. D. Owens, M. Houston, D. Luebke, S. Green, J. E. Stone, J. C.Phillips, “GPU computing,” Proceedings of the IEEE, vol. 96, no. 5, pp. 879–899, May 2008.
[13] U. Ray, T.K. Hazra, U.K. Ray, "Matrix Multiplication using Strassen’s Algorithm on CPU & GPU", International Journal of Computer Sciences and Engineering, Vol.4, Issue.10, pp.98-105, 2016.