Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks
K. Devkota1 , P. Bhattarai2
- Department of Electronics and Computer Engineering, Institute of Engineering (Tribhuvan University), Kathmandu, Nepal.
- Department of Electronics and Computer Engineering, Institute of Engineering (Tribhuvan University), Kathmandu, Nepal.
Correspondence should be addressed to: kdkapildevkota@gmail.com.
Section:Research Paper, Product Type: Journal Paper
Volume-5 ,
Issue-10 , Page no. 266-272, Oct-2017
CrossRef-DOI: https://doi.org/10.26438/ijcse/v5i10.266272
Online published on Oct 30, 2017
Copyright © K. Devkota, P. Bhattarai . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
View this paper at Google Scholar | DPI Digital Library
How to Cite this Paper
- IEEE Citation
- MLA Citation
- APA Citation
- BibTex Citation
- RIS Citation
IEEE Citation
IEEE Style Citation: K. Devkota, P. Bhattarai, “Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks,” International Journal of Computer Sciences and Engineering, Vol.5, Issue.10, pp.266-272, 2017.
MLA Citation
MLA Style Citation: K. Devkota, P. Bhattarai "Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks." International Journal of Computer Sciences and Engineering 5.10 (2017): 266-272.
APA Citation
APA Style Citation: K. Devkota, P. Bhattarai, (2017). Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks. International Journal of Computer Sciences and Engineering, 5(10), 266-272.
BibTex Citation
BibTex Style Citation:
@article{Devkota_2017,
author = {K. Devkota, P. Bhattarai},
title = {Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks},
journal = {International Journal of Computer Sciences and Engineering},
issue_date = {10 2017},
volume = {5},
Issue = {10},
month = {10},
year = {2017},
issn = {2347-2693},
pages = {266-272},
url = {https://www.ijcseonline.org/full_paper_view.php?paper_id=1511},
doi = {https://doi.org/10.26438/ijcse/v5i10.266272}
publisher = {IJCSE, Indore, INDIA},
}
RIS Citation
RIS Style Citation:
TY - JOUR
DO = {https://doi.org/10.26438/ijcse/v5i10.266272}
UR - https://www.ijcseonline.org/full_paper_view.php?paper_id=1511
TI - Parallel Implementation of Gradient Descent Algorithm for Backpropagation Networks
T2 - International Journal of Computer Sciences and Engineering
AU - K. Devkota, P. Bhattarai
PY - 2017
DA - 2017/10/30
PB - IJCSE, Indore, INDIA
SP - 266-272
IS - 10
VL - 5
SN - 2347-2693
ER -
![]() |
![]() |
![]() |
577 | 255 downloads | 200 downloads |




Abstract
The problem of computational efficiency in adaptive algorithms, which is current and pressing, can be solved through their implementation in parallel frameworks, like CUDA, OpenCL, etc. The approach taken to parallelize any complex operation requires its separation into several distinct and independent sub-operations. We employed the same procedure to parallelize the BP (or Backpropagation) network algorithm. The function breakdown of the BP network involved breaking its overall operation into Feed-forward and Back-Propagate sub-operations, which was further divided into smaller independent execution groups. We applied parallel constructs on those independent execution groups and used the MNIST dataset to compare the algorithm’s performance with respect to the sequential algorithm. Comparing their performances, we found that the efficiency of the algorithm depended on the size of the BP network. In the large network with massive number of weight connections, we saw a significant improvement in the convergence time. This makes our algorithm preferable in feedforward networks having large number of hidden layers, neurons and weight connections.
Key-Words / Index Term
Backpropagation, Supervised Learning, CUDA, parallel
References
[1] G. S. Almasi, A. Gottlieb, “Highly Parallel Computing”, Benjamin-Cummings Publishing Co., Inc., USA, 1989.
[2] D. E. Rumelhart, G. E. Hinton, R. J. Williams, “Learning representations by back-propagating errors,” Nature, vol. 323, no. 6088, pp. 533–536, 1986.
[3] C. Li, C. Yu, “Performance evaluation of public non-profit hospitals using a BP artificial neural network: The case of Hubei province in china,” International Journal of Environmental Research and Public Health, Aug 2013.
[4] Y. Li, Y. Fu, H. Li, and S. W. Zhang, “The improved training algorithm of back propagation neural network with self-adaptive learning rate,” in 2009 International Conference on Computational Intelligence and Natural Computing, pp. 73–76, 2009.
[5] J. Zhu, P. Sutton, “FPGA implementations of neural networks–a survey of a decade of progress,” Field Programmable Logic and Application, pp. 1062-1066, 2003.
[6] I. B. D. Steinkraus, P.Y. Simard, “Using GPUs for machine learning algorithms,” Document Analysis and Recognition, pp. 1115-1120, 2009.
[7] C. M. Bishop, “Pattern Recognition and Machine Learning”, Springer-Verlag New York, USA, 2006.
[8] Y. Yuan, “Step-sizes for the gradient method,” AMS/IP Studies in Advanced Mathematics, 1999.
[9] R. A. Jacobs, “Increased rate of convergence through learning rate adaptation,” Neural Networks, vol. 1, 1988.
[10] M. J. Flynn, “Some computer organizations and their effectiveness,” IEEE Transactions on Computers, vol. C-21, no. 9, pp. 948–960, 1972.
[11] E. Kussul, T. Baidyk, “Improved method of handwritten digit recognition tested on MNIST database,” Image and Vision Computing, vol. 22, no. 12, pp. 971 – 981, 2004.
[12] J. D. Owens, M. Houston, D. Luebke, S. Green, J. E. Stone, J. C.Phillips, “GPU computing,” Proceedings of the IEEE, vol. 96, no. 5, pp. 879–899, May 2008.
[13] U. Ray, T.K. Hazra, U.K. Ray, "Matrix Multiplication using Strassen’s Algorithm on CPU & GPU", International Journal of Computer Sciences and Engineering, Vol.4, Issue.10, pp.98-105, 2016.