|
[1] A. Krizhevsky, I. Sutskever, G.E. Hinton, Imagenet classification with deep convolutional neural networks, in: Advances in neural information processing systems, 2012, pp. 1097-1105. [2] Z. Zhong, L. Jin, Z. Xie, High performance offline handwritten chinese character recognition using googlenet and directional feature maps, in: Document Analysis and Recognition (ICDAR), 2015 13th International Conference on, IEEE, 2015, pp. 846-850. [3] L. Wang, S. Guo, W. Huang, Y. Qiao, Places205-vggnet models for scene recognition, arXiv preprint arXiv:1508.01667, (2015). [4] Y. LeCun, L. Bottou, Y. Bengio, P. Haffner, Gradient-based learning applied to document recognition, Proceedings of the IEEE, 86 (1998) 2278-2324. [5] H.M. Albeahdili, T. Han, N.E. Islam, Hybrid Algorithm for the Optimization of Training Convolutional Neural Network. [6] A. Toshev, C. Szegedy, Deeppose: Human pose estimation via deep neural networks, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 1653-1660. [7] R.A. Jacobs, Increased rates of convergence through learning rate adaptation, Neural networks, 1 (1988) 295-307. [8] A. Van Ooyen, B. Nienhuis, Improving the convergence of the back-propagation algorithm, Neural Networks, 5 (1992) 465-471. [9] V. Nair, G.E. Hinton, Rectified linear units improve restricted boltzmann machines, in: Proceedings of the 27th international conference on machine learning (ICML-10), 2010, pp. 807-814. [10] T. Yamasaki, T. Honma, K. Aizawa, Efficient Optimization of Convolutional Neural Networks Using Particle Swarm Optimization, in: Multimedia Big Data (BigMM), 2017 IEEE Third International Conference on, IEEE, 2017, pp. 70-73. [11] P.R. Lorenzo, J. Nalepa, M. Kawulok, L.S. Ramos, J.R. Pastor, Particle swarm optimization for hyper-parameter selection in deep neural networks, in: Proceedings of the Genetic and Evolutionary Computation Conference, ACM, 2017, pp. 481-488. [12] D. Zang, J. Ding, J. Cheng, D. Zhang, K. Tang, A Hybrid Learning Algorithm for the Optimization of Convolutional Neural Network, in: International Conference on Intelligent Computing, Springer, 2017, pp. 694-705. [13] W.-C. Yeh, An improved simplified swarm optimization, Knowledge-Based Systems, 82 (2015) 60-69. [14] T. Hastie, R. Tibshirani, J. Friedman, Overview of supervised learning, in: The elements of statistical learning, Springer, 2009, pp. 9-41. [15] T.D. Sanger, Optimal unsupervised learning in a single-layer linear feedforward neural network, Neural networks, 2 (1989) 459-473. [16] R.S. Sutton, A.G. Barto, Reinforcement learning: An introduction, MIT press Cambridge, 1998. [17] Y. Freund, L. Mason, The alternating decision tree learning algorithm, in: icml, 1999, pp. 124-133. [18] D.T. Larose, K‐nearest neighbor algorithm, Discovering Knowledge in Data: An Introduction to Data Mining, (2005) 90-106. [19] J.A. Suykens, J. Vandewalle, Least squares support vector machine classifiers, Neural processing letters, 9 (1999) 293-300. [20] D. Silver, A. Huang, C.J. Maddison, A. Guez, L. Sifre, G. Van Den Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanctot, Mastering the game of Go with deep neural networks and tree search, Nature, 529 (2016) 484-489. [21] V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, M. Riedmiller, Playing atari with deep reinforcement learning, arXiv preprint arXiv:1312.5602, (2013). [22] C.J.C.H. Watkins, Learning from delayed rewards, in, King's College, Cambridge, 1989. [23] C.J. Watkins, Peter. Dayan, Technical note Q-learning, Machine Learn, 8 (1992) 279-292. [24] F. Rosenblatt, The perceptron: A probabilistic model for information storage and organization in the brain, Psychological review, 65 (1958) 386. [25] Y. Bengio, P. Simard, P. Frasconi, Learning long-term dependencies with gradient descent is difficult, IEEE transactions on neural networks, 5 (1994) 157-166. [26] Y. LeCun, B. Boser, J.S. Denker, D. Henderson, R.E. Howard, W. Hubbard, L.D. Jackel, Backpropagation applied to handwritten zip code recognition, Neural computation, 1 (1989) 541-551. [27] G.E. Nasr, E. Badr, C. Joun, Cross Entropy Error Function in Neural Networks: Forecasting Gasoline Demand, in: FLAIRS Conference, 2002, pp. 381-384. [28] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G.S. Corrado, A. Davis, J. Dean, M. Devin, Tensorflow: Large-scale machine learning on heterogeneous distributed systems, arXiv preprint arXiv:1603.04467, (2016). [29] 武妍, 王守覺, 一種新的快速收斂的反向傳播算法, 同濟大學學報 (自然科學版), 32 (2004) 1092-1095. [30] W. Yeh, Study on quickest path networks with dependent components and apply to RAP, Rep. NSC, (2008) 97-2221. [31] C.-L. Huang, A particle-based simplified swarm optimization algorithm for reliability redundancy allocation problems, Reliability Engineering & System Safety, 142 (2015) 221-230. [32] W.-C. Yeh, Novel swarm optimization for mining classification rules on thyroid gland data, Information Sciences, 197 (2012) 65-76. [33] W.-C. Yeh, Y.-M. Yeh, C.-H. Chou, Y.-Y. Chung, X. He, A radio frequency identification network design methodology for the decision problem in Mackay Memorial Hospital based on swarm optimization, in: Evolutionary Computation (CEC), 2012 IEEE Congress on, IEEE, 2012, pp. 1-7. [34] W.-C. Yeh, Simplified swarm optimization in disassembly sequencing problems with learning effects, Computers & Operations Research, 39 (2012) 2168-2177. [35] W. Gao, C. Song, J. Jiang, C. Zhang, Simplified Particle Swarm Optimization Algorithm Based on Improved Learning Factors, in: International Symposium on Neural Networks, Springer, 2017, pp. 321-328. [36] W.C. Yeh, Y.T. Yang, C.M. Lai, A Hybrid Simplified Swarm Optimization Method for Imbalanced Data Feature Selection, Australian Academy of Business and Economics Review, 2 (2017) 263-275. [37] W.-C. Yeh, Y.-M. Yeh, P.-C. Chang, Y.-C. Ke, V. Chung, Forecasting wind power in the Mai Liao Wind Farm based on the multi-layer perceptron artificial neural network model with improved simplified swarm optimization, International Journal of Electrical Power & Energy Systems, 55 (2014) 741-748. [38] X. Zhang, W.-c. Yeh, Y. Jiang, Y. Huang, Y. Xiao, L. Li, A Case Study of Control and Improved Simplified Swarm Optimization for Economic Dispatch of a Stand-Alone Modular Microgrid, Energies, 11 (2018) 793. [39] I. Sutskever, J. Martens, G. Dahl, G. Hinton, On the importance of initialization and momentum in deep learning, in: International conference on machine learning, 2013, pp. 1139-1147. [40] D.P. Kingma, J. Ba, Adam: A method for stochastic optimization, arXiv preprint arXiv:1412.6980, (2014). [41] M.D. Zeiler, ADADELTA: an adaptive learning rate method, arXiv preprint arXiv:1212.5701, (2012). [42] T. Tieleman, G. Hinton, Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude, COURSERA: Neural networks for machine learning, 4 (2012) 26-31. [43] A.R. Syulistyo, D.M.J. Purnomo, M.F. Rachmadi, A. Wibowo, Particle swarm optimization (PSO) for training optimization on convolutional neural network (CNN), Jurnal Ilmu Komputer dan Informasi, 9 (2016) 52-58. [44] D.M. Hawkins, The problem of overfitting, Journal of chemical information and computer sciences, 44 (2004) 1-12. [45] S. Lawrence, C.L. Giles, Overfitting and neural networks: conjugate gradient and backpropagation, in: Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on, IEEE, 2000, pp. 114-119. [46] X. Glorot, Y. Bengio, Understanding the difficulty of training deep feedforward neural networks, in: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, 2010, pp. 249-256.
|