Page 316 - Artificial Intelligence in the Age of Neural Networks and Brain Computing
P. 316

References    309




                  are superior to those that can be developed by hand today; it is also likely to make it
                  possible to apply deep learning to a wider array of tasks and applications in the
                  future.



                  7. CONCLUSION
                  Evolutionary optimization makes it possible to construct more complex deep
                  learning architectures than can be done by hand. The topology, components, and
                  hyperparameters of the architecture can all be optimized simultaneously to fit the
                  requirements of the task, resulting in superior performance. This automated design
                  can make new applications of deep learning possible in vision, speech, language, and
                  other areas. Currently such designs are comparable with best human designs; with
                  anticipated increases in computing power, they should soon surpass them, putting
                  the power to good use.



                  REFERENCES
                   [1] Y. LeCun, L. Bottou, Y. Bengio, P. Haffner, Gradient-based learning applied to docu-
                      ment recognition, Proceedings of the IEEE 86 (1998) 22782324.
                   [2] S. Hochreiter, J. Schmidhuber, Long short-term memory, Neural Computation 9 (1997)
                      1735e1780.
                   [3] R. Collobert, J. Weston, A unified architecture for natural language processing: deep
                      neural networks with multitask learning, in: Proceedings of the 25th International
                      Conference on Machine Learning, ACM, 2008, pp. 160e167.
                   [4] A. Graves, A. Mohamed, G. Hinton, Speech recognition with deep recurrent neural
                      networks, in: 2013 IEEE International Conference on Acoustics, Speech and Signal
                      Processing, IEEE, 2013, pp. 6645e6649.
                   [5] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, Z. Wojna, Rethinking the inception
                      architecture for computer vision, in: Proceedings of Computer Vision and Pattern
                      Recognition Conference, 2016, pp. 2818e2826.
                   [6] Z. Che, S. Purushotham, K. Cho, D. Sontag, Y. Liu, Recurrent Neural Networks for
                      Multivariate Time Series with Missing Values, 2016. ArXiv 1606.01865.
                   [7] K. He, X. Zhang, S. Ren, J. Sun, Identity Mappings in Deep Residual Networks, 2016.
                      ArXiv 1603.05027.
                   [8] J.Y.-H. Ng, M.J. Hausknecht, S. Vijayanarasimhan, O. Vinyals, R. Monga, G. Toderici,
                      Beyond Short Snippets: Deep Networks for Video Classification, 2015. ArXiv
                      1503.08909.
                   [9] H. Hoos, Programming by optimization, Communications of the ACM 55 (2012) 70e80.
                  [10] C. Le Goues, T. Nguyen, S. Forrest, W. Weimer, Genprog: automatic bug correction in
                      real programs, ACM Transactions on Software Engineering 38 (2012).
                  [11] R. Miikkulainen, N. Iscoe, A. Shagrin, R. Rapp, S. Nazari, P. McGrath, C. Schoolland,
                      E. Achkar, M. Brundage, J. Miller, J. Epstein, G. Lamba, Sentient Ascend: AI-based
                      massively multivariate conversion rate optimization, in: Proceedings of Innovative
                      Applications of Artificial Intelligence, 2018.
   311   312   313   314   315   316   317   318   319   320   321