參考文獻 |
[1] Andrieu, C., De Freitas, N., Doucet, A., & Jordan, M. I. (2003). An introduction to MCMC for machine learning. Machine learning, 50(1), 5-43.
[2] Bengio, Y. (2009). Learning deep architectures for AI. Now Publishers Inc.
[3] Bengio, Y., Courville, A., & Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8), 1798-1828.
[4] Bishop, C. M. (2006). Pattern recognition. Machine learning, 128(9).
[5] Blei, D. M., Kucukelbir, A., & McAuliffe, J. D. (2017). Variational inference: A review for statisticians. Journal of the American statistical Association, 112(518), 859-877.
[6] Blundell, C., Cornebise, J., Kavukcuoglu, K., & Wierstra, D. (2015). Weight uncertainty in neural network. International Conference on Machine Learning.
[7] Elman, J. L. (1990). Finding structure in time. Cognitive science, 14(2), 179-211.
[8] Gal, Y., & Ghahramani, Z. (2015). Dropout as a Bayesian approximation: appendix. arXiv preprint arXiv:1506.02157, 420.
[9] Gal, Y., & Ghahramani, Z. (2016). Dropout as a bayesian approximation: Representing model uncertainty in deep learning. international conference on machine learning.
[10] Gers, F. A., Schmidhuber, J., & Cummins, F. (2000). Learning to forget: Continual prediction with LSTM. Neural computation, 12(10), 2451-2471.
[11] Girshick, R. (2015). Fast r-cnn. Proceedings of the IEEE international conference on computer vision.
[12] Goodfellow, I., Bengio, Y., Courville, A., & Bengio, Y. (2016). Deep learning (Vol. 1). MIT press Cambridge.
32
[13] Greff, K., Srivastava, R. K., Koutník, J., Steunebrink, B. R., & Schmidhuber, J. (2016). LSTM: A search space odyssey. IEEE transactions on neural networks and learning systems, 28(10), 2222-2232.
[14] Hahnloser, R. H., Seung, H. S., & Slotine, J.-J. (2003). Permitted and forbidden sets in symmetric threshold-linear networks. Neural computation, 15(3), 621-638.
[15] Hinton, G. E. (1987). Learning translation invariant recognition in a massively parallel networks. International Conference on Parallel Architectures and Languages Europe.
[16] Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural computation, 9(8), 1735-1780.
[17] Ioffe, S., & Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift. International conference on machine learning.
[18] Kawakami, K. (2008). Supervised sequence labelling with recurrent neural networks. Ph. D. thesis.
[19] Kingma, D. P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
[20] LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. nature, 521(7553), 436-444.
[21] Lipton, Z. C., Berkowitz, J., & Elkan, C. (2015). A critical review of recurrent neural networks for sequence learning. arXiv preprint arXiv:1506.00019.
[22] Logsdon, B. A., Hoffman, G. E., & Mezey, J. G. (2010). A variational Bayes algorithm for fast and accurate multiple locus genome-wide association analysis. BMC bioinformatics, 11(1), 1-13.
[23] Rojas, R. (2013). Neural networks: a systematic introduction. Springer Science & Business Media.
[24] Ruder, S. (2016). An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747.
33
[25] Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R. (2014). Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research, 15(1), 1929-1958.
[26] Szepesvári, C. (2010). Algorithms for reinforcement learning. Synthesis lectures on artificial intelligence and machine learning, 4(1), 1-103.
[27] Tsuda, K., Kin, T., & Asai, K. (2002). Marginalized kernels for biological sequences. Bioinformatics, 18(suppl_1), S268-S275.
[28] Yuan, X., Li, L., & Wang, Y. (2019). Nonlinear dynamic soft sensor modeling with supervised long short-term memory network. IEEE transactions on industrial informatics, 16(5), 3168-3176.
[29] Zhang, N., Lei, D., & Zhao, J. (2018). An improved Adagrad gradient descent optimization algorithm. 2018 Chinese Automation Congress (CAC). |