|
[1] C. Argueta, E. Saravia, and Y.S. Chen. Unsupervised graph-based patterns extraction for emotion classification. InProceedings of the 2015 IEEE/ACM international conference on Advances in Social Networks Analysis and Mining 2015. ACM, 2015. [2] M. Asgari, G. Kiss, J.P.H. van Santen, I. Shafran, and X. Song. Automatic measurement of affective valence and arousal in speech. InICASSP, 2014. [3] R.G. Bachu, S. Kopparthi, B. Adapa, and B.D. Barkana. Separation of voiced and un-voiced using zero crossing rate and energy of the speech signal. InAmerican Societyfor Engineering Education (ASEE) Zone Conference Proceedings, 2008. [4] Y.L. Boureau, J. Ponce, and Y. LeCun. A theoretical analysis of feature pooling invisual recognition. InProceedings of the 27th international conference on machinelearning (ICML-10), 2010. [5] C. Busso, M. Bulut, C.C. Lee, A. Kazemzadeh, E. Mower, S. Kim, J.N. Chang, S. Lee,and S.S. Narayanan. Iemocap: Interactive emotional dyadic motion capture database.Language resources and evaluation, 42(4), 2008. [6] F. Chollet et al. Keras, 2015. [7] F. Gouyon, F. Pachet, O. Delerue, et al. On the use of zero-crossing rate for an application of classification of percussive sounds. In Proceedings of the COST G-6 conference on Digital Audio Effects (DAFX-00), Verona, Italy, 2000. [8] J. Grekow. Audio features dedicated to the detection of arousal and valence in music recordings. In INnovations in Intelligent SysTems and Applications (INISTA), 2017 IEEE International Conference on. IEEE, 2017. [9] K. Han, D. Yu, and I. Tashev. Speech emotion recognition using deep neural network and extreme learning machine. In Fifteenth annual conference of the international speech communication association, 2014. [10] K.M.Han,T.Zin,andH.M.Tun.Extractionofaudiofeaturesforemotionrecognition system based on music. International Journal of Scientific & Technology Research, 4(8), 2015. [11] S. Haq, P.J.B. Jackson, and J. Edge. Speaker-dependent audio-visual emotion recog- nition. In AVSP, pages 53–58, 2009. [12] M.R. Hasan, M. Jamil, M. Rahman, et al. Speaker identification using mel frequency cepstral coefficients. variations, 1(4), 2004. [13] X.D. Huang, F. Alleva, H.W. Hon, M.Y. Hwang, K.F. Lee, and R. Rosenfeld. The sphinx-ii speech recognition system: an overview. Computer Speech & Language, 7(2), 1993. [14] Q. Jin, C.X. Li, S.Z. Chen, and H.M. Wu. Speech emotion recognition with acoustic and lexical features. In Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on. IEEE, 2015. [15] Y. Kim. Convolutional neural networks for sentence classification. arXiv preprint arXiv:1408.5882, 2014. [16] E. Lakomkin, M.A. Zamani, C. Weber, S. Magg, and S. Wermter. Emorl: Continuous acoustic emotion classification using deep reinforcement learning. arXiv preprint arXiv:1804.04053, 2018. [17] J. Lee and I. Tashev. High-level feature representation using recurrent neural network for speech emotion recognition. 2015. [18] S.R. Livingstone and F.A. Russo. The ryerson audio-visual database of emotional speech and song (ravdess): A dynamic, multimodal set of facial and vocal expressions in north american english. PloS one, 13(5), 2018. [19] B.McFee,C.Raffel,D.Liang,D.P.W.Ellis,M.McVicar,E.Battenberg,andO.Nieto. librosa: Audio and music signal analysis in python. In Proceedings of the 14th python in science conference, 2015. [20] A. Metallinou, S. Lee, and S. Narayanan. Decision level combination of multiple modalities for recognition and analysis of emotional expression. In Acoustics Speech and Signal Processing (ICASSP), 2010 IEEE International Conference on. IEEE, 2010. [21] T. Mikolov, K. Chen, G. Corrado, and J. Dean. Efficient estimation of word represen- tations in vector space. arXiv preprint arXiv:1301.3781, 2013. [22] V. Nair and G.E. Hinton. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th international conference on machine learning (ICML-10), 2010. [23] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blon- del, P. Prettenhofer, R. Weiss, V. Dubourg, et al. Scikit-learn: Machine learning in python. Journal of machine learning research, 12(Oct), 2011. [24] J.W. Pennebaker, R.J. Booth, and M.E. Francis. Linguistic inquiry and word count: Liwc [computer software]. Austin, TX: liwc. net, 2007. [25] J. Pennington, R. Socher, and C. Manning. Glove: Global vectors for word repre- sentation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), 2014. [26] E. Riloff and J. Wiebe. Learning extraction patterns for subjective expressions. In Proceedings of the 2003 conference on Empirical methods in natural language pro- cessing. Association for Computational Linguistics, 2003. [27] V. Rozgic, S. Ananthakrishnan, S. Saleem, R. Kumar, and R. Prasad. Ensemble of svm trees for multimodal emotion recognition. In Signal & Information Processing Association Annual Summit and Conference (APSIPA ASC), 2012 Asia-Pacific. IEEE, 2012. [28] E. Saravia, C.H. Chang, R.J. De Lorenzo, and Y.S. Chen. Midas: Mental illness detection and analysis via social media. In Advances in Social Networks Analysis and Mining (ASONAM), 2016 IEEE/ACM International Conference on. IEEE, 2016. [29] E. Saravia, H.C.T. Liu, and Y.S. Chen. Deepemo: Learning and enriching pattern- based emotion representations. arXiv preprint arXiv:1804.08847, 2018. [30] V. Sintsova, C.C. Musat, and P. Pu. Fine-grained emotion recognition in olympic tweets based on human computation. In 4th Workshop on computational approaches to subjectivity, sentiment and social media analysis, number EPFL-CONF-197185, 2013. [31] N.Srivastava,G.Hinton,A.Krizhevsky,I.Sutskever, and R.Salakhutdinov.Dropout: a simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research, 15(1), 2014. [32] C. Strapparava and R. Mihalcea. Semeval-2007 task 14: Affective text. In Proceedings of the 4th international workshop on semantic evaluations. Association for Computational Linguistics, 2007. [33] N. Yang, R. Muraleedharan, J. Kohl, I. Demirkol, W. Heinzelman, and M. Sturge- Apple. Speech-based emotion classification using multiclass svm with hybrid kernel and thresholding fusion. In Spoken Language Technology Workshop (SLT), 2012 IEEE. IEEE, 2012. [34] A. Yessenalina, Y. Yue, and C. Cardie. Multi-level structured models for document-level sentiment classification. In Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2010. [35] M.D. Zeiler. Adadelta: an adaptive learning rate method. arXiv preprint arXiv:1212.5701, 2012. [36] Y.ZhangandB.Wallace.Asensitivityanalysisof(and practitioners’ guide to) convolutional neural networks for sentence classification. arXiv preprint arXiv:1510.03820, 2015. [37] H. Zhou, M. Huang, T.Y. Zhang, X.Y. Zhu, and B. Liu. Emotional chatting machine: Emotional conversation generation with internal and external memory. arXiv preprint arXiv:1704.01074, 2017. |