|
[1] Narayanan, Shrikanth, and Panayiotis G. Georgiou. “Behavioral signal processing: Deriving human behavioral informatics from speech and language.” Proceedings of the IEEE, vol.101, no.5, pp.1203-1233, 2013. [2] Chen, Wei-Chen, et al. “Multimodal arousal rating using unsupervised fusion technique.” 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, pp.5296-5300, 2015. [3] Delaherche, Emilie, et al. “Assessment of the communicative and coordination skills of children with autism spectrum disorders and typically developing children using social signal processing.” Research in Autism Spectrum Disorders, vol.7, no.6, pp.741-756, 2013. [4] Hsiao, Shan-Wen, et al. “A Multimodal Approach for Automatic Assessment of School Principals' Oral Presentation During Pre-Service Training Program.” Sixteenth Annual Conference of the International Speech Communication Association. 2015. [5] Black, Matthew P., et al. “Toward automating a human behavioral coding system for married couples’ interactions using speech acoustic features.” Speech Communication, vol.55, no.1, pp.1-21, 2013. [6] P. Association, “Diagnostic and statistical manual of mental disorders (4th ed., text rev.).” American Psychiatric Publishing, Inc., 2000. [7] Botturi, Luca, Chiara Bramani, and Sara Corbino. “Digital storytelling for social and international development: from special education to vulnerable children.” International Journal of Arts and Technology, vol.7, no.1, pp.92-111, 2014. [8] Even, Cindy, et al. “Supporting Social Skills Rehabilitation with Virtual Storytelling.” Twenty-Ninth International Florida Artificial Intelligence Research Society Conference. AAAI publications, pp.329-334, 2016. [9] Ring, Lazlo, et al. “Addressing loneliness and isolation in older adults: Proactive affective agents provide better support.” Affective Computing and Intelligent Interaction (ACII), 2013 Humaine Association Conference on. IEEE, pp.61-66, 2013. [10] Tartaro, Andrea, and Justine Cassell. “Playing with virtual peers: bootstrapping contingent discourse in children with autism.” Proceedings of the 8th international conference on International conference for the learning sciences-Volume 2. International Society of the Learning Sciences, pp.382-389, 2008. [11] Anderson, Keith, et al. “The TARDIS framework: intelligent virtual agents for social coaching in job interviews.” Advances in computer entertainment. Springer International Publishing, pp.476-491, 2013. [12] Tartaro, Andrea, and Justine Cassell. “Using virtual peer technology as an intervention for children with autism.” Towards universal usability: designing computer interfaces for diverse user populations. Chichester: John Wiley, vol.231, pp.62, 2007. [13] S. S. Narayanan and A. Potamianos, “Creating conversational interfaces for children,” IEEE Transactions on Speech and Audio Processing, vol.10, no. 2, pp. 65–78, 2002. [14] Mower, Emily, et al. “Rachel: Design of an emotionally targeted interactive agent for children with autism.” Multimedia and Expo (ICME), 2011 IEEE International Conference on. IEEE, pp.1-6, 2011. [15] Lord, Catherine, et al. “The Autism Diagnostic Observation Schedule—Generic: A standard measure of social and communication deficits associated with the spectrum of autism.” Journal of autism and developmental disorders, vol.30, no.3, pp.205-223, 2000. [16] Lord, Catherine, et al. “Austism diagnostic observation schedule: A standardized observation of communicative and social behavior.” Journal of autism and developmental disorders, vol.19, no.2, pp.185-212, 1989. [17] Akshoomoff, Natacha, Christina Corsello, and Heather Schmidt. “The role of the autism diagnostic observation schedule in the assessment of autism spectrum disorders in school and community settings.” The California School Psychologist, vol.11, no.1, pp.7-19, 2006. [18] Baucom, Brian R., and Esti Iturralde. “A behaviorist manifesto for the 21 st century,” Signal & Information Processing Association Annual Summit and Conference (APSIPA ASC), 2012 Asia-Pacific. IEEE, pp.1-4, 2012. [19] Lord, Catherine, Michael Rutter, and Ann Le Couteur. “Autism Diagnostic Interview-Revised: a revised version of a diagnostic interview for caregivers of individuals with possible pervasive developmental disorders.” Journal of autism and developmental disorders, vol.24, no.5, pp.659-685, 1994. [20] Bailly, Gérard, Stephan Raidt, and Frédéric Elisei. “Gaze, conversational agents and face-to-face communication.” Speech Communication, vol.52, no.6, pp.598-612, 2010. [21] Bal, Elgiz, et al. “Emotion recognition in children with autism spectrum disorders: Relations to eye gaze and autonomic state.” Journal of autism and developmental disorders, vol.40, no.3, pp.358-370, 2010. [22] de Marchena, Ashley, and Inge‐Marie Eigsti. “Conversational gestures in autism spectrum disorders: Asynchrony but not decreased frequency.” Autism research, vol.3, no.6, pp.311-322, 2010. [23] Bellard, Fabrice, et.al. “FFmpeg.” Availabel from: https://ffmpeg.org/. [24] Boersma, Paul. “Praat, a system for doing phonetics by computer.” Glot international, vol.5, no.9/10, pp.341-345, 2002. [25] McFee, Brian, et al. “librosa: Audio and music signal analysis in python.” Proceedings of the 14th Python in Science Conference. 2015. [26] Davis, Steven, and Paul Mermelstein. “Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences.” IEEE transactions on acoustics, speech, and signal processing, vol.28, no.4, pp. 357-366, 1980. [27] Boersma, Paul. “Accurate short-term analysis of the fundamental frequency and the harmonics-to-noise ratio of a sampled sound.” Proceedings of the institute of phonetic sciences. vol.17, no.1193, pp.97-110, 1993. [28] Wang, Heng, et al. “Action recognition by dense trajectories.” Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on. IEEE, pp.3169-3176, 2011. [29] Wang, Heng, and Cordelia Schmid. “Action recognition with improved trajectories.” Proceedings of the IEEE International Conference on Computer Vision., pp.3551-3558, 2013. [30] Baraldi, Lorenzo, et al. “Gesture recognition in ego-centric videos using dense trajectories and hand segmentation.” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops., pp.688-693, 2014. [31] Baltru, Tadas, Peter Robinson, and Louis-Philippe Morency. “OpenFace: an open source facial behavior analysis toolkit.” 2016 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE, pp.1-10, 2016. [32] Baltrusaitis, Tadas, Peter Robinson, and Louis-Philippe Morency. “Constrained local neural fields for robust facial landmark detection in the wild.” Proceedings of the IEEE International Conference on Computer Vision Workshops., pp.354-361, 2013. [33] Wood, Erroll, et al. “Rendering of eyes for eye-shape registration and gaze estimation.” 2015 IEEE International Conference on Computer Vision (ICCV). IEEE, pp.3756-3764, 2015. [34] Baltrušaitis, Tadas, Marwa Mahmoud, and Peter Robinson. “Cross-dataset learning and person-specific normalisation for automatic Action Unit detection.” Automatic Face and Gesture Recognition (FG), 2015 11th IEEE International Conference and Workshops on, IEEE, vol.6, pp.1-6, 2015. [35] Matthews, Iain, and Simon Baker. “Active appearance models revisited.” International Journal of Computer Vision, vol.60, no.2, pp.135-164, 2004. [36] Cambria, Erik, and Bebo White. “Jumping NLP curves: a review of natural language processing research [review article].” IEEE Computational Intelligence Magazine, vol.9, no.2, pp.48-57, 2014. [37] Justice, Laura M., et al. “A scalable tool for assessing children's language abilities within a narrative context: The NAP (Narrative Assessment Protocol).” Early Childhood Research Quarterly, vol.25, no.2, pp.218-234, 2010. [38] Loria, Steven. “TextBlob: simplified text processing.” Secondary TextBlob: Simplified Text Processing (2014). [39] Csurka, Gabriella, et al. “Visual categorization with bags of keypoints.” Workshop on statistical learning in computer vision, ECCV, vol.1, no.1-22, 2004. [40] Sivic, Josef, and Andrew Zisserman. “Efficient visual search of videos cast as text retrieval.” IEEE transactions on pattern analysis and machine intelligence, vol.31, no.4, pp. 591-606, 2009 [41] Perronnin, Florent, and Christopher Dance. “Fisher kernels on visual vocabularies for image categorization.” 2007 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, pp.1-8, 2007. [42] Perronnin, Florent, Jorge Sánchez, and Thomas Mensink. “Improving the fisher kernel for large-scale image classification.” European conference on computer vision. Springer Berlin Heidelberg, pp.143-156, 2010. [43] Jégou, Hervé, et al. “Aggregating local descriptors into a compact image representation.” Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on. IEEE, pp.3304-3311, 2010. [44] Bone, Daniel, et al. “Intoxicated Speech Detection by Fusion of Speaker Normalized Hierarchical Features and GMM Supervectors.” INTERSPEECH, pp.3217-3220, 2011. [45] Csurka, Gabriela, and Florent Perronnin. “Fisher vectors: Beyond bag-of-visual-words image representations.” International Conference on Computer Vision, Imaging and Computer Graphics. Springer Berlin Heidelberg, 2010. [46] Sánchez, Jorge, et al. “Image classification with the fisher vector: Theory and practice.” International journal of computer vision, vol.105, no.3, pp.222-245, 2013. [47] Platt, John. “Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods.” Advances in large margin classifiers, vol.10, no.3, pp.61-74, 1999. [48] Schuller, Björn, Stefan Steidl, and Anton Batliner. “The INTERSPEECH 2009 emotion challenge.” INTERSPEECH, vol. 2009, pp.312-315, 2009.
|