@conference {60, title = {Automatic cry detection in early childhood education settings}, booktitle = {7th IEEE International Conference on Development and Learning, 2008. ICDL 2008}, year = {2008}, month = {08/2008}, publisher = {IEEE}, organization = {IEEE}, address = {Monterey, CA}, abstract = {
We present results on applying a novel machine learning approach for learning auditory moods in natural environments [1] to the problem of detecting crying episodes in preschool classrooms. The resulting system achieved levels of performance approaching that of human coders and also significantly outperformed previous approaches to this problem [2].
}, keywords = {Acoustic noise, auditory moods, automatic cry detection, behavioural sciences computing, Deafness, early childhood education settings, education, Educational robots, Emotion recognition, human coders, Humans, learning (artificial intelligence), Machine Learning, Mood, preschool classrooms, Prototypes, Robustness, Working environment noise}, isbn = {978-1-4244-2661-4}, author = {Ruvolo, P. and Movellan, J.} } @conference {62, title = {A discriminative approach to frame-by-frame head pose tracking}, booktitle = {8th IEEE International Conference on Automatic Face Gesture Recognition, 2008. FG {\textquoteright}08}, year = {2008}, month = {09/2008}, publisher = {IEEE}, organization = {IEEE}, address = {Amsterdam}, abstract = {We present a discriminative approach to frame-by-frame head pose tracking that is robust to a wide range of illuminations and facial appearances and that is inherently immune to accuracy drift. Most previous research on head pose tracking has been validated on test datasets spanning only a small (\< 20) subjects under controlled illumination conditions on continuous video sequences. In contrast, the system presented in this paper was both trained and tested on a much larger database, GENKI, spanning tens of thousands of different subjects, illuminations, and geographical locations from images on the Web. Our pose estimator achieves accuracy of 5.82deg, 5.65deg, and 2.96deg root-mean-square (RMS) error for yaw, pitch, and roll, respectively. A set of 4000 images from this dataset, labeled for pose, was collected and released for use by the research community.
}, keywords = {accuracy drift, continuous video sequence, controlled illumination condition, discriminative approach, face detection, face recognition, facial appearance, frame-by-frame head pose tracking, Humans, Image analysis, Image databases, Laboratories, Lighting, Magnetic heads, mean square error methods, pose estimation, Robustness, root-mean-square error tracking, System testing, Video sequences}, isbn = {978-1-4244-2153-4}, author = {Whitehill, J. and Movellan, Javier R.} }