Emotion-sensing computer software that models and responds to students' cognitive1 and emotional states -- including frustration2 and boredom3 -- has been developed by University of Notre Dame4 Assistant Professor of Psychology5 Sidney D'Mello, Art Graesser from the University of Memphis and a colleague from Massachusetts Institute of Technology. D'Mello also is a concurrent6 assistant professor of computer science and engineering. The new technology, which matches the interaction of human tutors, not only offers tremendous learning possibilities for students, but also redefines human-computer interaction.
"AutoTutor" and "Affective AutoTutor" can gauge7(测量,估计) the student's level of knowledge by asking probing questions; analyzing8 the student's responses to those questions; proactively(前摄地) identifying and correcting misconceptions; responding to the student's own questions, gripes and comments; and even sensing a student's frustration or boredom through facial expression and body posture9 and dynamically changing its strategies to help the student conquer those negative emotions.
"Most of the 20th-century systems required humans to communicate with computers through windows, icons10, menus and pointing devices," says D'Mello, who specializes in human-computer interaction and artificial intelligence in education.
"But humans have always communicated with each other through speech and a host of nonverbal cues such as facial expressions, eye contact, posture and gesture. In addition to enhancing the content of the message, the new technology provides information regarding the cognitive states, motivation levels and social dynamics11 of the students."
AutoTutor is an Intelligent Tutoring System (ITS) that helps students learn complex technical content in Newtonian physics, computer literacy and critical thinking by holding a conversation in natural language; simulating teaching and motivational strategies of human tutors; modeling students' cognitive states; using its student model to dynamically tailor the interaction to individual students; answering students' questions; identifying and correcting misconceptions; and keeping students engaged with images, animations12 and simulations. In addition to these capabilities13, Affective AutoTutor adds emotion-sensitive capabilities by monitoring facial features, body language and conversational14 cues; regulating negative states such as frustration and boredom; and synthesizing emotions via the content of its verbal responses, speech intonation15(声调) and facial expressions of an animated16 teacher.
D'Mello's study, titled "AutoTutor and Affective AutoTutor: Learning by Talking with Cognitively17 and Emotionally Intelligent Computers that Talk Back," that details this new technology will be published in special edition of ACM Transactions on Interactive18 Intelligent Systems that highlights innovative19 technology of the last decade.
"Much like a gifted human tutor, AutoTutor and Affective AutoTutor attempt to keep the student balanced between the extremes of boredom and bewilderment by subtly modulating20 the pace, direction and complexity21 of the learning task," D'Mello says.
Considerable empirical evidence has shown that one-on-one human tutoring is extremely effective when compared to typical classroom environments, and AutoTutor and Affective AutoTutor closely model the pedagogical(教育学的) styles, dialogue patterns, language and gestures of human tutors. They are also one of the few ITSs that help learning by engaging students in natural language dialogues that closely mirror human-human tutorial dialogues.
Tested on more than 1,000 students, AutoTutor produces learning gains of approximately one letter grade -- gains that have proven to outperform novice22 human tutors and almost reach the bar of expert human tutors.