TY - GEN
T1 - Animated conversation
T2 - 21st Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 1994
AU - Cassell, Justine
AU - Pelachaud, Catherine
AU - Badler, Norman
AU - Steedman, Mark
AU - Achorn, Brett
AU - Becket, Tripp
AU - Douville, Brett
AU - Prevost, Scott
AU - Stone, Matthew
N1 - Funding Information:
This research is partially supported by NSF Grants IRI90-18513, IRI91-17110, CISE Grant CDA88-22719, NSF graduate fellowships, NSF VPW GER-9350179; ARO Grant DAAL03-89-C-0031 including participation by the U.S. Army Research Laboratory (Aberdeen); U.S. Air Force DEPTH contract through Hughes Missile Systems F33615-91-C-000; DMSO through the University of Iowa; National Defense Science and Engineering Graduate Fellowship in Computer Science DAAL03-92-G-0342; and NSF Instrumentation and Laboratory Improvement Program Grant USE-9152503.
Publisher Copyright:
© ACM 1994.
PY - 1994/7/24
Y1 - 1994/7/24
N2 - We describe an implemented system which automatically generates and animates conversations between multiple human-like agents with appropriate and synchronized speech, intonation, facial expressions, and hand gestures. Conversations are created by a dialogue planner that produces the text as well as the intonation of the utterances. The speaker/listener relationship, the text, and the intonation in turn drive facial expressions, lip motions, eye gaze, head motion, and arm gesture generators. Coordinated arm, wrist, and hand motions are invoked to create semantically meaningful gestures. Throughout, we will use examples from an actual synthesized, fully animated conversation.
AB - We describe an implemented system which automatically generates and animates conversations between multiple human-like agents with appropriate and synchronized speech, intonation, facial expressions, and hand gestures. Conversations are created by a dialogue planner that produces the text as well as the intonation of the utterances. The speaker/listener relationship, the text, and the intonation in turn drive facial expressions, lip motions, eye gaze, head motion, and arm gesture generators. Coordinated arm, wrist, and hand motions are invoked to create semantically meaningful gestures. Throughout, we will use examples from an actual synthesized, fully animated conversation.
UR - http://www.scopus.com/inward/record.url?scp=85029641276&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85029641276&partnerID=8YFLogxK
U2 - 10.1145/192161.192272
DO - 10.1145/192161.192272
M3 - Conference contribution
AN - SCOPUS:85029641276
T3 - Proceedings of the 21st Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 1994
SP - 413
EP - 420
BT - Proceedings of the 21st Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 1994
PB - Association for Computing Machinery, Inc
Y2 - 24 July 1994 through 29 July 1994
ER -