1. Real-Time Vision Based Mouth Tracking and Parameterization for a Humanoid Imitation Task
- Author
-
Naomi Inoue, Gordon Cheng, and Sabri Gurbuz
- Subjects
Computer science ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Facial recognition system ,Human–robot interaction ,Motion (physics) ,Stereopsis ,Face (geometry) ,Robot ,Computer vision ,Artificial intelligence ,Software system ,business ,Humanoid robot - Abstract
Robust real-time stereo facial feature tracking is one of the important research topics for a variety multimodal Human-Computer, and human robot Interface applications, including telepresence, face recognition, multimodal voice recognition, and perceptual user interfaces (Moghaddam et al., 1996; Moghaddam et al., 1998; Yehia et al., 1988). Since the motion of a person's facial features and the direction of the gaze is largely related to person's intention and attention, detection of such motions with their 3D real measurement values can be utilized as a natural way of communication for human robot interaction. For example, addition of visual speech information to robot's speech recognizer unit clearly meets at least two practicable criteria: It mimics human visual perception of speech recognition, and it may contain information that is not always present in the acoustic domain (Gurbuz et al., 2001). Another application example is enhancing the social interaction between humans and humanoid agents with robots learning human-like mouth movements from human trainers during speech (Gurbuz et al., 2004; Gurbuz et al., 2005). The motivation of this research is to develop an algorithm to track the facial features using stereo vision system in real world conditions without using prior training data. We also demonstrate the stereo tracking system through a human to humanoid robot mouth mimicking task. Videre stereo vision hardware and SVS software system are used for implementing the algorithm. This work is organized as follows. In section 2, related earlier works are described. Section 3 discusses face RIO localization. Section 4 presents the 2D lip contour tracking and its extention to 3D. Experimental results and discussions are presented in Section 5. Conclusion is given in Section 6. Finally, future extention is described in Section 7.
- Published
- 2021