Home People Projects Teaching Publications Contact Jobs
 
+    
Reanimating Faces in Video and Images
   
  References:
Reanimating Faces in Video and Images [pdf]
Volker Blanz, Curzio Basso, Tomaso Poggio and Thomas Vetter

Computer Graphics Forum 22 (3), EUROGRAPHICS 2003, Granada, Spain, p. 641 - 650, 2003. Best Paper Award

Video 1 [mov]

Video 2 [mov]
       


Figure
Reconstructed from the original images (left column), 3D shape can be modified automatically to form different mouth configurations. The paintings are Vermeer's "Girl with a Pearl Earring", Tischbein's Goethe, Raphael's St. Catherine, and Edward Hopper's self-portrait. The bottom left image is a digital photograph. The wrinkles are not caused by texture, but entirely due to illuminated surface deformations. In the bottom-right image, they are emphasized by more directed illumination. Teeth are transferred from 3D scans. Note the open mouth in Vermeer's painting, closed by our algorithm (top row, second image).
       
 

Abstract
This paper presents a method for photo-realistic animation that can be applied to any face shown in a single image or a video. The technique does not require example data of the person's mouth movements, and the image to be animated is not restricted in pose or illumination. Video reanimation allows for head rotations and speech in the original sequence, but neither of these motions is required.
 
In order to animate novel faces, the system transfers mouth movements and expressions across individuals, based on a common representation of different faces and facial expressions in a vector space of 3D shapes and textures. This space is computed from 3D scans of neutral faces, and scans of facial expressions.
 
The 3D model's versatility with respect to pose and illumination is conveyed to photo-realistic image and video processing by a framework of analysis and synthesis algorithms: The system automatically estimates 3D shape and all relevant rendering parameters, such as pose, from single images. In video, head pose and mouth movements are tracked automatically. Reanimated with new mouth movements, the 3D face is rendered into the original images.