Research
Goal | Eye
Tracking |
Gaze
Tracking | Face
Tracking |Facial
Feature Tracking |Facial
Motion Recovery | Facial
Expression Recogntion
Real-Time
Facial Feature Tracking Under Significant Facial Expressions and
Various Face Orientations |
|
|
|
Facial
features, such as eyes, eyebrows, nose and mouth, and their spatial
arrangement, are important for the facial interpretation tasks based
on face images, such as face recognition, facial expression analysis
and face animation. Therefore, locating these facial features in a
face image accurately is a crucial step for these tasks to perform
well. However, in reality, the appearance of the facial features in
the images varies significantly among different individuals. Even,
for a specific person, the appearance of the facial features is
easily affected by the lighting conditions, face orientations and
facial expressions, etc. Therefore, accurate facial feature
detection and tracking still remains a very challenging task,
especially under different illuminations, face orientations and
facial expressions, etc.
In
our research, we proposed an effective approach to detect and track
twenty-eight facial features from the face images with different
facial expressions under various face orientations in real time. The
improvements in facial feature detection and tracking accuracy are
resulted from: (1) combination of the Kalman filtering with the eye
positions to constrain the facial feature locations; (2) the use of
pyramidal Gabor wavelets for efficient facial feature
representation; (3) dynamic and accurate model updating for each
facial feature to eliminate any error accumulation; (4)imposing the
global geometry constraints to eliminate any geometrical violations.
By these combinations, the accuracy of the facial feature tracking
reaches a practical acceptable level. Subsequently, the extracted
spatio-temporal relationships among the facial features can be used
to conduct the facial expression classification successfully.
|
Facial Feature and Pose Tracking based on 3D
Deformable Models |
|
|
|
In our
recent work, we compared our method with 3D
tracking method based on the
Candide-3
face model and particle filter. The 3D deformable
face model can be manually initialized or be
automatically initialized at the first frame with 28
detected facial feature points detected by our 2D
tracker. Demo 5 gives
the 3D tracking result.
Demo 6 shows the comparison of 2D and 3D
tracker and indicates that the 3D tracker
can also reliably track 28 salient facial feature
points, and provide more other facial feature points
like those on the cheek. Even though currently we
only take into account the rigid motion, the
Candide-3 face model also allows non-rigid motion
and the
Action Unit (AU) parameters of this model can be
directly applied for expression analysis. It can be
seen from Demo 6 that facial feature tracking
and pose estimation result of the 3D tracker are
not as smooth as those output by our 2D tracker, this can be explained
by the fact that this 3D tracker is purely global
feature based (in the 3D tracking method, the whole
face will be warped with the estimated parameters of
Candide-3 into geometrically-free facial patch, and
then compared with a template to output the
likelihood), while our 2D tracker is also based on
local search.
We also
tried directly instantiating the 3D face model with
28 facial feature points provided by the 2D tracker
at each frame. The benefits of doing this include
providing more facial feature points and the AU
parameters. The result is shown in
Demo 7. Exploiting the cooperation between
2D and 3D face model would be one of our future works.
|
|
|
|
Publications: |
(1) Yan
Tong and Qiang Ji, “Automatic Eye Position Detection and Tracking
under Natural Facial Movement”, Passive Eye Monitoring, Editor: Riad
I. Hammoud, pp. 67-84, Springer 2007. (2) Yan Tong, Yang Wang,
Zhiwei Zhu, and Qiang Ji, “Robust Facial Feature Tracking under
Varying Face Pose and Facial Expression”, Pattern Recognition, Vol.
40, No. 11, pp. 3195-3208, November 2007. (3) Zhiwei Zhu and
Qiang Ji, Robust Pose Invariant Facial Feature Detection and
Tracking in Real-Time, the 18th International Conference on Pattern
Recognition (ICPR), Hongkong, August, 2006. (4) Yan Tong, Yang
Wang, Zhiwei Zhu, and Qiang Ji, “Facial Feature Tracking using a
Multi-State Hierarchical Shape Model under Varying Face Pose and
Facial Expression”, the 18th International Conference on Pattern
Recognition (ICPR), Hongkong, August, 2006. (5) Yan Tong and
Qiang Ji, “Multiview Facial Feature Tracking with a Multi-modal
Probabilistic Model”, the 18th International Conference on Pattern
Recognition (ICPR), Hongkong, August, 2006.
|
Demos: |
|
|
|
|
|
|
Demo.1
Real-time facial feature tracking demo (short
version) |
Demo.2
Real-time facial Feature tracking demo (long
version) |
Demo.3 Real-time
facial feature tracking demo with estimated 3D head
pose |
|
|
|
|
Demo.4
Real time facial feature tracking under large pose
and scale changes |
|
|
Latest Demos: |
|
|
|
Demo.5 3D tracking based
on the
Candide-3 face model (Initialized by 28 detected
facial feature points), Bottom is the estimated
pose, top left is the geometrically-free face
template |
Demo.6 Comparison of 2D
and 3D facial feature/pose tracking, Bottom
left/right is the estimated pose of 3D/2D tracker |
Demo.7 3D face model
fitting with facial feature points provided by 2D tracker |
|