Chenyi Kuang

PhD student, Rensselaer Polytechnic Institute

About Me

Hello! I’m Chenyi Kuang. Currently I'm a PhD student in ECSE (Electrical, Computer and System Engineering) Department of Rensselaer Polytechnic Institute, under the supervision of Prof.Qiang Ji. My research mainly focus on computer vision tasks in the area of human facial behavior analysis using the tools of deep learning models and 3D geometric models. More concretely, my research is targeting at reconstructing two facial motions: the non-rigid 3D facial motion in terms of facial action units (FAUs) and the rigid 3D eyeball motion, for facial expression recognition and eye gaze tracking.

Research Area: Computer Vision, Deep Learning, 3D Reconstruction and Modeling.

Bio

Email
kuangc2@rpi.edu
Phone
+1 518-961-5132
Address
RPI, Troy, New York, U.S.A

Education

PhD in ECSE from Rensselaer Polytechnic Institute
2019 - present
3D Human Facial Bahaviour Analysis in Computer Vision
Supervisor: Prof.Qiang Ji
Bachelor of Electrical Engineering from University of Science and Technology of China
2015 - 2019
Department of Automation

Research Project

Human Facial Expression recognition & Analysis through 3D Face Modeling:

  • learning personalized 3D face models from images/3D scans.
  • accurate 3D face reconstruction for joint analysis of 3D head poses and facial expressions (for single/multiple subjects).
  • 3D facial action unit recognition through learning AU-aware 3D faces.
  • Facial Action Unit Dataset Labeling:
    We segment facial expression videos from BP4D+ and upload them to Amazon AMT website for online manual labeling. Then we collect worker annotation from the website to perform post-processing to generate frameby-frame data labeling and analyze the accuracy of the labels.
    3D Eye Modeling and Gaze Tracking:

  • constructing deformable 3D eye model for representing the anatomical eyeball structure applicable to various subjects, including eye data collection from a wearable device, data processing for camera calibration, calculation of 3D eyeball parameters (pupil center, cornea center, eyeball center and fovea position), and 3D deformable eyeball basis construction
  • Model-based 3D eyeball reconstruction & gaze estimation by 3DMM-Face-Eye fitting and regression.
  • Publication

    AU-aware 3D Face Reconstruction through Personalized AU-specific Blendshape Learning

    We present a multi-stage learning framework that recovers AU-interpretable 3D facial details by learning personalized AU-specific blendshapes from images. Our model explicitly learns 3D expression basis by using AU labels and generic AU relationship prior and then constrains the basis coefficients such that they are semantically mapped to each AU. Our AU-aware 3D reconstruction model generates accurate 3D expressions composed by semantically meaningful AU motion components. Furthermore, the output of the model can be directly applied to generate 3D AU occurrence predictions, which have not been fully explored by prior 3D reconstruction models.

    More details can be found at: https://sites.ecse.rpi.edu/~cvrl/3DFace_Eye/3DFace.html

    [1] Kuang, Chenyi, Zijun Cui, Jeffrey O. Kephart, and Qiang Ji. "AU-Aware 3D Face Reconstruction through Personalized AU-Specific Blendshape Learning." In European Conference on Computer Vision, pp. 1-18. Springer, Cham, 2022. paper

    Towards an accurate 3D deformable eye model for gaze estimation

    We present a method for constructing an anatomically accurate 3D deformable eye model from the IR images of eyes and demonstrate its application to 3D gaze estimation. The 3D eye model consists of a deformable basis capable of representing individual realworld eyeballs, corneas, irises and kappa angles. To validate the model’s accuracy, we combine it with a 3D face model (without eyeball) and perform image-based fitting to obtain eye basis coefficients The fitted eyeball is then used to compute 3D gaze direction. Evaluation results on multiple datasets show that the proposed method generalizes well across datasets and is robust under various head poses.

    More details can be found at: https://sites.ecse.rpi.edu/~cvrl/3DFace_Eye/3D_eye.html

    Chenyi Kuang, Jeffrey Kephart, and Qiang Ji. Towards an accurate 3D deformable eye model for gaze estimation. Towards a Complete Analysis of People: From Face and Body to Clothes (T-CAP Workshop at ICPR), 2022 paper

    AU-Aware Dynamic 3D Face Reconstruction from Videos with Transformer

    We present a framework for dynamic 3D face reconstruction from monocular videos, which can accurately recover 3D facial geometrical representations for facial action unit (AU). Specifically, we design a coarse-to-fine framework, where the ”coarse” 3D face sequences are generated by a pre-trained static reconstruction model; and the ”refinement” is performed through a Transformer-based network. We design 1) a Temporal Module used for modeling temporal dependency of facial motion dynamics; 2) an Spatial Module for modeling AU spatial correlations from geometry-based AU tokens; 3) feature fusion for simultaneous dynamic facial AU recognition and 3D expression capturing. Experimental results show the superiority of our method in generating AU-aware 3D face reconstruction sequences both quantitatively and qual- itatively.

    Kuang, Chenyi, Jeffrey O. Kephart, and Qiang Ji. "AU-Aware Dynamic 3D Face Reconstruction From Videos With Transformer." Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 2024. paper

    Interaction-aware Dynamic 3D Gaze Estimation in Videos

    We propose a novel method for dynamic 3D gaze estimation in videos by utilizing the human interaction labels. Our model contains a temporal gaze estimator which is built upon Autoregressive Transformer structures. Besides, our model learns the spatial relationship of gaze among multiple subjects, by constructing a Human Interaction Graph from predicted gaze and update the gaze feature with a structure-aware Transformer. Our model predict future gaze conditioned on historical gaze and the gaze interactions in an autoregressive manner. We propose a multi-state training algorithm to alternately update the Interaction module and dynamic gaze estimation module, when training on a mixture of labeled and unlabeled sequences. We show significant improvements in both within-domain gaze estimation accuracy and cross-domain generalization on the physically-unconstrained gaze estimation benchmark.

    Kuang, Chenyi, Jeffrey O. Kephart, and Qiang Ji. "Interaction-aware Dynamic 3D Gaze Estimation in Videos." In NeuRIPS 2023 Workshop on Gaze Meets ML. 2023. paper

    Model-based 3D Gaze Estimation

    This work is under-review.

    Physics-informed 4D face reconstruction

    This work is undergoing.

    Contact

    kuangc2@rpi.edu
    kuangcy1998@gmail.com