210 likes | 322 Views
This study presents a novel method that combines performance-driven facial motion capture with advanced shape-from-shading techniques to accurately capture and represent facial details such as wrinkles and dimples. By leveraging captured images under controlled lighting and implementing spatial and temporal constraints, the proposed method improves the reliability of facial expression reproduction in animation. It highlights the effectiveness of our optimization-based approach in estimating motion dynamics and the potential for application across various reflectance models, paving the way for more realistic character animations.
E N D
Capturing Facial Details by Space-time Shape-from-shading Yung-Sheng Lo*, I-Chen Lin*, Wen-Xing Zhang*, Wen-Chih Tai†, Shian-Jun Chiou† CAIG Lab, Dept. of CS, National Chiao Tung University* Chunghwa Picture Tubes, LTD.†
Outline • Introduction • Acquisition of facial motion • Space-time shape-from-shading • Experiment and results • Conclusion
Introduction • Performance-driven method is one of the most straightforward method for facial animation. • Expression details, e.g. wrinkles, dimples, are key factors but difficult for motion capture. Original captured images Deformation without details
Introduction (cont.) • Physics-based simulation and blend shape methods try to mimic the details. • But, the synthesized details are not the exact expressions. Blend shape [Z. Deng et al. 2006] Muscle-based [E. Sifakis et al. 2005]
Introduction (cont.) • Our goal is to enhance existing motion capture tech. and capture facial details. • With the captured images and directional lighting, our optimization-basedshape-from-shading (SFS) can estimate details from shading in video. Captured video With facial details
The proposed method • Combines the benefits of motion capture and shape-from-shading. • Motion capture and stereo reconstruction • accurate on feature points and general geometry. • Unreliable corresponding matching at textureless regions • Shape-from-shading • don’t need detailed point correspondencefor textureless regions. • Estimate relative undulation. • Sensitive to noise. Motion capture + Space-time shape-from-shading.
Approximate geometry by Mocap • Tracking by block matching and stereo reconstruction. • Deforming a generic face model by radial-basis functions (RBF).
Input image Facial details by SFS • Estimating time-varying details by iterative approximating shape V and reflectance R.
Space-time constraints • Only SFS is not enough. • For more reliable detailed motions, we proposed using space-time constraints. Highly sensitive to noise After applying our spatial constraints
Spatial constraints • Mostly continuous surface • High spatial coherence • For J Є Neighbor(p) • Neighbor(p) denotes the 8-neighbor pixel set • Wj is an adaptive weight • Kcs is the weight for spatial constraints. noise noise Ztp Ztj Ztj Ztj Ztp Ztp Reduce the noise
Temporal constraints • Still flicker • According to biomechanics properties: • A human facial surface should gradually transit between expressions. A video image sequence flicker T2 T2 T0 T0 T1
Space-time shape-from-shading • Finally, our objective function becomes spatial constraints temporal constraints Space-time shape-from-shading + + = shading constraints
Performance issue • if applied our optimization to the whole face. • DOF is too large • Assigned some small windows. • We preferred areas with more wrinkles and creases. Video image sequence N … D.O.F=N*M (pixels) *i(frames) F3 M Fi F2 F1
Experiment • Illumination-controlled (single light source) • Two video streams.(HDV, 1280*720 ,30 fps) • We pasted 25 to 30 markers on human face. {C1} {C2}
Result of synthesis • Generic model: • 6078 vertices • 6315 polygons deformation (RBF) subdivision per-pixel normal mapping
Conclusion & Future work • We propose capturing detailed motion by conventional Mocap and advanced shape-from-shading. • Doesn’t need additional devices, paint pigments, or restrict the wrinkle shape. • With spatial and temporal constraints, our optimal shape-from-shading is more reliable. • Reflectance parameters are also estimated.
Conclusion & Future work • In addition to Phong model, we will extend the concept to other reflectance models. • E.g Cook-Torrance BRDF model, BSSRDF, etc) • Currently, SFS is only applied to designated segments. An more efficient SFS for the whole face will make our animation more realistic.
Thank for your attention! forehead details between eyebrows smile