US20240338915
2024-10-10
Physics
G06T19/20
A method has been developed to create controllable, dynamic appearances for neural 3D portraits. This involves projecting color onto specific points in a digital video portrait, taking into account factors such as location, surface normal, and viewing direction. The technique allows for the projection of dynamic face normals that change according to head poses and facial expressions captured in the video, enabling a realistic representation of movement and emotion.
Controllable 3D portraits are essential for augmented reality (AR) and virtual reality (VR) applications, where an immersive experience is crucial. By digitally recording an individual under controlled lighting conditions, a video is created which serves as the basis for training a model that can manipulate head movements and facial expressions accurately. This technology enhances the realism of virtual interactions by allowing for precise control over the animated portraits.
Creating photo-realistic moving portraits typically requires strict lighting conditions to ensure consistent color across surfaces. However, real-world capture often introduces complications such as shadows, specularities, and interreflections that vary with different head poses and expressions. These challenges have limited the creation of animated portraits to professional setups, making it difficult to produce high-quality outputs in everyday environments.
The proposed method addresses these limitations by enabling the capture of training videos in any lighting condition, including those that produce shadows or varying skin reflectance. This flexibility allows users to create animated portraits using mobile devices or less sophisticated equipment without the need for a controlled environment. Consequently, it simplifies the process of generating animated content for various applications such as videos or AR presentations.
The process utilizes a video processing application that disentangles appearance data from captured videos to create dynamic lighting normals and specularity. Users can provide inputs to control how the portrait is animated over time, specifying details like head position and facial expressions. The resulting neural 3D portrait can be stored or rendered for display, allowing for high-quality animations that are easily produced with accessible technology.