US20240212262
2024-06-27
Physics
G06T15/20
A 3D conversation system enables users to engage in conversations within an augmented reality environment, simulating face-to-face interactions. The system operates through a series of data processing stages, including calibration, capture, tagging, filtering, compression, decompression, reconstruction, rendering, and display. This pipeline captures images from the sending user and transforms them into a viewpoint that aligns with the receiving user, ensuring synchronized audio and visual output.
The technology focuses on enhancing communication methods by facilitating three-dimensional (3D) conversations between multiple devices. Traditional video conferencing lacks the depth of in-person interactions due to its reliance on two-dimensional (2D) displays that hinder understanding of body language and spatial dynamics. The 3D conversation system aims to overcome these limitations by allowing participants to move relative to one another in a more immersive experience.
The pipeline involves several key stages:
The calibration stage is crucial for ensuring accurate data capture. It gathers intrinsic parameters (fixed camera features) and extrinsic parameters (contextual conditions) that characterize how light is interpreted by the camera. Each camera involved in the conversation can be calibrated individually, allowing for precise mapping between camera coordinates and pixel coordinates in images.
The system is designed to adapt dynamically based on conversation context, which includes factors like processing capacity, bandwidth availability, and user settings. This flexibility allows for optimized performance during conversations by utilizing appropriate resources and configurations based on real-time conditions, enhancing the overall user experience in augmented reality interactions.