X3D AR Requirements and Use cases
Requirements and use cases of X3D functions to support AR and MR visualization
By Augmented Reality Working Group, Web3D Consortium
August 17, 2011
The new set of X3D specification for supporting AR and MR visualization must include the following functions and features:
- Use live video stream as a texture in the X3D scene.
- Use live video stream as a background of the X3D scene.
- Retrieve tracking information of the position and orientation of physical objects (such as the camera device and markers).
- Use tracking information to change the position and orientation of arbitrary nodes in the X3D scene.
- Synchronization between video image and tracking information.
- Retrieve calibration information of the camera device providing the video stream.
- Use calibration information to set properties of (virtual) camera nodes.
- Specify key color for the live video stream texture chroma keying, making pixels in this color appear transparent.
- Specify a group of nodes as representatives of physical objects, and render those nodes into depth buffer and not into color buffer. As a result, revealing background video on those part where physical objects are rendered, showing correct occlusion between physical objects and virtual objects.
2. Use cases
The functions and features could be used in the following use cases:
- Augmented Reality applications, where live video stream is shown on the background and the 3D scene is shown as registered in the physical space of the live video stream. (Correct occlusion between virtual and physical objects can be achieved by preparing 3D models of physical objects and specifying them as a representative of physical objects.)
- Augmented Virtuality (or virtual studio) applications, where live video stream of physical objects can be placed within the 3D scene. (Only the foreground objects can be shown in the live video stream, if the scene in the video is prepared with color matte on the background.)