Difference between revisions of "X3D AR Requirements and Use cases"

From Web3D.org
Jump to: navigation, search
Line 10: Line 10:
  
 
- Use live video stream as a texture in the X3D scene.
 
- Use live video stream as a texture in the X3D scene.
 +
 
- Use live video stream as a background of the X3D scene.
 
- Use live video stream as a background of the X3D scene.
 +
 
- Retrieve tracking information of the position and orientation of physical objects (such as the camera device and markers).
 
- Retrieve tracking information of the position and orientation of physical objects (such as the camera device and markers).
 +
 
- Use tracking information to change the position and orientation of arbitrary nodes in the X3D scene.
 
- Use tracking information to change the position and orientation of arbitrary nodes in the X3D scene.
 +
 
- Synchronization between video image and tracking information.
 
- Synchronization between video image and tracking information.
 +
 
- Retrieve calibration information of the camera device providing the video stream.
 
- Retrieve calibration information of the camera device providing the video stream.
 +
 
- Use calibration information to set properties of (virtual) camera nodes.
 
- Use calibration information to set properties of (virtual) camera nodes.
 +
 
- Specify key color for the live video stream texture chroma keying, making pixels in this color appear transparent.
 
- Specify key color for the live video stream texture chroma keying, making pixels in this color appear transparent.
 +
 
- Specify a group of nodes as representatives of physical objects, and render those nodes into depth buffer and not into color buffer. As a result, revealing background video on those part where physical objects are rendered, showing correct occlusion between physical objects and virtual objects.
 
- Specify a group of nodes as representatives of physical objects, and render those nodes into depth buffer and not into color buffer. As a result, revealing background video on those part where physical objects are rendered, showing correct occlusion between physical objects and virtual objects.
  
Line 22: Line 30:
 
== 2. Use cases ==
 
== 2. Use cases ==
 
The functions and features could be used in the following use cases:
 
The functions and features could be used in the following use cases:
 +
 
- Augmented Reality applications, where live video stream is shown on the background and the 3D scene is shown as registered in the physical space of the live video stream. (Correct occlusion between virtual and physical objects can be achieved by preparing 3D models of physical objects and specifying them as a representative of physical objects.)
 
- Augmented Reality applications, where live video stream is shown on the background and the 3D scene is shown as registered in the physical space of the live video stream. (Correct occlusion between virtual and physical objects can be achieved by preparing 3D models of physical objects and specifying them as a representative of physical objects.)
 +
 
- Augmented Virtuality (or virtual studio) applications, where live video stream of physical objects can be placed within the 3D scene. (Only the foreground objects can be shown in the live video stream, if the scene in the video is prepared with color matte on the background.)
 
- Augmented Virtuality (or virtual studio) applications, where live video stream of physical objects can be placed within the 3D scene. (Only the foreground objects can be shown in the live video stream, if the scene in the video is prepared with color matte on the background.)
  
 
[http://web3d.org/x3d/wiki/images/6/62/ARWG-Requirements_and_Usecases.pdf pdf version]
 
[http://web3d.org/x3d/wiki/images/6/62/ARWG-Requirements_and_Usecases.pdf pdf version]

Revision as of 23:16, 19 March 2012

Requirements and use cases of X3D functions to support AR and MR visualization

Augmented Reality Working Group, Web3D Consortium

August 17, 2011


1. Requirements

The new set of X3D specification for supporting AR and MR visualization must include the following functions and features:

- Use live video stream as a texture in the X3D scene.

- Use live video stream as a background of the X3D scene.

- Retrieve tracking information of the position and orientation of physical objects (such as the camera device and markers).

- Use tracking information to change the position and orientation of arbitrary nodes in the X3D scene.

- Synchronization between video image and tracking information.

- Retrieve calibration information of the camera device providing the video stream.

- Use calibration information to set properties of (virtual) camera nodes.

- Specify key color for the live video stream texture chroma keying, making pixels in this color appear transparent.

- Specify a group of nodes as representatives of physical objects, and render those nodes into depth buffer and not into color buffer. As a result, revealing background video on those part where physical objects are rendered, showing correct occlusion between physical objects and virtual objects.


2. Use cases

The functions and features could be used in the following use cases:

- Augmented Reality applications, where live video stream is shown on the background and the 3D scene is shown as registered in the physical space of the live video stream. (Correct occlusion between virtual and physical objects can be achieved by preparing 3D models of physical objects and specifying them as a representative of physical objects.)

- Augmented Virtuality (or virtual studio) applications, where live video stream of physical objects can be placed within the 3D scene. (Only the foreground objects can be shown in the live video stream, if the scene in the video is prepared with color matte on the background.)

pdf version