Creating an Immersive Sensor

This page shows how to create an Immersive Sensor to visualize what is viewed from a specific point of view of a scene. An immersive sensor allows you to generate an .speos360 file that can be visualized in Virtual Reality Lab.

Important: This feature is only available under Speos Premium or Enterprise license.

To create an Immersive Sensor:

  1. From the Light Simulation tab, click VR > Immersive .

    Spatial indicators appear in the 3D view to demonstrate how the sensor is oriented in space (front, back, bottom, top, left and right positions are indicated).



  2. If you want to adjust the Axis system of the sensor:
    • Click to select another origin than the absolute origin of the assembly.
    • Click and select a line to define the Front direction (corresponding by default to Y axis).
    • Click and select a line to define the Top direction (corresponding by default to Z axis).
    • or click and select a coordinate system to autofill the Axis System.
    Important: Make sure the sensor is not tangent to a geometry.
    Note: Make sure no geometry intersects the sensor. Otherwise you may generate unexpected black result in CPU simulation.
    Note: If you define manually one axis only, the other axis is automatically (and randomly) calculated by Speos in the 3D view. However, the other axis in the Definition panel may not correspond to the axis in the 3D view. Please refer to the axis in the 3D view.
  3. In General, edit the Sampling value to recalculate the Central resolution.
    The Central resolution is automatically calculated and depends on the Sampling value.
  4. If you want to define an interocular distance, set Stereo to True and type the interpupillary distance in mm.
    Note: When you define a stereo sensor, make sure that the Front direction is horizontal, the Top direction is vertical, and the Central resolution matches the intended 3D display.
  5. From the Layer drop-down list:
    • Select None to get the simulation's results in one layer.

    • Select Source if you have created more than one source and want to include one layer per active source in the result.

      Tip: You can change the source's power or spectrum with the Virtual Lighting Controller in the Virtual Photometric Lab or in the Virtual Human Vision Lab.
  6. If needed, adjust the Integration angle in case you add your Immersive sensor to a Direct simulation.
  7. In Wavelength, define the spectral excursion of the sensor:
    • Edit the Start (minimum wavelength) and End (maximum wavelength) values to determine the wavelength range to be considered by the sensor.
    • If needed, in Sampling, adjust the number of wavelengths to be computed during simulation.

      The Resolution is automatically computed according to the sampling and wavelength start and end values.

      Note: The sensor sampling is used for each face.

      Every wavelength beyond the defined borders are not taken into account by the sensor.

  8. In Faces, define which faces to include/exclude.
  9. In Optional or advanced settings :
    • If needed, in Visualization size, adjust the preview of the sensor's size.
    • In Preview, from the Active Vision Field drop-down list, select which face of the sensor should be used as default viewpoint for the Automatic framing option.

The Immersive Sensor is created and appears both in Speos tree and in the 3D view.