...
The JSON configuration file defines the input/output and processing components within LPVR. More specifically there are on overall header tag and four sub-module types:
JSON Tag Name | Function | |
---|---|---|
1 | PoseMachineConfig | Header tag that preceeds the definition of all sub-modules |
2 | absoluteSources | Defines the absolute positioning (optical tracking) devices used to acquire global poses |
3 | imuSources | Defines sources for IMU orientation data |
4 | trackedObjects | Defines how objects are tracked i.e. their data sources and how the different data sources are combined |
5 | emitters | Defines the output target of the sensor fusion done for trackedObjects. This is usually an HMD. |
The configuration file consists of these components to define the flow of input signals to output signals. The input to a typical system would consist of an absoluteSources structure for an optical tracking input, one or more imuSources and one or more emitters to output the information. All input data is combined in the trackedObjects structure that defines the parameters of the sensor fusion and signal routing. The figure below shows an overview of the structure of a configuration file that works for differential IMU headset tracking.
Info |
---|
The outer-most structure “PoseMachineConfig” : { } should only be added when editing settings.json directly. When editing the configuration in the integrated editor on the configuration page, it needs to be left awayomitted. |
...
Optical Tracking Sources
The absoluteSources tag describes an absolute position and orientation source like an optical tracking system. Currently LPVR supports VICON, Optitrack and ART tracking system natively, as well as the common VR communication protocol VRPN. "absoluteSources"
is an object whose properties describe the optical tracking systems in use. The name of each property is used to reference the tracking system in the reaminder of the configuration. Examples for various tracking systems are given below.
Optical Tracking System | Example Code Block | Explanation | |
---|---|---|---|
ART |
|
|
|
|
|
|
name
type Must be DTrack settings: axisPermutation Optional axis permutation setting to adjust coordinate system settings: host Address of the host PC settings: port Port number of host PC |
type
Must be DTrack
VICON |
|
|
|
|
|
|
|
|
name
type Must be Vicon settings: host IP address of the VICON host computer running VICON Tracker, Blade etc. settings: |
trackedObjects |
Name of the rigid body in VICON software.
settings: subjectName
Should be the same as the segmentName
type
an array of JSON objects. Each contains | |||
Optitrack |
|
|
|
|
|
name
type Must be OptiTrack settings: connectionType |
Can be Multicast or Unicast settings: localAddress Local address of the Optitrack client settings: remoteAddress Address of the Optitrack server settings: serverCommandPort Must be 1510 settings: serverDataPort Must be 1511 |
type
Must be OptiTrackVRPN |
|
|
|
|
|
|
name
type Must be VRPN settings: tracker Name and address of VRPN server |
type
Must be VRPNIMU Sources
The imuSources
tag describes the IMU attached to the headset. At the moment only the LP-REASEARCH LPMS IMU is supportedIMUs in use. It is a JSON object whose properties define the individual IMUs.
Example Code Block | Explanation | |
---|---|---|
|
|
|
|
|
|
|
|
|
|
id
Defines
| Each member of the type Can either be
settings: name Specifies the ID of the connected IMU (not required for ViveHeadset). settings: autodetectType For |
Output of Pose Calculation
The orientation and position calculated by sensor fusion of IMU and optical tracking is output to a headset by an emitter. The emitter allows setting parameters determining the position and orientation offset between IMU coordinate system and the optical coordinate system of the headset.
Example Code Block | Explanation | |
---|---|---|
|
|
| name Defines the name of the output device. Any name is good. Will be referenced in trackedObjects tag. HMD Emitter Emits orientation to HMD settings: |
imuFromEyeQuat Rotation from |
eye frame to the |
IMU frame in which the graphics are rendered. To operate in the optical coordinate system, imuFromEyeQuat must the the inverse of absoluteFromImuFrameQuat. Note: imuToEyeQuat is not a valid parameter, please don’t use. settings: |
imuFromEyeVect Translation from |
eye frame to |
IMU frame in which the graphics are rendered. Note: imuToEyeVect is not a valid parameter, please don’t use. settings: type Must be HMD type Must be OpenVR Console Emitter Displays orientation as SteamVR console output settings: interval Interval between log outputs in ms name Must be Console |
Pose Calculation
The actual pose of an object is calculated by combining IMU data and optical tracking information. This tag combines the modules we defined above. The result pose is forwarded to the emitter block.
Example Code Block | Explanation | ||
---|---|---|---|
| absoluteSource:name Name of the previously defined absolute source (VICON etc.). absoluteSource:trackingId ID of object tracked by optical system. combinerType Type of sensor fusion used.
emitterName Name of emitter to output data to imuSource Name of IMU source declared above (headset IMU) settings: absoluteFromImuFrameQuat Orientation of the tracked body frame relative to the IMU frame. Make sure this parameter is spelt correctly in your configuration, there’s been some confusion also internally here at LP about the exact naming convention. settings: absoluteFromImuFrameVect Translation of the tracked body frame relative to the IMU frame settings: ignoreGravity If true, acceleromter data of the headset IMU is not used to correct HMD pitch and roll orientation. This should be true for in vehicle applications and false for stationary installations. settings: opticalWeight Impact of the optical orientation tracking on oprientation measurements. Default = 0.005 settings: reference_imu Name of the IMU to be used as reference (fixed to vehicle) IMU settings: referenceOrientationQuat Orientation of the reference IMU body inside the optical tracking space settings: referenceToOpticalQuat Rotation to translate from reference IMU internal coordinate system to optical tracking coordinate system |
LPVIZ: Procedure for Manual Adjustment of HMD Marker Target
Adjust the optical target first: set imuSource to no_imu (need to create no_imu first) and adjust the target live in DTrack (ART) or Motive (Optitrack). Make sure the coordinate system center in the Steamvr grid is eactly in the center between the two Smarttrack cameras or whereever you set the center of your coordinate system.
Then, switch on the the HMD IMU (imuSource: hmd imu) and calibrate optical vs. imu.
To operate in the optical coordinate system, imuFromEyeQuat must the the INVERSE of absoluteFromImuFrameQuat.
Check a simple Unreal or Unity test scene with a single object at (0, 0, 0) and player start at (0, 0, 0) rot=(0, 0, 180). you should observe very little distortion and shifting when looking around the object.
Only after this has been confirmed, test in the moving vehicle.