North Star software 1: Basic Intel T265 Integration

Categories:   ARUnity3D

If you just want to read how to do the unity integration, jump to Unity Integration


  • Pose: Combination of a position and an orientation.
  • Depth: Estimation of the distance between the user and the environment (Points).
  • Inside-out tracking: Spatial tracking method based on the cameras/sensors being on the tracked device.

Spatial Tracking

If you haven’t done any modifications to your NorthStar headset, you have a 0 degrees of freedom (dof) device, this means it won’t take into account any movement you do, everything is “fixed” to the world coordinate system, even your headset. If you place a cube in front of you and turn your head in any of the three axis (pitch, roll, yaw), the cube will follow your movements. Same thing happens if you move along any of the three axis (forward-back, up-down, left-right).

Six degrees of freedom

We can upgrade our headset to have 6 dof, if we do this, our headset will move relative to the world coordinate system, allowing us to move around everything situated in both the real and virtual world (We can now fix objects in real-world positions and they will stay there!).


There are a few options to track the headset in space, some of them (the most used ones) are:

  Devices       Pose      Inside-out tracking     Depth  
Intel Realsense T265 Yes Yes No*
Occipital Structure Core Yes Yes Yes
Vive tracker Yes No** No

*While not oficially supported, the Intel page states that it is possible to compute a suboptimal dense depth. This is left as an exercise for the reader.

**To use a Vive tracker, you must have a Lighthouse system.

I decided to use an Intel Realsense T265 device, mainly because of it’s price and my needs, so that is what we will be integrating in Unity next.

Unity Integration

The first thing you want to do is go to the Intel Realsense Unity Wrapper github page and then download the realsense.unitypackage.

Once you have downloaded it, drag and drop it into your NorthStar Unity project.

Your hierarchy should look like this:

Clean project

Navigate through your project folders Assets -> RealSenseSDK2.0 -> Prefabs and drag and drop the RsDevice prefab in your hierarchy.

RsDevice prefab

Once you have the RsDevice gameobject in your hierarchy, click it and look at the inspector, it should have three profiles by default:

RsDevice three profiles

Change it to one, then, set the Stream variable to Pose,the Format to Six DOF and Framerate,Stream Index,Width and Height to zero.

RsDevice pose profile

Now, navigate to ARCameraRig -> Leap Odometry Origin in your hierarchy, click Add Component and look for Rs Pose Stream Transformer

Add Pose Stream to Leap Odometry

Drag and drop your RsDevice in the Source variable of the Rs Pose Stream Transformer you just added.

RsDevice to source variable

Now you can add a cube to your scene, scale it down a bit, place it in front of the headset and hit play, and you should be able to move around the cube!

Related links

Written on January 27, 2020