Thumbnail

Intel Realsense D435i

by Intel

Connect the D435i Intel Realsense depth cameras to the ARC navigation messaging system (NMS) for distance detection and mapping.

Requires ARC v20 (Updated 1/2/2025) Hardware Info

How to add the Intel Realsense D435i robot skill

  1. Load the most recent release of ARC (Get ARC).
  2. Press the Project tab from the top menu bar in ARC.
  3. Press Add Robot Skill from the button ribbon bar in ARC.
  4. Choose the Navigation category tab.
  5. Press the Intel Realsense D435i icon to add the robot skill to your project.

Don't have a robot yet?

Follow the Getting Started Guide to build a robot and use the Intel Realsense D435i robot skill.


How to use the Intel Realsense D435i robot skill

Connect the D435i Intel Realsense camera to the ARC Navigation Messaging System (NMS) for distance detection and mapping. This should also work with other Intel RealSense depth camera models because the FOV and parameters are used from the device.

User-inserted image

Main Screen

User-inserted image

  1. Select the device by the serial number. This allows multiple Intel Realsense devices to be added to a project.

  2. START/STOP button connecting to the RealSense device.

  3. Log window for status and errors.

Image Align

The Image Align tab allows you to select the area of the image that you wish to detect for the point cloud. For example, setting the detection too low will detect the floor at a close distance. Ideally, set the bar high enough to avoid detecting the floor.

User-inserted image

  1. Adjust the top of the detected image range.

  2. Adjust the bottom of the detected image range.

  3. The detected image range is highlighted in purple.

Configuration

The configuration screen displays several options for customizing the device to work with your robot project.

User-inserted image

Mounted Radius Offset in mm (input)

Offset of the sensor from the center of the robot. A positive number means the sensor is toward the rear; a negative number means the sensor is toward the front. The measurement is from the center of the robot. The sensor must sit aligned with the center of the robot, not toward the left or right. The sensor must face 0 degrees heading.

Sensor Width, Height, and Frame Rate (input)

Sensor resolution during initialization. If the values are changed from the default, obtain them from the Intel RealSense viewer. Changing these values may be necessary if using a sensor different from the D435i. Set both width & height to 0 for auto-size, which isn't the most efficient for performance.

Skip Scan Points X and Y (input)

The number of depth pixels to skip when processing the depth data. This is useful for lower CPU PCs or when high-resolution depth is not required. Since the sensor has a limited field of view (FOV), processing every depth pixel may not be necessary. If the sensor resolution is 1280x720, the CPU would be processing 921,600 depth points for every scan of the specified framerate. At 1280x720 @ 15 FPS, that's 13,824,000 depth points per second. If the Skip Scan Points for X & Y are set to 10 for that resolution/fps, the CPU would only need to process 9,216 depth points per scan (or 138,240 per second).

The editor will recommend a value for the best balance between performance and detection based on the sensor FOV.

Enable Video Stream (checkbox)

The camera video from RealSense can be pushed into a selected Camera Device. In the selected camera device, choose CUSTOM as the device type. This will accept the video feed from this robot skill. Ensure START is pressed on the selected camera device as well.

Use NMS (checkbox)

Enable pushing the data to the NMS to combine with a navigation robot skill.

Enable the Fake Pose hint checkbox

  • Do not use if another robot skill provides pose data to NMS
  • 3d Cameras do not provide enough data for SLAM systems to identify the position, so this isn't very useful. You would want at least two cameras (facing opposite directions) for this to work properly. Or combine with a lidar, etc..

SLAM systems, such as The Better Navigator (Hector), will require the robot's pose/position in cartesian coordinates to identify its position. This is generally provided by an odometry setup such as wheel encoders. If your robot does not have this, you can fake the pose hint by selecting this option. This will send 0,0 as the pose location of the robot to force The Better Navigator to render. Do not check this if you have another robot skill submitting pose data to the NMS!

Getting Started

  1. At the time of this guide, the Intel Realsense Viewer being used is available here, in the Assets section of the release: https://github.com/IntelRealSense/librealsense/releases/tag/v2.55.1

  2. Ensure the computer's USB 3 port is used. Second, load the Intel RealSense Viewer, which is available in the Intel Realsense SDK package. A prompt to "Enable USB Meta Data" may be displayed, a system-wide change. Ensure you answer ENABLE to that option. Lastly, update the firmware if there is a prompt to do so.

    User-inserted image

  3. If the sensor works in the Intel RealSense viewer, it will work with this robot skill. Depending on the version of the sensor, you may need to configure the capture width/height/framerate. This can be done in the robot skill configuration screen. The RGB and Depth camera resolution values must be the same. You can use the Intel RealSense Viewer to see what values work for your camera. If the RGB camera is 640x480 x 15 FPS, the depth camera must have a supporting resolution with the same settings. This robot skill will parse the depth and RGB data together.

  4. Using the RealSense viewer, select a profile and note the resolution it is set for. The "Hand" profile seems to work well, although a lower camera imager is ideal to limit the amount of processing on your PC.

Variables

This skill will create four global variables that scripts can reference.

User-inserted image

$D435IsRunning - status of the D435 robot skill, whether connected to the sensor or not.

$D435FurthestDistanceCM - furthest distance detected

$D435FurthestDistanceDegree - furthest degree of the furthest distance CM variable.

$D435NearestDistanceCM - nearest distance detected

$D435NearestDistanceDegree - nearest degree of the closest distance CM variable.

ARC Navigation Messaging System

This skill is part of the ARC navigation messaging system. You are encouraged to read more about the messaging system to understand the available skills HERE. This skill is in level #3, group #1 in the diagram below. It contributes to telemetry positioning in the cartesian positioning channel of the NMS. Combining this skill with Level #3 Group #2 skills for obstacle avoidance works well. For Level #1, the Better Navigator robot skill is recommended.

User-inserted image


Control Commands for the Intel Realsense D435i robot skill

There are Control Commands available for this robot skill which allows the skill to be controlled programmatically from scripts or other robot skills. These commands enable you to automate actions, respond to sensor inputs, and integrate the robot skill with other systems or custom interfaces. If you're new to the concept of Control Commands, we have a comprehensive manual available here that explains how to use them, provides examples to get you started and make the most of this powerful feature.

Control Command Manual


  • controlCommand("Intel Realsense D435i", "Start")
    Starts the D435i robot skill.



  • controlCommand("Intel Realsense D435i", "Stop")
    Stops the D435i robot skill.




ARC Pro

Upgrade to ARC Pro

Your robot can be more than a simple automated machine with the power of ARC Pro!

PRO
Synthiam
#1   — Edited

Fixed Intel's bug so the device can be started and stopped

There is a small performance improvement as well, which will be noticeable on SBCs

Also, this only displays depth sensors in the drop-down

PRO
Synthiam
#2   — Edited

Pretty big bug fix that correctly calculates the point-cloud distances into distance scans. You may have noticed that flag objects, such as walls were rounded. This is no longer the case.

Here you can see the wall, corner of the wall, and the edge of the monitor. 

User-inserted image

And here's a low scan of just the table in front - where it's flat and no longer curved.

User-inserted image

PRO
Synthiam
#3   — Edited

Updated with a huge performance improvement. I didn't realize the Intel sample code was copying the depth data into a managed array on every frame ouch. So I'm referencing the unmanaged memory pointer directly. Should have known better than to copy that bit of code from their sample:D

Should experience a 10 times improvement. I went from 27% cpu to 2% cpu

User-inserted image

PRO
Synthiam
#4  

Like The Navigator, there are variables for the nearest and furthest object detected... just in case this is being used without an NMS skill like the navigator

User-inserted image

PRO
Canada
#5  

Thank you so much for making this.  Just installed on the Rock Pi X.  Seems to work well although took me a bit of time to get it working. I have to run it off the USB 3.0 port as I get an error on USB 2.0 but the T265 seems to work on the USB 2.0 port ok

Couple of questions Did you manage to get to work off the USB 2.0 port or do we not have enough bandwidth ? If you can use off usb 2.0 could it work with 2 or more D435i ? If you mount the camera sideways could you set image range vertically instead of horizontally?

PRO
Synthiam
#6   — Edited

The device is usb c which requires usb 3. I’m sure there’s more information about that in the product manual

PRO
Canada
#7   — Edited

Ok it works with USB 2.0 with the realsense view app provided by intel.

I guess I could try and  use the USB C port for 1 and the USB 3.0 port for another.

PRO
Synthiam
#8  

I don’t know how that could be. It’s usb-c which is usb 3. If it’s true, then maybe their app is doing some wizardry that uses a different driver and somehow defies usb-c specifications lol