Introduction to Autonomous Navigation – LIDAR, Sensor Fusion, Kalman Filter

Srinivas Rahul Sapireddy 11 Apr, 2021 • 5 min read
This article was published as a part of the Data Science Blogathon.

Introduction

This article is an introduction to autonomous navigation. First, we will go through the basic concepts of LIDAR and see how sensor fusion helps to minimize the biases during measurement of the surrounding environment of the autonomous robot.

LIDAR stands for Light Detection and Ranging

LIDAR emits light waves into the surrounding environment, and these pulses bounce back and return to the sensor. Then the sensor measures the time to return and calculates the distance it traveled. Repeating this process creates a 3D environment map that can be used for the safe navigation of the robot.

LIDAR autonomous navigation

Types of LIDAR sensor output

[1] LaserScan messages

[2] Point Cloud messages

Any LIDAR sensor can scan the area in front of it with a vision range. We define this vision range as angle_min and angle_max. We also consider angle increment at which particular angle the sensor takes the sample measurements. We also need to provide the time_increment which helps to take the samples at every time interval we define. For LIDAR, we have to detect if there is an obstacle in front of it or not. For this, we consider range_min and range_max. Here the range is an array. This array size depends on angle_increment and it stores the values of the maximum distance of the obstacle at that degree. If no obstacle is present, we take that element of the array as infinity (meaning no obstacle). If there is an obstacle, we assign that distance to it.

Using LIDAR, we can detect the objects and their exact object positions.

autonomous navigation detect objects with LIDAR

A laser scan was taken using a gazebo.

Dead reckoning is calculating the current position of the robot using the previous positions (as with Inertial Measurement Unit – IMU Sensor). But this process is highly susceptible to errors. Thus we use the concept of sensor fusion for localization.

 

Sensor Fusion

We are considering measurements from the combination of multiple sensors so that one sensor can compensate for the drawbacks of the other sensors. This is known as sensor fusion. We implemented sensor fusion using filters.

Types of filters:

[1] Kalman Filter

[2] Complementary Filter

[3] Particle Filter

Kalman Filter

Let us consider two sensors measuring distances from the sensor to the obstacles. Of which sensor 1 can measure short distances with high accuracy and sensor 2 can measure long distances with high accuracy. We want our robot to measure all the distances properly. So which sensor we consider for our measurements for the robot?

We take n measurements from these sensors and estimate the statistical properties of the measurement using PDF (probability density function).

Kalman Filter autonomous navigation
mean and SD

Here we can see that the probability density function is characterized by mean and standard deviation.

Here we take the n measurements from sensor 1 and sensor 2 and find the mean and standard deviation of these sensors. Then we find the probability density function for both the sensors.

autonomous navigation PDF

Probability Density Functions of sensor 1 and sensor 2.

Then how do we consider the true value of distance? Will you consider sensor 1 readings or sensor 2 readings? Here is where we use the Kalman filter.

Here we find the new mean and standard deviation using the formulas below using the Kalman filter.

new mean and standard deviation

From the figure, we can say that we are more confident of our measurements from the new mean and new standard deviation.

measurements from the new mean and new standard deviation.

We can see that the new reading is narrower, which shows that we are more confident about our final readings with the new mean and standard deviation. This is an example of a static estimation case.

Example of dynamic estimation case

Example of dynamic estimation case

Blue line–Actual path of the robot

Black line–Position measurements using IMU (We can see that the error using IMU sensor by comparing the measurements from the blue path)

Green dots–Measurements from GPS

Red line–Sensor fusion using Kalman filter measurements considering measurements from IMU and GPS.

From the figure, we can see that we measure the actual path using sensor fusion on fusing sensors. From this, we can say that we are more confident about our final measurements by using the concept of Kalman filters.

Code for Kalman Filter in Python

import numpy as np
import matplotlib.pyplot as plt
import math
sensor_data1 = 100*np.ones(20)+8*np.random.rand(20)
sensor_data2 = 105*np.ones(20)+5*np.random.rand(20)
print('Sensor1_data :' , sensor_data1)
print('Sensor2_data :' , sensor_data2)
def find_meanvariance(sensor_data):
    mean = np.mean(sensor_data)
    variance = np.var(sensor_data)
    return [mean,variance]
m1,v1 = find_meanvariance(sensor_data1)
m2,v2 = find_meanvariance(sensor_data2)
print('mean1 = ', m1)
print('variance1 = ', v1)
print('mean2 = ', m2)
print('variance2 = ', v2)
def gaussian_dist(x,mean,variance):
    y = (1/np.power(2*np.pi*variance,0.5))*np.exp(-np.power(x - mean, 2.) / (2 * variance))
    return y
x1 = np.linspace(m1 - 5*v1, m1 + 5*v1, 100)
x2 = np.linspace(m2 - 5*v2, m2 + 5*v2, 100)
y1 = gaussian_dist(x1,m1,v1)
y2 = gaussian_dist(x2,m2,v2)
plt.plot(x1,y1,'g-',label='sensor1')
plt.plot(x2,y2,label='sensor2')
plt.legend(loc='upper left')
plt.show()
def m_v_new(m1,m2,v1,v2):
    #insert code here
    m_new=(m1*v2+m2*v1)/(v1+v2)
    v_new=(v1*v2)/(v1+v2)
    return [m_new,v_new]
# Plot the new gaussian distribution
m_new,v_new = m_v_new(m1,m2,v1,v2)
x_new = np.linspace(m_new - 5*v_new, m_new + 5*v_new, 100)
y_new = gaussian_dist(x_new,m_new,v_new)
plt.plot(x1,y1,'g-',label='sensor1')
plt.plot(x2,y2,label='sensor2')
plt.plot(x_new,y_new,label='final')
plt.legend(loc='upper left')
plt.show()

References

[1]https://taylor.raack.info/2017/09/autonomous-vehicle-technology-sensor-fusion-and-extended-kalman-filters-for-object-tracking/

[2]https://www.ness.com/sensor-fusion-algorithms-for-autonomous-driving-part-1%E2%80%8A- %E2%80%8Athe-kalman-filter-and-extended-kalman-filter/

[3]https://medium.com/lansaar/what-is-sensor-fusion-cb596f7cf446

[4]https://towardsdatascience.com/kalman-filter-an-algorithm-for-making-sense-from-the-insights-of-various-sensors-fused-together-ddf67597f35e

The media shown in this article on autonomous navigation are not owned by Analytics Vidhya and is used at the Author’s discretion. 

Frequently Asked Questions

Lorem ipsum dolor sit amet, consectetur adipiscing elit,

Responses From Readers

Clear