Overview
Here is a post about the LiDAR Working Principle and different types of LiDARs.
One lidar model is the Range-Azimuth-Elevation model (RAE)
Its cartesian coordinates are:
If given
Some point clouds have meta data such as reflectance (intensity), RGB, etc. Reflectance can be used for ground detection, lane detection. Based on the above, LiDAR manufacturers transmit packets (usually) through UDP after measurements. Velodyne HDL-64S3 is a 64 line LiDAR. At each azimuth angle azimuth, (rotational position in the chart), a message contains on laser block that contains 32 lines (along the elevation angle). Its packet looks like:
- So, to find the
[x,y,z]
of a single lidar point,elevation_angle = 32*block_id + array_index
azimuth = rotational_position * 0.01 deg / 180 deg * pi
(Velodyne lidar azimuth increments in 0.01 deg)range = array[array_index]
- In total, we need
2 + 2 + 32 *3 = 100 bytes
to represent the LiDAR Data. If we use thepcl::PointCloud
, each point would be [x, y, z, intensity], and that’s32 * (4 + 4 + 4 + 1) = 416bytes
. So packets are compressed LiDAR data.
Aside from 3D Point Cloud, another 3D representation is a Surface Element Map (Surfel). A surface element is a small 3D patch that contains:
x,y,z
- Surface normal vector
- Color / texture
- Size / Radius
- Confidence
Introduction: Point Cloud Query
In SLAM, we often need to query a point cloud map about the occupancy of points. In effect, this is a query function: bool is_there_a_point(point)
. Or, we try to find the nearest neighbor of that point. This is a mapping problem. Before jumping into the classical algorithms, we note that this can be learned by neaural networks (NN) without labelling. NN can also learn the implicit relationships in the pointcloud, like connectivity of the points; NN can also learn a signed-distance-field (SDF) as well. Note that an NN for querying one point cloud map may not be usable for another point cloud. But if we have a huge point cloud that doesn’t change for a good amount of time, NN can generate a good querying result:
(See how smooth the reconstructed pointcloud is?)
Some other reaserchers focus on finding descriptors of point clouds. For SLAM, a tiny bit of performance improvement can create a huge difference! Therefore, we must consider parallelism in implementation.
Now, let’s get into the conventional methods. Most There are 2 popular ways for point cloud querying in SLAM
- Point cloud is spatial. We can store them in spatial representations for faster indexing, like in a pixel / voxel grid.
- We can store a point cloud in a tree, a binary-search tree /KD tree, or a quad-tree or octo tree.
- Other trees like B-Tree are suitable for large static point clouds. In SLAM, we need to rapidly query in different point clouds.
Brute Force KNN Search
Brute-Force KNN is very easy to parallelize, and it could be more efficient than more sophisticated algorithms.
Brute force KNN is:
- Given a point, search through all points to find their distances. Find the K shortest distances
For K-Nearest-Neighbor search, we need an extra sorting process to find the k shortest distances. Of course, we are able to achieve 100% recall and precision using the brute force search
One-Pass Grid Search Method
A faster method is to put point cloud into pixel / voxels. Here is how it’s done:
- For a given point
p
, calculate the coordinatec
of with the known resolution r- Note that adjacent points may have the same coordinate.
- Hash the coordinate into an integer
- We use the Locality-Sensitive-Hashing function in “Optimized Hashing Function for Collision Checks”. Reference
hash(x,y,z) = ( x p1 xor y p2 xor z p3) mod n
, where p1, p2, p3 are large prime numbers, in our case 73856093, 19349663, 83492791.
- Store the point into an
unordered_map
into an array with hash.
During Search:
- Calculate the hash of the query.
- Fetch all points of the pixel / voxels under the same hash, and their neighbors (see below).
- Brute force search through these points, find K nearest neighbors.
For 2D grids, we can have center (0-neighbor) 4-neighbors, 8 Neighbors.
For 3D voxel grids, there are 6-neighbors, 18-neighbors, 26-neighbors.
KD Tree
In a KD Tree, there are 2 branches of a non-leaf node. A Leaf node represents an actual point, a non-leaf node doesn’t represent an actual point, but it just helps direct query searches to the right leaf node. A branch represents a “split” (or a “hyper plane”) at the specified dimension. KD-tree can be easily implemented in the recursive form. One thing to note is recursion may cause stack overflow. Using for-loops would not.
Tree Building:
- Input: a group of points
[x, y, z]
- If the group is empty, return
- If the group contains only 1 point, insert a new leaf node into the tree, return
- Otherwise, the group has multiple points. Calculate the mean and variance of these points across every dimension.
- Find the dimension with the largest variance. That will be our split dimension. The mean will be the split threshold.
- Partition the point group into two, the group below split threshold is inserted into the left child node, the other is inserted into the right child node.
- For each child node, repeat from step 1.
Tree searching:
The reason why tree searching is faster than brute force is because of tree-pruning. Tree pruning is: if the worst distance between my current candidates and the query is better than that between the query and a given split, we don’t need to look into that split.
- Input: current tree node, query point p, K
- Output: list of K nearest neighbors
- For the current node
nc
:- If
nc
is a leaf,- calculate its distance to p,
d_cp
- If
d_cp
< max(current_k_neighbor distances),d_max
, pop the most distant neighbor, and addnc
to the k nearest neighbors.
- calculate its distance to p,
- Otherwise, evaluate which child node should be evaluated first
- According to
nc
’s split dimensions_d
and thresholds_v
, find dimensionm_1
to evaluate - Repeat from step 1 for
m_1
until return - Evaluate if the other dimension,
m_2
needs to be explored.- Caculate the split distance of
m_2
on dimensions_d
to the point,d_split
- if
d_split < d_max
, explorem_2
, otherwise we don’t.
- Caculate the split distance of
- According to
- If
The worst case scenario of KD Tree is to traverse through the entire tree. That is, O(n Log(n))
, which could be worse than the brute force. But most cases, KD Tree is much faster. To speed up searching, we can moderately sacrifice precision and recall by utilizing “approximated nearest neighbor” (ANN):
- Approximated nearest neighbor is
d_split < alpha * d_max
- KD Tree without approximated nearest neighbor should be able to find all K nearest neighbors
NanoFLANN: what’s a leaf node size?
In many textbook descriptions of kd trees, each leaf is thought of as holding a single point, and internal nodes only store splitting thresholds along a chosen dimension. However, in practical implementations such as nanoflann, the tree is built as a “bucket kd tree.” Here’s how it works:
- The tree is built by recursively splitting the dataset. At each internal node, nanoflann chooses a splitting dimension (typically the one with the largest spread) and a splitting threshold (often the median of the points along that dimension). Points are partitioned into two groups: those with coordinates below (or equal to) the threshold go to the left child, and those above go to the right child.
- Leaf Nodes as Buckets:
- Instead of splitting until each leaf holds exactly one point, nanoflann stops splitting when the number of points in a node is less than or equal to a specified maximum (provided as a parameter in
KDTreeSingleIndexAdaptorParams
, e.g., 10). At that point, the node becomes a leaf node. - Leaf nodes in nanoflann do not represent a single point; they are buckets that contain several points. When a nearest neighbor query reaches a leaf node, it performs a linear search over all the points in that bucket.
- Instead of splitting until each leaf holds exactly one point, nanoflann stops splitting when the number of points in a node is less than or equal to a specified maximum (provided as a parameter in
Octo Tree
Octo-Tree is similar to KD tree, it spatially segments a point cloud into tree nodes. It is a tree data structure in which each internal node has exactly eight children. Octrees are most often used to partition a three-dimensional space. In KD Tree, we have “splits”, whereas in octotree, we have 3D boxes.
Tree building:
- Input: point group
- Initialize the octo tree with 1 node, which is bounded by the min and max values of
x, y, z
- Initialize the octo tree with 1 node, which is bounded by the min and max values of
- If the current point group is empty, return
- If the current point group has 1 point, add a new leaf node to the tree and return.
- If the current point group has more than 1 point:
- Create 8 children under the current node.
- Calculate the bounding boxes of each children. For each dimension in
x,y,z
, we create the bounding boxes by0.5 * current_boundary
- Partition the point group into 8 subgroups
- For each child, repeat from step 1 to insert their subgroup of points
Tree search:
- Input: current tree node, query point p, K
- Output: list of K nearest neighbors
- For the current node
nc
:- If
nc
is a leaf,- calculate its distance to p,
d_cp
- If
d_cp
< max(current_k_neighbor distances),d_max
, pop the most distant neighbor, and addnc
to the k nearest neighbors.
- calculate its distance to p,
- Otherwise, evaluate which child node should be evaluated first
- Find boxes [
m_1
, …m_8
] that is closest to the query point in ascending order. - For each box,
- Evaluate
d_box < d_max
. If not, return. If so, repeat from step 1 form_1
to update the k-nearest-neighbor list until return.
- Evaluate
- Find boxes [
- If
Performance Summary
The speeds of above methods (Query Per Second, QPS) are:
1
grid search > KD tree >Octo Tree >> Brute Force
In reality, we 80% precision and recall is acceptable for point cloud registration.
Related Issues not found
Please contact @ricojia to initialize the comment