back
Back to Glossary

What is LiDAR annotation?

LiDAR point cloud labeling, often called LiDAR annotation, is the process of adding structured labels to LiDAR point clouds so machine learning models can learn to detect, classify, and track objects in 3D.

A LiDAR sensor emits laser pulses and measures the return time, producing a point cloud made up of 3D points with x, y, and z coordinates and, in many systems, an intensity value. Annotation turns that raw geometry into training data by attaching meaning to points, point clusters, or 3D shapes that represent real-world objects.

Teams use LiDAR labeling when depth and spatial geometry matter. Common use cases include autonomy and robotics, advanced driver assistance systems (ADAS), warehouse automation, and mapping. Many projects combine LiDAR with camera images, radar, and GPS/IMU data.

When sensors are synchronized, labels can be cross-checked across views: a 3D box drawn in the point cloud should correspond to the object in the camera frame, and object motion should stay consistent across frames in a sequence.

The most common labeling primitive in LiDAR is the 3D bounding box, also called a cuboid. A cuboid defines an object’s position, size, and orientation in 3D space. Datasets often store attributes alongside the cuboid, such as class, occlusion level, truncation, and whether the object is moving. When models need richer geometry than a box, teams use point-level labels, often described as 3D semantic segmentation, where each point is assigned a class like road, curb, vehicle, or vegetation. 3D instance segmentation adds an object identity on top of the class, so two cars are separated as two instances.

The right output depends on what the model must produce in production. Boxes are efficient to label and work well for 3D detection and tracking. Point-wise segmentation supports tasks like drivable area estimation and obstacle surface modeling, but it is slower to label and harder to quality check. Many real pipelines mix both: cuboids for dynamic objects and segmentation for a limited set of static classes that drive safety or navigation.

LiDAR annotation has predictable hard cases. Point clouds get sparse at long range, so distant objects may have only a handful of points. Occlusions create partial shapes, and reflective surfaces can add artifacts.
These conditions can cause reviewer disagreement: one annotator may label a partial object tightly around visible points, while another may infer full physical extent. A good project resolves this upfront with clear policies for labelability, boundary placement, and orientation rules.

Example

Imagine labeling a highway sequence for a 3D detection model. A sedan is partially hidden behind a truck. In the point cloud, only points from the sedan’s roofline and rear are visible. If your policy represents full physical extent, the annotator should place a full-size cuboid that matches typical sedan dimensions, aligned to the vehicle’s heading. If your policy represents only visible extent, the cuboid should be tight around visible points. Mixing both styles inside the same dataset usually creates training noise, because the model sees two different “truths” for the same visual pattern.

Quality checks for LiDAR labeling go beyond whether an object exists. Reviewers check box placement, yaw angle accuracy, class correctness, and temporal consistency when tracking is required. Temporal consistency matters because tracking models rely on stable object IDs and smoothly moving boxes, not labels that jump from frame to frame. In multimodal setups, reviewers also validate projections: a cuboid should align with the object in camera views, and segmentation should not bleed into neighboring structures.

When evaluating tooling or vendors, test on hard samples, not clean demo frames. Check how quickly annotators can switch between point cloud and camera views, how efficiently they can label sequences, and how disagreements are handled in review. Export quality is equally important. A usable export preserves coordinates, timestamps, object IDs, attributes, and sensor metadata, and it matches your training and evaluation formats without manual cleanup.