Row-wise LiDAR Lane Detection Network
with Lane Correlation Refinement
Dong-Hee Paek1, Kevin Tirta Wijaya2, and Seung-Hyun Kong∗1
Abstract— Lane detection is one of the most important func-
tions for autonomous driving. In recent years, deep learning-
based lane detection networks with RGB camera images have
shown promising performance. However, camera-based meth-
ods are inherently vulnerable to adverse lighting conditions
such as poor or dazzling lighting. Unlike camera, LiDAR
sensor is robust to the lighting conditions. In this work, we
propose a novel two-stage LiDAR lane detection network with
row-wise detection approach. The first-stage network produces
lane proposals through a global feature correlator backbone
and a row-wise detection head. Meanwhile, the second-stage
network refines the feature map of the first-stage network via
attention-based mechanism between the local features around
the lane proposals, and outputs a set of new lane proposals.
Experimental results on the K-Lane dataset show that the
proposed network advances the state-of-the-art in terms of F1-
score with 30% less GFLOPs. In addition, the second-stage
network is found to be especially robust to lane occlusions,
thus, demonstrating the robustness of the proposed network
for driving in crowded environments.
I. INTRODUCTION
To be able to navigate from a point to another, an au-
tonomous driving agent needs to plan a safe and efficient
route according to the environmental conditions. Therefore,
the ability of perceiving the environment through raw sensor
measurements is crucial for the autonomous driving. One
of the perception tasks for autonomous driving is the lane
detection task, where the autonomous driving agent needs to
detect the location of lane lines on the roads.
Extensive studies have been conducted on the lane de-
tection task, particularly with the RGB camera sensors. In
the earlier days, rule-based and heuristic systems have been
developed to provide lane detection capability in limited
predefined environments [1][2][3]. Recently, data-driven ap-
proaches become popular owing to the advancements of
deep learning. Various neural networks for camera-based
lane detection have been developed [4][5][6], with promising
accuracy in most of driving conditions.
However, RGB camera has an inherent weakness towards
harsh lighting conditions such as low or dazzling light. This
is evident in the widely-used CULane benchmark [7], where
the performance degradation of various camera-based lane
detection networks occurs in the dazzling light and dark. As
∗corresponding author
1Dong-Hee Paek and Seung-Hyun Kong are with the CCS Grad-
uate School of Mobility, Korea Advanced Institute of Science and
Technology, 193, Munji-ro, Yuseong-gu, Daejeon, Republic of Korea
{donghee.paek, skong}@kaist.ac.kr
2Kevin Tirta Wijaya is with the Robotics Program, Korea Advanced
Institute of Science and Technology, 291, Daehak-ro, Yuseong-gu, Daejeon,
Republic of Korea kevin.tirta@kaist.ac.kr
an autonomous driving agent needs to be robust in various
driving conditions, the vulnerability of the existing camera-
based lane detection methods towards adverse lighting con-
ditions need to be resolved.
One possible solution to the problem is to use Light
Detection and Ranging (LiDAR) sensor. Since the LiDAR
sensor emits infrared signals, which is hardly interfered by
the visible light, adverse lighting conditions do not affect
its measurement capability significantly. Moreover, unlike
camera images, the LiDAR point cloud does not require a
bird’s eye view (BEV) projection for motion planning, which
often causes lane line distortions.
Despite the several advantage of the LiDAR sensor, only
a handful of studies have proposed deep learning-based
LiDAR lane detection methods. This is largely due to the
absence of publicly-available LiDAR lane detection datasets.
As seen in the deep camera-based lane detection field, the
majority of the lane detection networks are developed after
the publication of open lane detection datasets such as [7][8].
Recently, [9] opens a large-scale LiDAR lane detection
dataset, K-Lane, to the public, along with a segmentation-
based LiDAR lane detection network (LLDN) baseline. The
baseline consists of a projection network, a global feature
correlator (GFC), and a segmentation head, which generates
a feature map in the BEV format, extracting features via
global correlation, and predicting lanes per grid, respectively.
While a segmentation-based lane detection network is ca-
pable of producing lane detection of various shapes, it is
computationally expensive due to the need of processing
each grid of the final feature map through shared multi-layer
perceptron (MLP).
In this work, we propose a novel LiDAR lane detec-
tion network that is computationally efficient and especially
effective to the severe occlusion cases. We set the lane
detection problem as a row-wise prediction task instead of
a semantic segmentation task, so that the detection head of
the network performs row-wise MLP operations instead of
grid-wise MLP operations. The row-wise formulation leads
to a significantly less computational cost than the prior work
[9], with about 30% less GFLOPs.
Furthermore, we design an additional second-stage net-
work that refines the output feature map of the first-stage
network via correlation of features around the lane pro-
posals. The correlation process is implemented with the
dot-product attention mechanism, which allows the network
to exchange information between the features of the lane
proposals globally. As the lane lines often have some degree
of regularity in terms of shapes and distances between each
arXiv:2210.08745v1 [cs.CV] 17 Oct 2022