基于多源传感器融合的抗天气干扰目标检测方法

Robust target detection method against weather interference based on multi-source sensor fusion

  • 摘要: 在复杂天气条件下,单一传感器的目标检测性能易受影响,难以满足自动驾驶与智能交通等应用对鲁棒性的需求。针对这一问题,本文提出一种基于4D毫米波雷达与激光雷达数据融合的目标检测方法——SeparateFusion。该方法利用两类传感器在感知能力上的互补性,通过神经网络模型实现多源信息的高效融合。首先,设计了三维早期融合模块GSE编码器,将两类点云映射至相同的柱状体视图,并分别对Lidar和Radar点云的几何信息与语义信息进行增强处理,再提取柱状体特征,实现多模态数据的底层融合。其次,提出二维特征提取增强模块BMM,在鸟瞰图(BEV)视图下引入MambaVisionMixer结构增强空间特征建模能力,并结合门控机制自适应过滤冗余信息,提升特征表达的有效性。在公开的多模态数据集View-of-Delft(VOD)上的实验表明,该方法在一般天气及雨雾等恶劣天气下的目标检测精度和稳定性均优于多种现有的目标检测方法,能够有效减弱天气干扰对检测性能的影响。研究结果验证了结合GSE编码器与BMM模块的SeparateFusion网络在多源传感器融合与抗干扰目标检测中的有效性,为全天候智能感知提供了一种可行方案。

     

    Abstract: Robust object detection under adverse weather conditions remains a pressing challenge in autonomous driving and intelligent transportation, as single-sensor systems are prone to performance degradation in rain, fog, or snow. To address this issue, we propose SeparateFusion, a novel multi-sensor fusion framework that integrates 4D millimeter-wave radar and LiDAR data through a deep neural network. By exploiting radar’s resilience to weather interference and LiDAR’s high spatial resolution, SeparateFusion delivers accurate and stable perception across diverse environments. The architecture comprises two key modules: the Geometry–Semantic Enhancement (GSE) encoder for early 3D fusion, and the BEV Feature Enhancement Module (BMM) for 2D feature refinement. In the first stage, LiDAR and radar point clouds are independently projected into a shared pillar grid, ensuring spatial alignment. The GSE encoder enhances geometric and semantic information of each modality separately—geometric features capture structural layouts from point coordinates, while semantic features encode attributes such as intensity, Doppler velocity, and reflectivity. Following this enhancement, pillar-level features are extracted, enabling early-stage multi-modal fusion that aligns and preserves modality-specific advantages. In the second stage, the fused features are transformed into a bird’s-eye view (BEV) representation. The BMM module then processes this representation using the MambaMixer structure to capture both local and long-range dependencies in the spatial domain. Additionally, a gating mechanism is applied to suppress redundant or noisy signals, allowing the network to focus on discriminative information for detection. This two-stage design provides a balance between fine-grained geometry–semantic modeling in 3D space and high-level spatial reasoning in BEV space, contributing to strong robustness against weather-related degradation. Extensive experiments on the View-of-Delft (VoD) dataset show that our method consistently outperforms both state-of-the-art single-sensor detectors and existing multi-sensor fusion approaches, achieving 70.8% mean Average Precision (mAP) across the entire test area and 85.46% within the driving corridor, demonstrating notable gains in both global and lane-focused detection scenarios. Furthermore, additional evaluations on a fog-simulation dataset confirm that SeparateFusion maintains clear advantages over previous methods in low-visibility conditions, indicating strong generalization capability. Ablation studies further validate the contributions of the GSE encoder and BMM module, showing that removing either component results in a significant drop in detection accuracy, which highlights the complementary nature of early 3D geometry–semantic enhancement and later-stage BEV feature gating. In summary, SeparateFusion introduces a structured two-stage fusion approach for integrating radar and LiDAR data, incorporating both early geometry–semantic enhancement and later-stage BEV refinement with adaptive gating. The method achieves significant improvements over both powerful single-sensor and existing fusion-based object detection methods under challenging weather, providing a promising foundation for next-generation all-weather intelligent perception systems that must operate reliably in safety-critical scenarios.

     

/

返回文章
返回