基于小波和动态互补滤波的图像与事件融合方法

Image and event fusion method based on wavelet and dynamic complementary filtering

  • 摘要: 本文探讨了事件相机和传统帧相机数据的融合,引入了一种旨在增强复杂光照条件下成像质量的新型融合方法. 事件相机作为创新性视觉传感器的代表,以其高时域分辨率和极低功耗受到广泛关注. 然而,事件相机产生的数据往往存在噪声和特征丢失等问题. 相反,传统帧相机具有较高的空间分辨率,但在捕捉涉及快速运动或包含广泛动态范围场景时表现不佳. 基于此,本文引入了一种创新方法,将离散小波变换与动态增益互补滤波相结合,有效地融合图像和事件数据. 该方法首先通过计算图像的信息熵评估曝光水平. 然后,采用离散小波变换来从事件数据和帧图像数据中分离高频和低频细节. 随后,该方法应用动态增益互补滤波器来实现图像和事件数据的融合. 该方法的核心在于其能够自适应地平衡每种数据源的贡献,从而确保在不同条件下实现最佳的图像重建质量. 通过利用事件相机的高频时域信息和帧相机的高分辨率空间信息,该方法旨在克服每种传感器固有的局限性. 这种融合不仅可以减轻事件相机数据中遇到的噪声和特征丢失,还可以解决帧相机在捕捉高速运动和具有明显亮度变化的场景时的缺点. 该融合方法的有效性已经在HDR Hybrid Event-Frame数据集上进行了测试,该数据集具有高动态范围和复杂光照环境的真实场景. 实验结果展示了该方法在提高图像质量方面实现的改进. 与传统图像重建方法相比,本文方法在几个关键指标上表现出色:均方误差为0.0199,结构相似性指数为0.90,Q-score为6.07. 这些结果不仅验证了所提出的融合方法在改善在挑战性条件下成像质量方面的有效性,还凸显了整合不同类型视觉数据以实现更优重建结果的潜力.

     

    Abstract: This study investigates the fusion of data from event cameras and traditional frame cameras, introducing a novel fusion approach designed to enhance image quality under complex lighting conditions. Event cameras are an innovative class of vision sensors that are known for their high temporal resolution and minimal power consumption; however, their output is often plagued by noise and feature loss. Conversely, traditional frame cameras boast commendable spatial resolution; however, they struggle to capture fast-moving scenes or scenes with a vast dynamic range. To address these challenges, the study proposes an innovative method that combines discrete wavelet transform with dynamic gain complementary filtering to fuse image and event data. The process begins by evaluating the exposure level of incoming image frames using the image entropy of a metric. Following this assessment, the discrete wavelet transform segregates the high- and low-frequency components from the event stream and frame image data. A dynamic gain complementary filter is applied to seamlessly integrate image and event data. The proposed method capitalizes on its ability to balance the contribution of each data source adaptively, thereby ensuring optimal reconstruction quality under varying conditions. By leveraging the high-frequency temporal information from event cameras and the high-resolution spatial information from frame cameras, the proposed method attempts to overcome the limitations inherent in each type of sensor. This fusion not only mitigates the noise and feature loss in event camera data but also improves the capture of high-speed movements and scenes with significant brightness variations. The efficacy of this fusion approach was rigorously tested on the HDR Hybrid Event-Frame Dataset, which includes high dynamic range and complex lighting environments in real-world scenarios. The experimental results underscored a notable improvement in image quality, outperforming traditional image reconstruction methods. Our proposed approach has demonstrated superior performance, as evidenced by its scores on several key metrics: a mean squared error of 0.0199, a structural similarity index measure of 0.90, and a Q-score of 6.07. These results not only validate the effectiveness of the proposed fusion method in enhancing imaging quality under challenging conditions but also highlight the potential of integrating disparate types of visual data to achieve superior reconstruction outcomes.

     

/

返回文章
返回