Abstract
In severe foggy weather, the visibility of the driving environment is extremely low. This seriously affects the driver's vision and safety. To address the challenges of manual driving in severe foggy weather, this paper proposes a foggy image prediction algorithm for road scenes based on Transformer. The aim is to enhance the visual perception and prediction capabilities of autonomous driving systems under adverse weather conditions. Leveraging the long-range dependency modeling capability of Transformer. We adopt a Transformer improved by Taylor-expanded multi-head self-attention. The Taylor series expansion of the softmax function significantly reduces computational costs. Additionally, a multi-branch architecture with multi-scale patch embedding is introduced in the Transformer. This embeds features through overlapping deformable convolutions of different scales. These improvements enable our algorithm to achieve good image prediction results with relatively low computational performance requirements. The performance of the proposed method was tested on three sets of custom haze road scene image datasets, with the experimental results showing a PSNR of 12.9836 and an SSIM of 0.6278. The experimental results indicate that our method can effectively predict real images under hazy weather, improving visibility in haze conditions. This addresses the serious driving safety issues during heavy haze and contributes to the development of autonomous driving technology.