Deep stereo networks perform effectively when both training and testing data come from the same domain. However, their accuracy tends to drop significantly in efficiency-focused target scenarios due to domain shifts between training and testing datasets. These shifts often arise from differences in factors such as color, lighting, contrast, and texture. Additionally, the architecture of deep networks generally results in processing times that are unsuitable for real-time applications. To address these issues, this paper proposes a lightweight and robust stereo matching approach tailored for diverse driving environments. It leverages attention mechanisms for feature extraction and uses evolutionary algorithms for optimizing parameters. The method outperforms existing deep learning and traditional stereo matching techniques in terms of both processing speed and the percentage of bad pixels, as demonstrated on three challenging outdoor datasets: KITTI, HCI, and Driving Stereo. These results indicate that the proposed solution is highly effective for real-world applications where both precision and flexibility are essential.
Copyrights © 2025