Depth estimation is a critical task for autonomous driving. It’s necessary to estimate the distance to cars, pedestrians, bicycles, animals, and obstacles.
The popular way to estimate depth is LiDAR. However, the hardware price is high, LiDAR is sensitive to rain and snow, so there is a cheaper alternative: depth estimation with a stereo camera. This method is also called stereo matching.
In general, the idea behind the stereo matching is pretty straightforward.
We have two cameras with collinear optical axes, which have a horizontal displacement only. We can find a corresponding pixel on the right camera frame for every pixel on the left camera frame. We can estimate the point’s depth if we know the distance between pixels related to this point on the left and right frames. As one can see from the depth of the point is inversely proportional to the distance between the images of this point. This distance is called a disparity:
Classical Approaches For Stereo Matching
A classical approach to disparity estimation consists of the following steps:
- Extract features from the image to get more valuable information than raw color intensities and improve the point’s matching.
- Construct the cost volume to estimate how the left and the right feature maps match each other on different disparity levels. For example, we can use absolute intensity differences or cross-correlation.
- Calculate the disparity from the cost volume using the disparity computation module. For example, a brute-force algorithm can look for a disparity level at which left and right feature maps match best.
- Refine the disparity if the initially predicted disparity map is too coarse.
You can refer to the survey by Scharstein and Szeliski that provides an overview of the pre-deep-learning methods for stereo matching.
Datasets for stereo matching
Data plays a crucial role in computer vision, so a few words about the datasets for stereo matching:
- Middlebury dataset is one of the first datasets for this task. It contains 33 static indoor scenes. The paper describes the dataset acquisition.
- The large synthetic SceneFlow dataset contains more than 39 thousand stereo frames in 960×540 resolution. Many authors used it for the pretraining of stereo-matching neural networks. There is a detailed description of the dataset in the paper.
- KITTI is a popular benchmark for an autonomous driving scenario. Data is collected from a moving vehicle, and a LiDAR estimates depth.
Deep Learning-Based Approaches for Stereo Matching
Nowadays, deep learning methods combine many of the steps described above into an end-to-end algorithm. A very early example is GCNet. StereoNet and PSMNet follow the same idea. We can focus deeply on the PSMNet approach.
You can see the list of its building blocks in Figure 1.
The approach uses backbones with shared weights to extract features from both left and right images. Interestingly, the authors adopt the Spatial Pyramid Pooling block from semantic segmentation nets to combine features from different scales.
Next, the authors concatenate features from the left image with features from the right image shifted horizontally by pixels. They use in the range and combine the result into a 4D volume.
As the next step, 3D convolution encoder-decoder architecture improves and reduces cost volume. The final layer is a 3D convolution that produces a feature map with size.
Finally, the authors apply the soft argmin function to predict the disparity. Every channel of the output feature map represents a different disparity. The authors normalize the probability volume across the disparity dimension with the softmax operation, . They combine the disparities, , with the weights corresponding to their probabilities. The formula below represents this operation:
If we look at the KITTI benchmark results in Figure 3 (dated by the middle of August 2020), we can notice that the top stereo-matching approaches are slow. The fastest work from the top-10 on the KITTI 2015 benchmark provides only 3.3 frames per second using GPU. The reason is the top models use 3D convolutions to improve the quality, but the cost is the algorithm’s speed.
There are stereo-matching methods that use a correlation layer (MADNet, DispNetC), but they lose the battle for the quality to 3D convolution-based methods.
One of the recent approaches trying to achieve state-of-the-art results while being significantly faster than other methods is AANet.
The general method’s architecture is the following:
- A shared network (feature extractor) extracts feature pyramids from the left and right images: and , correspondingly.
denotes the number of scales, is the scale index, and represents the largest scale.
- A correlation layer constructs multi-scale 3D cost volumes:
- Several Adaptive Aggregation Modules aggregate cost volumes. These modules replace 3D encoder-decoder network used earlier in [GCNet, StereoNet, PSMNet],
- At the final step, the refinement module upsamples low-resolution disparity prediction to the original resolution. AANet uses two refinement modules proposed in StereoDRNet.
The figure below depicts AANet architecture.
Adaptive Aggregation Module (AAModule)
Let’s focus on the Adaptive Aggregation Module (AAModule) that is the major novelty of the paper.
It consists of two major parts:
- Intra-Scale Aggregation (ISA)
- Cross-Scale Aggregation (CSA)
Deep learning-based stereo matching methods usually perform window-based cost aggregation:
where is an aggregated cost at pixel for disparity candidate , pixel belongs to the neighbors of , is the aggregation weight, and is the raw matching cost at pixel for disparity candidate .
However, it should work well only when the disparity is continuous, but the object boundaries violate this assumption.
That’s why we should design the weighting function carefully and eliminate the influence of pixels on the disparity discontinuities.
This problem is called an edge-fattening issue.
The authors of AANet propose the adaptive sampling method instead of regular sampling used in simple 2D and 3D convolutions to cope with the edge-fattening problem. The idea is identical to deformable convolutions.
The cost aggregation strategy for AANet learns an additional offset to the regular sampling location:
where is the number of sampling points, is the aggregation weight for -th point, is the fixed offset to , is the learnable offset.
The authors also make convolution weights content-adaptive with the modulation mechanism from Deformable ConvNets v2: More Deformable, Better Results.
The final cost aggregation formulation looks as follows:
where is a content-specific weight at pixel .
The only difference with deformable convolution is the number of groups. Feature map consists of G groups, and each group of channels has its own offsets and weights , unlike the original implementation of deformable convolution, where all channels in the feature map share the same offsets and weights.
The authors of AANet propose to combine cost aggregation on multiple scales. Cost aggregation on the high resolution is required to cope with fine details, while cost aggregation on the coarse resolution is beneficial for low-texture and textureless regions.
AANet combines the features in the following way:
where is the resulting cost volume after cross-scale cost aggregation, is the aggregated cost volume at scale , is a function to combine cost volumes with different shapes.
The idea here is identical to the concept of HRNet and is adopted from that paper.
where is an identity function, convolutions with are used to downsample a feature map times, is a bilinear upsampling.
The only difference with HRNet is the number of channels for feature maps with smaller resolutions. HRNet has more channels in low-resolution feature maps than high-resolution ones as feature extractors typically provide them this way. In contrast, AANet has fewer channels in coarse-scale feature maps than in high-scale ones. The reason is that the number of channels represents the number of disparity candidates, so coarse representation needs fewer disparity levels.
The authors make an ablation study to prove the importance of the proposed modules. They remove ISA and CSA modules from AANet, and the quality worsens.
They also replaced 3D convolution encoder-decoders with ISA & CSA modules into several popular approaches and achieved a significant speedup and lower memory consumption.
More than that, quality becomes better for StereoNet, GCNet, and PSMNet. There is only a negligible quality drop for GANet.
The authors denote modified architectures with AA suffix.
The authors modified GANet-AA’s refinement modules with hourglass networks with deformable convolutions and call the resulting model as AANet+.
They compare AANet and AANet+ with the counterparts on the SceneFlow (fig. 7) and KITTI (fig. 8) datasets.
The authors trained models with SceneFlow and KITTI datasets; however, the neural network shows good results on the other domain. For example, you can see the results on the sample from the Middlebury dataset on the fig. 10:
|Left image||Right image|
|Groundtruth disparity||AANet prediction||AANet+ prediction|
The authors provide the code for training, evaluation, and inference in the repository.
Leave a Reply