• Nenhum resultado encontrado

6.3 Dense Stereo Matching using a Bayesian Neural Network

6.3.1 Comparison to the Deterministic Baseline

Table 6.3:Quantitative comparison of the deterministic and probabilistic variant of GC-Net.

The probabilistic variant proposed in this work is compared against the deterministic baseline presented by Kendall et al. (2017b) based on the disparity error metrics described in Section 5.4.1.

Pixel Error Rate [%] MAE [px]

RMSE [px]

τ= 1 τ= 3 τ= 5 Sceneflow

deterministic 10.4 5.5 4.3 1.8 10.5 probabilistic 12.1 5.6 3.7 0.9 3.7

KITTI

deterministic 64.3 33.1 22.2 6.1 14.7 probabilistic 64.9 33.2 20.0 3.7 7.0

Middlebury

deterministic 34.2 18.4 13.5 4.7 13.1 probabilistic 42.8 24.4 16.8 3.9 7.3

InStereo2K

deterministic 65.9 44.4 36.3 17.5 34.1 probabilistic 71.7 51.0 41.0 10.0 15.9

(a) Trained on Sceneflow

Pixel Error Rate [%] MAE [px]

RMSE [px]

τ= 1 τ= 3 τ= 5 Sceneflow

24.0 11.7 9.6 3.9 13.3

23.9 15.0 12.5 4.6 14.2

KITTI

35.8 7.5 4.0 1.5 4.2

34.3 8.8 5.2 1.7 5.4

Middlebury

32.4 20.0 16.0 4.5 10.9

33.9 20.4 16.2 4.4 11.0

InStereo2K

17.9 8.9 7.1 3.7 8.0

19.1 10.4 7.8 2.5 5.7

(b) Fine-tuned on InStereo2K

reference image. Under the assumption that potentially matching pixels located in such a region are assigned similar cost values, the correct match can only be identified by chance. However, averaging the disparities assigned to these potentially matching pixels, the risk of choosing an incorrect pixel far away from the correct solution and thus making a gross mistake can be reduced. Consequently, the superior performance of GC-Netprob with respect to weakly textured image regions can mainly be explained by the Monte Carlo sampling approach applied to approximate the distribution of the disparity prediction.

To obtain a deeper insight into the differences between the deterministic and the probabilistic variant and to better understand which effects arise from the Monte Carlo sampling and which are due to the stochastic nature of the employed BNN approach, GC-Netprob is additionally examined using a single Monte Carlo sample instead of 50 as proposed in Section 5.4.5. As shown in Table 6.4, this probabilistic variant using a single Monte Carlo sample estimates disparities with similar pixel error rates as the deterministic baseline. The MAE and the RMSE, however, are slightly improved, which implies that the stochastic nature of this approach itself is already beneficial to reduce the number of large errors. In comparison, GC-Netprob using 50 Monte Carlo samples demonstrates a further clear improvement with respect to the MAE and the RMSE, while the pixel error rates deteriorate, as already described above. This behaviour is also visible in Figure 6.10, which shows the distribution of the estimated disparities with respect to the corresponding ground truth values.

While an optimal dense stereo matching method would have an associated distribution that shows a thin diagonal line, the three evaluated variants show a relatively strong correlation in this sense but also a noticeable amount of larger deviations. On the one hand, GC-Netdet tends to rather under- estimate the true disparity, on the other hand, the probabilistic variant using a single Monte Carlo sample shows a more balanced distribution with a similar amount of deviations above and below the diagonal. More important is the fact that most deviations of the latter are clearly smaller than for

Reference Image

Reference DisparityRegionMaskErrorMap (Deterministic)

ErrorMap (Probabilistic)

Figure 6.9: Qualitative comparison of the disparity error. From left to right, the figure shows an example for the Sceneflow, Middlebury, InStereo2K and KITTI dataset, respectively. The refer- ence disparity maps shows small values in dark blue and large ones in red, while grey indicates pixels with unknown reference disparity. The region masks highlights regions that are especially challenging in the context of dense stereo matching, showing weakly textured areas in beige, occluded areas in red and pixels close to depth discontinuities in orange. The error maps show small values in white and large ones in dark red / black.

Table 6.4:Evaluation of the effects arising from the probabilistic convolutional layers and the Monte Carlo sampling. Three variants of GC-Net are compared, the deterministic baseline and the probabilistic variant presented in this work, using a single Monte Carlo sample or 50 samples (as proposed in Sec. 5.4.5), respectively. All variants are fine-tuned and tested on the InStereo2K dataset. Besides the disparity error, the training time in seconds per sample, the inference time in seconds per image and the memory footprint during training are listed for every variant.

Pixel Error Rate [%] MAE [px]

RMSE [px]

Training time [s]

Inference time [s]

Memory footprint [GB]

τ = 1 τ = 3 τ = 5

deterministic 17.9 8.9 7.1 3.7 8.0 0.41 5 5.0

probabilistic

(1 MC sample) 18.1 8.8 6.8 2.8 7.7 0.55 7 5.2

probabilistic

(50 MC samples) 19.1 10.4 7.8 2.5 5.7 0.55 351 5.2

the deterministic variant. This applies even more for GC-Netprob using 50 Monte Carlo samples, for which the distribution is more centralised along the diagonal. Consequently, the conclusion drawn based on the numeric results that both, the stochastic nature as well as the sampling approach used in the probabilistic variant, contribute to improve the average disparity error, are also supported by the analysis of the correlation between estimated and reference disparity. Moreover, these results indicate that the increased number of small errors, as expressed by the higher pixel error rates for small thresholds, is purely caused by the strategy employed to aggregate the results of the individual Monte Carlo samples. To overcome this drawback, it may be beneficial to determine the mode of the posterior distribution as final disparity estimate instead of using the mean. However, because the approximation of the posterior is computed based on samples, the result is typically a discrete distribution which further depends on hyper-parameter settings, such as the number and width of bins used for the discretisation, which may lead to different modes for the same set of samples. Thus, further investigations are required in this context that need to be carried out in future work.

To close the comparison between the deterministic baseline and the probabilistic variant of GC- Net presented in this work, some practical aspects are examined. For this purpose, the training time per sample (an image extract of size 512×128 pixels as described in Sec. 5.3.3), the inference time per image (with respect to the images of the InStereo2K dataset as specified in Tab. 5.1) and the memory footprint during training are determined as the average values over all training or test samples of the InStereo2K dataset, respectively, and are listed in Table 6.4. To minimise the risk of effects that arise from anything else than the probabilistic adaptation proposed in this work, both variants, GC-Netdetand GC-Netprob, are implemented using the same framework (Tensorflow 2.3) and are trained and tested on the same hardware (AMD Ryzen 7 2700X 8×3.7 GHz, 32 GB memory, Nvidia Titan V). Analysing the training time and the memory footprint, it can be seen that the stochastic forward pass together with the Bayesian backpropagation employed to optimise the weights of the variational distribution of GC-Netprobare computationally more complex and require more memory than their deterministic counter-parts. This becomes particularly evident when keeping in mind that the number of feature channels and thus the size of the intermediate feature

(a) Deterministic (b) Probabilistic (1 sample) (c) Probabilistic (50 samples)

Figure 6.10:Distribution of the estimated disparity with respect to the reference. Three variants of GC-Net are compared, the deterministic baseline and the probabilistic variant presented in this work, using a single Monte Carlo sample and 50 samples (as proposed in Sec. 5.4.5), respectively. All variants are fine-tuned and tested on the InStereo2K dataset. The logarithmic colour scale denotes the percentage of pixels assigned the respective estimated and ground truth disparity, while a thin diagonal line would indicate an optimal result.

maps is reduced by 25% for the probabilistic variant compared to the baseline (see Sec. 4.3.1). In turn, this reduction helps to mitigate the difference between both variants, leading to an increase of training time of about 34% and a similar memory footprint. Note that the training time and the memory footprint are identical for the two probabilistic variants, because they only differ in the number of Monte Carlo samples used to approximate the posterior distribution, which only affects the behaviour at test time and has no influence on the training characteristics.

In contrast, the comparison of the inference time reveals significant differences and demonstrates a drawback of approaches that are based on Monte Carlo sampling: While the difference between the deterministic and the probabilistic variant using a single Monte Carlo sample is comparable to the one on the training time, the inference time increases linearly with the number of samples drawn. Thus, the probabilistic variant using 50 Monte Carlo samples requires about six minutes to compute a single disparity map, while this task is completed by the baseline in about five seconds. It may be argued that such a long inference time limits the practical relevance of the proposed approach, but it should be noted that several options exist to optimise the inference time:

The computation of the individual samples can be fully parallelised or a more light-weight model could be used as basis (e.g. the adapted variant of GC-Net presented by Tulyakov et al. (2018)).

Alternatively, the number of Monte Carlo samples drawn can be reduced if uncertainty estimates of lower accuracy are acceptable (cf. Fig. 5.3). Finally, first approaches are presented in the literature that allow to estimate epistemic uncertainty without relying on any kind of sampling, which clearly reduces the computational over-head (Postels et al., 2019).