Back

Autonomous Vision Perceiving Systems Members Publications

Computer Vision Performance Evaluation

Computer vision performance evaluation
The KITTI Vision Benchmark Suite (left) and the MPI Sintel Benchmark (right) provide ground truth data and evaluation servers for benchmarking vision algorithms. So far, more than 400 methods have been evaluated on our benchmarks.

Members

Thumb ticker sm upper body tiny
Autonomous Vision, Perceiving Systems
Guest Scientist
Thumb ticker sm thumb jonas winter 1
Perceiving Systems
  • Doctoral Researcher
Thumb ticker sm headshot2021
Perceiving Systems
Director
Thumb ticker sm moritz menze
Perceiving Systems, Autonomous Vision
Thumb ticker sm danbutlerphoto
Perceiving Systems

Publications

Perceiving Systems Autonomous Vision Conference Paper Object Scene Flow for Autonomous Vehicles Menze, M., Geiger, A. In IEEE Conf. on Computer Vision and Pattern Recognition (CVPR) 2015, 3061-3070, IEEE, IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), June 2015
This paper proposes a novel model and dataset for 3D scene flow estimation with an application to autonomous driving. Taking advantage of the fact that outdoor scenes often decompose into a small number of independently moving objects, we represent each element in the scene by its rigid motion parameters and each superpixel by a 3D plane as well as an index to the corresponding object. This minimal representation increases robustness and leads to a discrete-continuous CRF where the data term decomposes into pairwise potentials between superpixels and objects. Moreover, our model intrinsically segments the scene into its constituting dynamic components. We demonstrate the performance of our model on existing benchmarks as well as a novel realistic dataset with scene flow ground truth. We obtain this dataset by annotating 400 dynamic scenes from the KITTI raw data collection using detailed 3D CAD models for all vehicles in motion. Our experiments also reveal novel challenges which can't be handled by existing methods.
pdf abstract suppmat DOI BibTeX

Perceiving Systems Autonomous Vision Article Vision meets Robotics: The KITTI Dataset Geiger, A., Lenz, P., Stiller, C., Urtasun, R. International Journal of Robotics Research, 32(11):1231 - 1237 , Sage Publishing, September 2013
We present a novel dataset captured from a VW station wagon for use in mobile robotics and autonomous driving research. In total, we recorded 6 hours of traffic scenarios at 10-100 Hz using a variety of sensor modalities such as high-resolution color and grayscale stereo cameras, a Velodyne 3D laser scanner and a high-precision GPS/IMU inertial navigation system. The scenarios are diverse, capturing real-world traffic situations and range from freeways over rural areas to inner-city scenes with many static and dynamic objects. Our data is calibrated, synchronized and timestamped, and we provide the rectified and raw image sequences. Our dataset also contains object labels in the form of 3D tracklets and we provide online benchmarks for stereo, optical flow, object detection and other tasks. This paper describes our recording platform, the data format and the utilities that we provide.
pdf DOI BibTeX

Perceiving Systems Conference Paper A naturalistic open source movie for optical flow evaluation Butler, D. J., Wulff, J., Stanley, G. B., Black, M. J. In European Conf. on Computer Vision (ECCV), 611-625, Part IV, LNCS 7577, (Editors: A. Fitzgibbon et al. (Eds.)), Springer-Verlag, October 2012
Ground truth optical flow is difficult to measure in real scenes with natural motion. As a result, optical flow data sets are restricted in terms of size, complexity, and diversity, making optical flow algorithms difficult to train and test on realistic data. We introduce a new optical flow data set derived from the open source 3D animated short film Sintel. This data set has important features not present in the popular Middlebury flow evaluation: long sequences, large motions, specular reflections, motion blur, defocus blur, and atmospheric effects. Because the graphics data that generated the movie is open source, we are able to render scenes under conditions of varying complexity to evaluate where existing flow algorithms fail. We evaluate several recent optical flow algorithms and find that current highly-ranked methods on the Middlebury evaluation have difficulty with this more complex data set suggesting further research on optical flow estimation is needed. To validate the use of synthetic data, we compare the image- and flow-statistics of Sintel to those of real films and videos and show that they are similar. The data set, metrics, and evaluation website are publicly available.
pdf dataset youtube talk supplemental material BibTeX

Perceiving Systems Conference Paper Lessons and insights from creating a synthetic optical flow benchmark Wulff, J., Butler, D. J., Stanley, G. B., Black, M. J. In ECCV Workshop on Unsolved Problems in Optical Flow and Stereo Estimation, 168-177, Part II, LNCS 7584, (Editors: A. Fusiello et al. (Eds.)), Springer-Verlag, October 2012 pdf dataset poster youtube BibTeX

Perceiving Systems Article A Database and Evaluation Methodology for Optical Flow Baker, S., Scharstein, D., Lewis, J. P., Roth, S., Black, M. J., Szeliski, R. International Journal of Computer Vision, 92(1):1-31, March 2011
The quantitative evaluation of optical flow algorithms by Barron et al. (1994) led to significant advances in performance. The challenges for optical flow algorithms today go beyond the datasets and evaluation methods proposed in that paper. Instead, they center on problems associated with complex natural scenes, including nonrigid motion, real sensor noise, and motion discontinuities. We propose a new set of benchmarks and evaluation methods for the next generation of optical flow algorithms. To that end, we contribute four types of data to test different aspects of optical flow algorithms: (1) sequences with nonrigid motion where the ground-truth flow is determined by tracking hidden fluorescent texture, (2) realistic synthetic sequences, (3) high frame-rate video used to study interpolation error, and (4) modified stereo sequences of static scenes. In addition to the average angular error used by Barron et al., we compute the absolute flow endpoint error, measures for frame interpolation error, improved statistics, and results at motion discontinuities and in textureless regions. In October 2007, we published the performance of several well-known methods on a preliminary version of our data to establish the current state of the art. We also made the data freely available on the web at http://vision.middlebury.edu/flow/ . Subsequently a number of researchers have uploaded their results to our website and published papers using the data. A significant improvement in performance has already been achieved. In this paper we analyze the results obtained to date and draw a large number of conclusions from them.
pdf pdf from publisher Middlebury Flow Evaluation Website DOI BibTeX

Perceiving Systems Conference Paper An additive latent feature model for transparent object recognition Fritz, M., Black, M., Bradski, G., Karayev, S., Darrell, T. In Advances in Neural Information Processing Systems 22, NIPS, 558-566, MIT Press, 2009 pdf slides BibTeX