Back

Autonomous Learning Members Publications

Combinatorial Optimization as a Layer / Blackbox Differentiation

Schema depicting an architecture with an embedded solver on the top. Thanks to our surrogate gradient, training can be done as usual with stochastic gradient descent. The bottom row illustrates the applications we tackled so far: keypoint correspondence in pairs of images, directly optimizing ranked-ranked-based losses for improved image retrieval, and embedded planning for stronger generalization in reinforcement learning.
Optimizing Rank-based Metrics Deep Graph Matching CombOptNet

Members

Empirical Inference, Autonomous Learning
Senior Research Scientist
no image
Autonomous Learning
Autonomous Learning
  • Doctoral Researcher
Autonomous Learning
Empirical Inference
  • Guest Scientist

Publications

Autonomous Learning Conference Paper CombOptNet: Fit the Right NP-Hard Problem by Learning Integer Programming Constraints Paulus, A., Rolínek, M., Musil, V., Amos, B., Martius, G. In Proceedings of the 38th International Conference on Machine Learning, 139:8443-8453, Proceedings of Machine Learning Research, (Editors: Meila, Marina and Zhang, Tong), PMLR, The Thirty-eighth International Conference on Machine Learning (ICML), July 2021
Bridging logical and algorithmic reasoning with modern machine learning techniques is a fundamental challenge with potentially transformative impact. On the algorithmic side, many NP-hard problems can be expressed as integer programs, in which the constraints play the role of their ``combinatorial specification.'' In this work, we aim to integrate integer programming solvers into neural network architectures as layers capable of learning both the cost terms and the constraints. The resulting end-to-end trainable architectures jointly extract features from raw data and solve a suitable (learned) combinatorial problem with state-of-the-art integer programming solvers. We demonstrate the potential of such layers with an extensive performance analysis on synthetic data and with a demonstration on a competitive computer vision keypoint matching benchmark.
Arxiv Code Pdf Spotlight video @ ICML 2021 Poster @ ICML 2021 URL BibTeX

Autonomous Learning Conference Paper Neuro-algorithmic Policies Enable Fast Combinatorial Generalization Vlastelica, M., Rolinek, M., Martius, G. In Proceedings of the 2021 International Conference on Machine Learning (ICML), The Thirty-eighth International Conference on Machine Learning (ICML), July 2021
Although model-based and model-free approa\-ches to learning the control of systems have achieved impressive results on standard benchmarks, generalization to task variations is still lacking. Recent results suggest that generalization for standard architectures improves only after obtaining exhaustive amounts of data. We give evidence that generalization capabilities are in many cases bottlenecked by the inability to generalize on the combinatorial aspects of the problem. We show that, for a certain subclass of the MDP framework, this can be alleviated by a neuro-algorithmic policy architecture that embeds a time-dependent shortest path solver in a deep neural network. Trained end-to-end via blackbox-differentiation, this method leads to considerable improvement in generalization capabilities in the low-data regime.
arXiv Spotlight PDF BibTeX

Autonomous Learning Conference Paper Deep Graph Matching via Blackbox Differentiation of Combinatorial Solvers Rolínek, M., Swoboda, P., Zietlow, D., Paulus, A., Musil, V., Martius, G. In Computer Vision – ECCV 2020, 28:407-424, Lecture Notes in Computer Science, 12373, (Editors: Vedaldi, Andrea and Bischof, Horst and Brox, Thomas and Frahm, Jan-Michael), Springer, Cham, 16th European Conference on Computer Vision (ECCV 2020) , August 2020 (Published)
Building on recent progress at the intersection of combinatorial optimization and deep learning, we propose an end-to-end trainable architecture for deep graph matching that contains unmodified combinatorial solvers. Using the presence of heavily optimized combinatorial solvers together with some improvements in architecture design, we advance state-of-the-art on deep graph matching benchmarks for keypoint correspondence. In addition, we highlight the conceptual advantages of incorporating solvers into deep learning architectures, such as the possibility of post-processing with a strong multi-graph matching solver or the indifference to changes in the training setting. Finally, we propose two new challenging experimental setups.
Code Arxiv Long Spotlight Short Spotlight pdf DOI BibTeX

Autonomous Learning Conference Paper Optimizing Rank-based Metrics with Blackbox Differentiation Rolínek, M., Musil, V., Paulus, A., Vlastelica, M., Michaelis, C., Martius, G. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), 7617 - 7627, IEEE, Piscataway, NJ, IEEE/CVF International Conference on Computer Vision and Pattern Recognition (CVPR 2020), June 2020, Best paper nomination (Published)
Rank-based metrics are some of the most widely used criteria for performance evaluation of computer vision models. Despite years of effort, direct optimization for these metrics remains a challenge due to their non-differentiable and non-decomposable nature. We present an efficient, theoretically sound, and general method for differentiating rank-based metrics with mini-batch gradient descent. In addition, we address optimization instability and sparsity of the supervision signal that both arise from using rank-based metrics as optimization targets. Resulting losses based on recall and Average Precision are applied to image retrieval and object detection tasks. We obtain performance that is competitive with state-of-the-art on standard image retrieval datasets and consistently improve performance of near state-of-the-art object detectors.
Paper @ CVPR2020 Long Oral Short Oral Arxiv Code Pdf DOI URL BibTeX

Autonomous Learning Conference Paper Differentiation of Blackbox Combinatorial Solvers Vlastelica*, M., Paulus*, A., Musil, V., Martius, G., Rolínek, M. In International Conference on Learning Representations, ICLR’20, May 2020, *Equal Contribution Arxiv Code pdf URL BibTeX