Search
Publication Authors

Prof. Dr. Didier Stricker

Dr. Alain Pagani

Dr. Gerd Reis

Eric Thil

Keonna Cunningham

Dr. Oliver Wasenmüller

Dr. Gabriele Bleser
Dr. Bruno Mirbach

Dr. Jason Raphael Rambach

Dr. Bertram Taetz
Dr. Muhammad Zeshan Afzal

Sk Aziz Ali

Mhd Rashed Al Koutayni
Murad Almadani
Alaa Alshubbak
Yuriy Anisimov

Jilliam Maria Diaz Barros

Ramy Battrawy
Hammad Butt

Mahdi Chamseddine
Steve Dias da Cruz

Fangwen Shu

Torben Fetzer

Ahmet Firintepe
Sophie Folawiyo

David Michael Fürst
Kamalveerkaur Garewal

Christiano Couto Gava
Leif Eric Goebel

Tewodros Amberbir Habtegebrial
Simon Häring
Khurram Hashmi

Jigyasa Singh Katrolia

Andreas Kölsch
Onorina Kovalenko

Stephan Krauß
Paul Lesur

Muhammad Jameel Nawaz Malik
Michael Lorenz
Markus Miezal

Mina Ameli

Nareg Minaskan Karabid
Mohammad Minouei

Pramod Murthy

Mathias Musahl

Peter Neigel

Manthan Pancholi
Qinzhuan Qian

Engr. Kumail Raza
Dr. Nadia Robertini
María Alejandra Sánchez Marín
Dr. Kripasindhu Sarkar

Alexander Schäfer
Pascal Schneider

René Schuster

Mohamed Selim
Lukas Stefan Staecker

Dennis Stumpf

Yongzhi Su

Xiaoying Tan
Yaxu Xie

Dr. Vladislav Golyanik

Dr. Aditya Tewari

André Luiz Brandão
LiDAR-Flow: Dense Scene Flow Estimation from Sparse LiDAR and Stereo Images
LiDAR-Flow: Dense Scene Flow Estimation from Sparse LiDAR and Stereo Images
Ramy Battrawy, René Schuster, Oliver Wasenmüller, Qing Rao, Didier Stricker
International conference in robotics and intelligent systems. IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS-2019) IEEE/RSJ International Conference on Intelligent Robots and Systems November 4-8 MACAO, CHINA Macao SAR of China IEEE 2019 .
- Abstract:
- We propose a new approach called LiDAR-Flow to robustly estimate a dense scene flow by fusing a sparse LiDAR with stereo images. We take the advantage of the high accuracy of LiDAR to resolve the lack of information in some regions of stereo images due to textureless objects, shadows, ill-conditioned light environment and many more. Additionally, this fusion can overcome the difficulty of matching unstructured 3D points between LiDAR-only scans. Our LiDAR-Flow approach consists of three main steps; each of them exploits LiDAR measurements. First, we build strong seeds from LiDAR to enhance the robustness of matches between stereo images. The imagery part seeks the motion matches and increases the density of scene flow estimation. Then, a consistency check employs LiDAR seeds to remove the possible mismatches. Finally, LiDAR measurements constraint the edge-preserving interpolation method to fill the remaining gaps. In our evaluation we investigate the individual processing steps of our LiDAR-Flow approach and demonstrate the superior performance compared to image-only approach.