News Archive
  • October 2024
  • September 2024
  • July 2024
  • June 2024
  • May 2024
  • April 2024
  • March 2024
  • February 2024
  • January 2024
  • December 2023
  • November 2023
  • October 2023

News

Two papers accepted at BMVC 2017

Our research group got two papers accepted at British Machine Vision Conference (BMVC) 2017:

Fast dense feature extraction with convolutional neural networks that have pooling or striding layers by Christian Bailer, Tewodros Habtegebrial, Kiran Varanasi and Didier Stricker.

Introduction to Coherent Depth Fields for Dense Monocular Surface Recovery by Vladislav Golyanik, Torben Fetzer and Didier Stricker.

Both papers will be soon available on our webpage.

The BMVC is one of the major international conferences on computer vision and related areas. It is organised by the British Machine Vision Association (BMVA) and will take place from 4th to 7th September 2017 in the Imperial College London.

GPU + Real-time Paper Accepted@ ICIP 2017

The paper analyses applicability of current GPUs in real-time applications with hard real-time
constraints and provides a solution for a single dedicated GPU. This work was done in collaboration with
Max Planck Institute for Software Systems (MPI-SWS).

Towards Scheduling Hard Real-Time Image Processing Tasks on a Single GPU
Vladislav Golyanik, Mitra Nasri, Didier Stricker

The IEEE International Conference on Image Processing (ICIP) 2017 will feature world-class speakers, tutorials, and industry sessions, and creates an excellent forum to foster innovation and entrepreneurship, and network with the brightest minds in academia and industry that are working in this field. The conference  will be held at the China National Convention Center in Beijing, China from 17-20 September 2017.

Three papers accepted at ISMAR 2017

Our research group got three papers accepted at 16th IEEE International Symposium on Mixed and Augmented Reality (ISMAR) 2017:

Paper #1: A Probablistic Combination of CNN and RNN Estimates for Hand Gesture Based Interaction in Car by Aditya Tewari, Bertram Taetz, Frederic Grandidier, Didier Stricker

Paper #2: Augmented Things: Enhancing AR Applications leveraging the Internet of Things and Universal 3D Object Tracking by Jason Rambach, Alain Pagani, Didier Stricker

Paper #3: Fusion of unsynchronized optical tracker and inertial sensor in EKF framework for in-car Augmented Reality delay reduction by Jason Rambach, Alain Pagani, Sebastian Lampe, Ruben Reiser, Manthan Pancholi, Didier Stricker

The IEEE ISMAR is the leading international academic conference in the fields of Augmented Reality and Mixed Reality. The congress jointly organized by Centrale Nantes and Inria will be held at La Cité, Nantes Events Center in Nantes (France), on October 9–13, 2017.

Optical-flow & Deep-Learning paper @ CVPR 2017

Our paper on deep-learning and optical flow will be presented @ CVPR 2017.

Please find the the details here:

CNN-based Patch Matching for Optical Flow with Thresholded Hinge Embedding Loss

CVPR is the premier annual computer vision event comprising the main conference and several co-located workshops and short courses. With its high quality and low cost, it provides an exceptional value for students, academics and industry researchers.

 

New project DAKARA started

Low-cost, extremely small and energy-efficient: New camera matrix provides precise depth images for automated driving and industrial applications

It is no bigger than a 1-cent piece: a new, ultra-compact and adaptive camera, which, in addition to pictures, also delivers accurate depth information in real-time. As a novel sensor, it offers a wide range of applications in the area of automated driving or manual assembly processes. The system is developed in the project “DAKARA” – design and application of an ultra-compact, energy-efficient and configurable camera matrix for spatial analysis. In the project funded by the German Federal Ministry of Education and Research (BMBF), five partners from industry and research work together.

The camera matrix consists of sixteen square-shaped single cameras, which together function not only as an imaging device, but also as a depth camera. They are arranged on a so-called “wafer”, an approximately 1 millimeter thick structure of polycrystalline semiconductor blanks. The new camera technology comes from AMS Sensors Germany GmbH. With their technology, the camera will not be larger than ten by ten millimeters and about three millimeters thick.

Stephan Voltz, CEO of AMS Sensors Germany, explains the function of the new technology: “The structure as a camera matrix captures the scene from sixteen slightly displaced perspectives in order to calculate the scene geometry (a depth image) using the light field principle. Because such calculations are very high-intensity, an efficient processor is embedded directly into the periphery of the camera matrix to enable real-time applications. ”

The Augmented Vision (AV) research area at the German Research Center for Artificial Intelligence (DFKI) in Kaiserslautern is developing the algorithms for depth image calculations. They are executed in a real-time manner, directly in the camera system. In addition, various applications can be run on the embedded chip for further processing of the generated image data.

Professor Didier Stricker, Head of the Augmented Vision department at DFKI, said: “The depth information provided by the camera system alongside the color information allows a wide range of new applications. The ultra-compact design makes it possible to integrate the new camera into very small filigree components and use it as a non-contact sensor “.

The structure of the camera matrix is reconfigurable. As a result, a more specific layout can be used, depending on the application, for example a different arrangement of the camera matrix in L-shape. The depth image computation can also be adapted to specific requirements for the depth information.

Cameras that provide depth information already exist. However, these send out light to calculate the depth. Disadvantages are the high energy consumption, the large design and high costs. Other passive systems have much lower energy consumption, but are still in the research stage and generally have large shapes and low image rates.

“The DAKARA camera matrix will be the first passive system to provide both color and depth images in real-time, with high image rates, adaptive features, low energy consumption, and a very compact design,” said Oliver Wasenmüller, DFKI Project Manager and co-initiator of the project. The new system is to be used by well-known users from different domains.

Two application scenarios are used to check and demonstrate the developments of DAKARA: A rear-view camera from the partner ADASENS Automotive GmbH is designed to better interpret the rear vehicle environment. This means that finer structures such as curbs or posts can also be detected during automated parking. In addition, the system is designed to recognize people and send warning signals in an emergency. As a result, a tremendous increase in the safety of automated or partially automated driving can be expected. The Bosch Rexroth AG and DFKI (department Innovative Factory Systems) with the Living Lab SmartFactory KL e.V. will be installing a manual assembly process for the workplace assistant. The camera matrix captures objects as well as the hands of the worker by means of the algorithms of the partner CanControls GmbH. The particular challenge is to clearly distinguish objects such as tools or workpieces from the operator’s hands. The depth information of the DAKARA camera is intended to make this separation easier and more precise.

In the next three years the new camera matrix is to be designed, developed and extensively tested in the mentioned scenarios. A first prototype is to be realized by late summer 2018.

The project “DAKARA” is funded by the Federal Ministry of Education and Research (BMBF) within the framework of the “Photonics Research Germany – Digital Optics” program. The project volume totals 3.8 million euros; almost half of which is generated by the industry partners involved.

For more details, please visit the project page: DAKARA

DAKARA Kick-Off at AMS in Nürnberg. Persons (l.t.r.): Stephan Voltz (AMS), Yuriy Anisimov (DFKI), Benjamin Klimczak (CanControls), Christoph Schiferle (Bosch Rexroth), Alwin Tuschmann (AMS), Oliver Wasenmüller (DFKI), Florian Baumann (ADASENS), Didier Stricker (DFKI), Ariadna Bartra (ADASENS), Martin Geist (Bosch Rexroth), Matthias Möller (Bosch Rexroth), Patrick Bertram (DFKI), Fabian Quint (DFKI)

DAKARA Kick-Off at AMS in Nürnberg.
Persons (l.t.r.): Stephan Voltz (AMS), Yuriy Anisimov (DFKI), Benjamin Klimczak (CanControls), Christoph Schiferle (Bosch Rexroth), Alwin Tuschmann (AMS), Oliver Wasenmüller (DFKI), Florian Baumann (ADASENS), Didier Stricker (DFKI), Ariadna Bartra (ADASENS), Martin Geist (Bosch Rexroth), Matthias Möller (Bosch Rexroth), Patrick Bertram (DFKI), Fabian Quint (DFKI)

AV-Forscher auf 4. Software Campus Summit feierlich verabschiedet

Nach dem erfolgreichen Abschluss seines Software Campus Projekts BodyAnalyzer wurde Oliver Wasenmüller vom Forschungsbereich Erweiterte Realität des DFKI Kaiserslautern am 27. März 2017 feierlich als Absolvent des Software Campus Programms verabschiedet. Oliver Wasenmüller stellte sein, in Kooperation mit der SIEMENS AG entstandenes Projekt BodyAnalyzer, Erfassung und Analyse von 3D Körpermodellen, in einer Abschlusspräsentation auf dem 4. Software Campus Summit in Berlin vor, bevor er von Prof. Dr. Wahlster und Dr. Harald Schöning, Leiter Forschung der Software AG, verabschiedet wurde.

Der Software Campus sucht herausragende Doktorandinnen und Doktoranden der Informatik sowie informatiknaher Disziplinen mit großem Interesse an Führungsaufgaben in der Wirtschaft oder Unternehmensgründung. Insgesamt 18 Partner aus Industrie und Forschung unterstützen das Führungskräfteentwicklungsprogramm, das vom Bundesministerium für Bildung und Forschung (BMBF) gefördert wird. Die Teilnehmerinnen und Teilnehmer setzen im Software Campus ihr eigenes IT-Projekt um. Sie managen den gesamten Prozess ihres IT-Projekts selbständig mit Unterstützung der Forschungs- und Industriepartner: von der Projektplanung über die Beantragung finanzieller Mittel und das Management, die Koordination des eigenen Forscher-Teams bis zum Abschluss des Vorhabens. Jedes Projekt wird mit bis zu 100.000 Euro über die Projektlaufzeit gefördert. Darüber hinaus wird im Rahmen des Software Campus das vorhandene Potenzial der Teilnehmerinnen und Teilnehmer gezielt gefördert. Die Industriepartner des Software Campus bringen in das Programm ihre besten Führungskräftetrainings ein. Die Teilnehmerinnen und Teilnehmer können in sechs Modulen ihre Führungs-, Methoden- sowie Sozial- und Selbstkompetenzen weiterentwickeln.

Wir gratulieren Herrn Wasenmüller recht herzlich zur Aufnahme in den Software Campus sowie zum erfolgreichen Abschluss seines eigenen Forschungsprojekts BodyAnalyzer.

Oliver Wasenmüller (li.) erhält die Urkunde von Prof. Dr. W. Wahlster

Oliver Wasenmüller (li.) erhält die Abschluss-Urkunde von Prof. Dr. W. Wahlster (Quelle: Software Campus/M. Braumann)

Augmented Things at the CeBIT 2017

From 20 to 24 March 2017 we will be presenting our demonstrator “Augmented Things” in hall 6, booth B48.

Although machines are designed to simplify our lives, it is often difficult to understand how to use them at first glance. Their use would be much simplified if they would explain us themselves how they are meant to be used. This is now possible using Augmented Reality with the “Augmented Things” concept. Thanks to a dedicated object recognition and tracking algorithm, any object can deliver instantaneously and on demand its instructions manual to the user. To this aim, the user needs only one single application on a tablet or a smartphone and can scan virtually any object to access meta-information in an AR viewer. This tool is easily scalable to many new objects and is an extremely intuitive and natural way of seeking additional information.

imge_1

We are looking forward to welcoming you at CeBIT 2017!

Two papers accepted at WACV 2017

Two papers have been accepted at IEEE Winter Conference on Applications of Computer VisionWACV 2017

Accurate 3D Reconstruction of Dynamic Scenes from Monocular Image Sequences with Severe Occlusions
Vladislav Golyanik, Torben Fetzer, Didier Stricker

Dense Batch Non-Rigid Structure from Motion in a Second
Vladislav Golyanik, Didier Stricker

The papers will be shortly available on our web page.

The WACV is one of the most renowned conferences covering all areas of Computer Vision. It will be held March 27-30, 2017 at the Hyatt Vineyard Creek Hotel, Santa Rosa, CA, USA.

The program schedule is available at http://pamitc.org/wacv2017/program-schedule/.