IV 2021 : 3D-DLAD-v3 third workshop on 3D Deep Learning for Autonomous Driving at Intelligent Vehicules 2021

 

CALL FOR PAPERS 3D-DLAD-v3 2021

 

3D-DLAD-v3 (third 3D Deep Learning for Autonomous Driving) workshop is the 6th workshop organized as part of DLAD workshop series. It is organized as a part of the flagship automotive conference Intelligent Vehicles https://2021.ieee-iv.org/.

 

Deep Learning has become a de-facto tool in Computer Vision and 3D processing with boosted performance and accuracy for diverse tasks such as object classification, detection, optical flow estimation, motion segmentation, mapping, etc. Lidar sensors are playing an important role in the development of Autonomous Vehicles as they overcome some of the many drawbacks of a camera based system, such as degraded performance under changes in illumination and weather conditions. In addition, Lidar sensors capture a wider field of view, and directly obtain 3D information. This is essential to assure the security of the different agents and obstacles in the scene. It is a computationally challenging task to process more than 100k points per scan in realtime within modern perception pipelines. Following the said motivations, finally to address the growing interest in deep representation learning for lidar point-clouds, in both academic as well as industrial research domains for autonomous driving, we invite submissions to the current workshop to disseminate the latest research.

 

We are soliciting contributions in deep learning on 3D data applied to autonomous driving in (but not limited to) the following topics. Please feel free to contact us if there are any questions.

 

TOPICS

Deep Learning for Lidar based clustering, road extraction object detection and/or tracking.

Deep Learning for Radar pointclouds

Deep Learning for TOF sensor-based driver monitoring

New lidar based technologies and sensors.

Deep Learning for Lidar localization, VSLAM, meshing, pointcloud inpainting

Deep Learning for Odometry and Map/HDmaps generation with Lidar cues.

Deep fusion of automotive sensors (Lidar, Camera, Radar).

Design of datasets and active learning methods for pointclouds

Synthetic Lidar sensors & Simulation-to-real transfer learning

Cross-modal feature extraction for Sparse output sensors like Lidar.

Generalization techniques for different Lidar sensors, multi-Lidar setup and point densities.

Lidar based maps, HDmaps, prior maps, occupancy grids

Real-time implementation on embedded platforms (Efficient design & hardware accelerators).

Challenges of deployment in a commercial system (Functional safety & High accuracy).

End to end learning of driving with Lidar information (Single model & modular end-to-end)

Deep learning for dense Lidar point cloud generation from sparse Lidars and other modalities

 

 

 

Location : Nagoya, Japan

Submission : 15th March 2021 (firm deadline, no extension)

Acceptance Notification : 25th April 2021

Workshop Date : 11th July 2021

 

Workshop Organizers:

B Ravi Kiran, Navya, France

Senthil Yogamani, Valeo Vision Systems, Ireland

Victor Vaquero, Research Engineer, IVEX.ai

Patrick Perez, Valeo.AI, France

Bharanidhar Duraisamy, Daimler, Germany

Dan Levi, GM, Israel

Abhinav Valada, University of Freiburg, Germany

Lars Kunze, Oxford University, UK

Markus Enzweiler, Daimler, Germany

Ahmad El Sallab, Valeo AI Research, Egypt

Sumanth Chennupati, Wyze Labs, USA

Stefan Milz, Spleenlab.ai , Germany

Hazem Rashed, Valeo AI Research, Egypt

Jean-Emmanuel Deschaud, MINES ParisTech, France

Kuo-Chin Lien, Appen USA

Naveen Shankar Nagaraja, BMW Group, Munich

Both comments and pings are currently closed.

Comments are closed.

Design by 2b Consult