Giter VIP home page Giter VIP logo

aydnzn / enhancing-lidar-based-3d-object-detection Goto Github PK

View Code? Open in Web Editor NEW
6.0 2.0 1.0 15.62 MB

This repository contains a Master's Thesis on enhancing LiDAR-based 3D Object Detection in autonomous vehicles using synthetic data. The research explores the use of the Ansys AVxcelerate Sensors Simulator (AVX) to create synthetic point clouds. It includes a hybrid training approach, experimental design, results, and related scripts.

License: MIT License

Python 99.77% Shell 0.23%
autonomous-driving kitti-dataset lidar-point-cloud pointpillars synthetic-data carmaker 3d-object-detection deep-learning

enhancing-lidar-based-3d-object-detection's Introduction

Enhancing LiDAR-based 3D Object Detection through Simulation

This repository is dedicated to the Master's Thesis in Electrical Engineering and Information Technology at the Institute for Measurement Systems and Sensor Technology, Technische Universität München.

This Master's thesis investigates the enhancement of LiDAR-based 3D Object Detection algorithms for autonomous vehicles, using synthetic point cloud data generated from the Ansys AVxcelerate CarMaker Co-Simulation process. The study focuses on integrating and aligning synthetic and real-world data, and applying fine-tuning techniques within the Pointpillars network to optimize the model. The research reveals challenges in ensuring model generalization across different data types, especially when identifying complex entities like pedestrians. The study indicates that a balanced combination of synthetic and real-world data yields promising results. Additionally, a hybrid training approach, consisting of initial pre-training with synthetic data followed by fine-tuning with real-world data, exhibits potential, particularly under conditions of real-world data scarcity. This study thus provides valuable insights to guide future improvements in the training and testing methodologies for autonomous driving systems.

Problem Statement

Despite the accuracy of depth perception provided by LiDAR technology, training deep learning algorithms for LiDAR-based object detection poses a significant challenge due to the scarcity of large-scale annotated data.

Synthetic data generation through simulation software is a potential solution, but often fails to accurately mimic real-world sensory data due to a reliance on handcrafted 3D assets and simplified physics, creating a 'synthetic-to-real gap'. Furthermore, models trained solely on synthetic data may not perform well in real-world scenarios due to data distribution differences.

As part of this research, I'll be investigating how to bridge this gap using the Ansys AVxcelerate Sensors Simulator (AVX). AVX offers a virtual testing environment for sensors used in autonomous vehicles, potentially helping bridge the synthetic-to-real gap. However, the accuracy of this simulator in replicating real-world data and its impact on the performance of algorithms needs to be critically evaluated.

Objectives

The main objectives are outlined below.

  • Generate a replica of the renowned KITTI dataset using synthetic data from the Velodyne HDL-64E LiDAR model in Ansys AVxcelerate Sensors Simulator (AVX), in co-simulation with CarMaker software.
  • Apply bounding box extraction algorithms to synthetic point clouds and create KITTI-compatible labels.
  • Investigate the potential of synthetic data in enhancing the performance of object detection algorithms.
  • Compare performance metrics of models trained on diverse data types (synthetic versus real-world).
  • Evaluate the influence of modifying the ratio of synthetic to real-world data on the performance.
  • Assess the viability and efficacy of a hybrid training strategy involving pre-training on synthetic data with subsequent fine-tuning on real-world data.
  • Analyze the impact of pre-training duration on the optimization of model parameters.
  • Conduct a detailed qualitative analysis of the trained networks.

Through these objectives, I aim to provide valuable insights into the benefits and challenges of using synthetic data in training object detection algorithms for autonomous vehicles.

Contents

  • Thesis: This is my Master's thesis PDF document.
  • Methodology: This section outlines the research methodology, emphasizing the LiDAR sensor modeling. It provides a detailed explanation of the Ansys AVxcelerate CarMaker Co-Simulation process, the processing of simulation outputs, and how simulated scenarios are scaled.
  • Experimental_Design: This section describes the experimental design, specifying the datasets used, network settings, evaluation metrics, and the adaptation of KITTI difficulty levels for synthetic dataset evaluation. It also presents the different experiments carried out.
  • Results: This section delves into the results from the experiments. It provides a quantitative analysis of the results from each experiment, along with an assessment of pre-training and training duration impact on the Average Precision for 3D object detection (AP 3D) scores. It also includes a qualitative analysis on the AVX test set and KITTI test set.
  • Python_scripts: These are the Python scripts required to process the synthetic point clouds to create the KITTI labels, calibration files, etc. See the README.md for usage instructions.
  • VM_scripts: These are scripts required for training, evaluation, data preparation, and point cloud visualization and need to be transferred to the virtual machine. Refer to the README.md for usage instructions.
  • cfgs: These are configuration files required for training and evaluation according to OpenPCDet.
  • kitti_models: These are the Pointpillars network models required for training and evaluation, according to OpenPCDet.
  • docs: These are some necessary documents for the other README's I have created.
  • RUN: This README explains how to run the whole framework. It includes creating the synthetic point clouds, their labels, preparing them for training, conducting training and evaluation, and visualization instructions.

enhancing-lidar-based-3d-object-detection's People

Contributors

aydnzn avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

Forkers

junyang0412

enhancing-lidar-based-3d-object-detection's Issues

AVX data set

Hi, I found this place and spent some time reading the paper and readme.md.
A lot of experiments regarding combining KITTI and AVX dataset with different proportions have been done.

I have a few questions that I would like to see them clarified.

  1. Does the AVX dataset contain reflectance(intensity) values for each point cloud? There's some synthetic data that contains x,y,z info of the point clouds but their intensity values are set to 0.
  2. I know that KITTI dataset is open to public. I'm wondering if AVX dataset is provided.

Thanks.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.