Giter VIP home page Giter VIP logo

rellis-3d's Introduction

RELLIS-3D: A Multi-modal Dataset for Off-Road Robotics

Texas A&M University    The DEVCOM Army Research Laboratory

Peng Jiang1, Philip Osteen2, Maggie Wigness2 and Srikanth Saripalli1
1. Texas A&M University;  2. CCDC Army Research Laboratory
[Website] [Paper] [Github]

Overview

Semantic scene understanding is crucial for robust and safe autonomous navigation, particularly so in off-road environments. Recent deep learning advances for 3D semantic segmentation rely heavily on large sets of training data; however, existing autonomy datasets represent urban environments or lack multimodal off-road data. We fill this gap with RELLIS-3D, a multimodal dataset collected in an off-road environment containing annotations for 13,556 LiDAR scans and 6,235 images. The data was collected on the Rellis Campus of Texas A&M University and presents challenges to existing algorithms related to class imbalance and environmental topography. Additionally, we evaluate the current state of the art deep learning semantic segmentation models on this dataset. Experimental results show that RELLIS-3D presents challenges for algorithms designed for segmentation in urban environments. Except for the annotated data, the dataset also provides full-stack sensor data in ROS bag format, including RGB camera images, LiDAR point clouds, a pair of stereo images, high-precision GPS measurement, and IMU data. This novel dataset provides the resources needed by researchers to develop more advanced algorithms and investigate new research directions to enhance autonomous navigation in off-road environments.

LiDAR Scans Statics

Recording Platform

Sensor Setup

Sensor Setup Illustration

Annotated Data:

Ontology:

With the goal of providing multi-modal data to enhance autonomous off-road navigation, we defined an ontology of object and terrain classes, which largely derives from the RUGD dataset but also includes unique terrain and object classes not present in RUGD. Specifically, sequences from this dataset includes classes such as mud, man-made barriers, and rubble piles. Additionally, this dataset provides a finer-grained class structure for water sources, i.e., puddle and deep water, as these two classes present different traversability scenarios for most robotic platforms. Overall, 20 classes (including void class) are present in the data.

Ontology Definition (Download 18KB)

Images Statics:

Images Statics

Image Download:

Image with Annotation Examples (Download 3MB)

Full Images (Download 11GB)

Full Image Annotations (Download 94MB)

Image Split File (44KB)

LiDAR Scans Statics:

LiDAR Scans Statics

LiDAR Download:

LiDAR with Annotation Examples (Download 24MB)

LiDAR with Color Annotation PLY Format (Download 26GB)

The header of the PLY file is described as followed:

element vertex
property float x
property float y
property float z
property float intensity
property uint t
property ushort reflectivity
property uchar ring
property ushort noise
property uint range
property uchar label
property uchar red
property uchar green
property uchar blue

To visualize the color of the ply file, please use CloudCompare or Open3D. Meshlab has problem to visualize the color.

LiDAR SemanticKITTI Format (Download 14GB)

To visualize the datasets using the SemanticKITTI tools, please use this fork: https://github.com/unmannedlab/point_labeler

LiDAR Annotation SemanticKITTI Format (Download 174MB)

LiDAR Scan Poses files (Download 174MB)

LiDAR Split File (75KB)

Calibration Download:

Camera Instrinsic (Download 2KB)

Camera to LiDAR (Download 3KB)

Benchmarks

Image Semantic Segmenation

models sky grass tr ee bush concrete mud person puddle rubble barrier log fence vehicle object pole water asphalt building mean
HRNet+OCR 96.94 90.20 80.53 76.76 84.22 43.29 89.48 73.94 62.03 54.86 0.00 39.52 41.54 46.44 9.51 0.72 33.25 4.60 48.83
GSCNN 97.02 84.95 78.52 70.33 83.82 45.52 90.31 71.49 66.03 55.12 2.92 41.86 46.51 54.64 6.90 0.94 44.18 11.47 50.13

Image Semantic Segmenation Video

LiDAR Semantic Segmenation

models sky grass tr ee bush concrete mud person puddle rubble barrier log fence vehicle object pole water asphalt building mean
SalsaNext - 64.74 79.04 72.90 75.27 9.58 83.17 23.20 5.01 75.89 18.76 16.13 23.12 - 56.26 0.00 - - 40.20
KPConv - 56.41 49.25 58.45 33.91 0.00 81.20 0.00 0.00 0.00 0.00 0.40 0.00 - 0.00 0.00 - - 18.64

LiDAR Semantic Segmenation Video

Benchmark Reproduction

To reproduce the results, please refer to here

ROS Bag Raw Data

Data included in raw ROS bagfiles:

Topic Name Message Tpye Message Descriptison
/img_node/intensity_image sensor_msgs/Image Intensity image generated by ouster Lidar
/img_node/noise_image sensor_msgs/Image Noise image generated by ouster Lidar
/img_node/range_image sensor_msgs/Image Range image generated by ouster Lidar
/imu/data sensor_msgs/Imu Filtered imu data from embeded imu of Warthog
/imu/data_raw sensor_msgs/Imu Raw imu data from embeded imu of Warthog
/imu/mag sensor_msgs/MagneticField Raw magnetic field data from embeded imu of Warthog
/nerian_stereo/left_image sensor_msgs/Image Left image from Nerian Karmin2
/nerian_stereo/right_image sensor_msgs/Image Right image from Nerian Karmin2
/odometry/filtered nav_msgs/Odometry A filtered local-ization estimate based on wheel odometry (en-coders) and integrated IMU from Warthog
/os1_cloud_node/imu sensor_msgs/Imu Raw imu data from embeded imu of Ouster Lidar
/os1_cloud_node/points sensor_msgs/PointCloud2 Point cloud data from Ouster Lidar
/os1_node/imu_packets ouster_ros/PacketMsg Raw imu data from Ouster Lidar
/os1_node/lidar_packets ouster_ros/PacketMsg Raw lidar data from Ouster Lidar
/vectornav/GPS sensor_msgs/NavSatFix INS data from VectorNav-VN300
/vectornav/IMU sensor_msgs/Imu Imu data from VectorNav-VN300
/vectornav/Mag sensor_msgs/MagneticField Raw magnetic field data from VectorNav-VN300
/vectornav/Odom nav_msgs/Odometry Odometry from VectorNav-VN300
/vectornav/Pres sensor_msgs/FluidPressure
/vectornav/Temp sensor_msgs/Temperature
/velodyne_points sensor_msgs/PointCloud2 PointCloud produced by the Velodyne Lidar

ROS Bag Download

ROS Bag Examples (2GB)

Sequence 00000: Synced data: (12GB) Filtered data: (23GB) Full-stack Raw data: (29GB)

Sequence 00000 Video

Sequence 00001: Synced data: (8GB) Filtered data: (16GB) Full-stack Raw data: (22GB)

Sequence 00001 Video

Sequence 00002: Synced data: (14GB) Filtered data: (28GB) Full-stack Raw data: (37GB)

Sequence 00002 Video

Sequence 00003:Synced data: (8GB) Filtered data: (15GB) Full-stack Raw data: (19GB)

Sequence 00003 Video

Sequence 00004:Synced data: (7GB) Filtered data: (14GB) Full-stack Raw data: (17GB)

Sequence 00004 Video

ROS Environment Installment

The ROS workspace includes a plaftform description package which can provide rough tf tree for running the rosbag.

Warthog in RVIZ

Full Data Download:

Access Link

Citation

@misc{jiang2020rellis3d,
      title={RELLIS-3D Dataset: Data, Benchmarks and Analysis}, 
      author={Peng Jiang and Philip Osteen and Maggie Wigness and Srikanth Saripalli},
      year={2020},
      eprint={2011.12954},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Collaborator

The DEVCOM Army Research Laboratory

License

All datasets and code on this page are copyright by us and published under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 License.

Updates

  • 11/26/2020 v1.0 release

Related Work

SemanticUSL: A Dataset for Semantic Segmentation Domain Adatpation

LiDARNet: A Boundary-Aware Domain Adaptation Model for Lidar Point Cloud Semantic Segmentation

rellis-3d's People

Contributors

maskjp avatar maskjptamu avatar phil0stine avatar

Watchers

 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.