Giter VIP home page Giter VIP logo

scenenn's Introduction

SceneNN

SceneNN is an RGB-D dataset with more than 100 indoor scenes. All scenes are carefully annotated with instance-level segmentation.

Please refer to our project page at http://www.scenenn.net for the papers and scene previews in the supplementary document.

To access the dataset, please use following repositories on Google Drive:

1. Instance-level segmentation

https://drive.google.com/drive/folders/0B-aa7y5Ox4eZWE8yMkRkNkU4Tk0

This includes all 90+ scenes captured by Asus Xtion PRO.

Please use the Python script scenenn_download to download a few specific scenes or the entire dataset. To run this script, please install PyDrive.

3D annotation

Each scene is identified by an ID with three digits, e.g., 076, 123. The data of the scene is structured as follows:

SceneNN
├── 123
│   ├── 123.ply                 /* the reconstructed triangle mesh  */
│   ├── 123.xml                 /* the annotation                   */
│   ├── trajectory.log          /* camera pose (local to world)     */
└── oni
    └── 123.oni                 /* the raw RGBD video               */
└── intrinsic
    └── asus.ini                /* intrinsic matrix data for Asus Xtion camera  */
	└── kinect2.ini             /* intrinsic matrix data for Kinect v2 camera   */
  • Depending on which application you are working on, you might want to download only the mesh with annotation (~100 - 200MB per scene), or include the raw RGBD (~1 GB per scene).

  • The object tag and axis-aligned bounding box are stored separately in the XML file. If you need object poses, please find them in the beta repository, below.

  • The format of the oriented bounding box is: box center (x, y, z), box dimension (along x, y, z), and a quaternion (x, y, z, w) that represents the rotation of the box.  Please note that the obbox are computed using eigen decomposition, so it is only accurate when the object is elongated enough. 

Color meshes

The color meshes for each scene could be downloaded from our experimental repository here: [https://drive.google.com/drive/folders/0B2BQi-ql8CzeakFnWWpqS3hVWlE]

2D annotation

Please download them here: https://drive.google.com/drive/folders/0B2BQi-ql8CzeaDNQQXNmZHdnSFE

We encode the label to color channels in RGB format. To decode, please use this simple C++ code.

Raw RGB-D videos

We also provide raw RGB-D files that are output from the Asus Xtion camera. They are stored in the oni folder in both repositories. (You might need to scroll down to the bottom of the scene list to see it.)

In order to extract depth and color images from ONI files, please use the tool in the playback folder from this repository. After extraction, depth map is stored in 16-bit unsigned short format. The depth unit is in millimeter

For Windows, a precompiled binary is provided. Note that OpenNI 1.5.4 x64 must be installed.

For Linux, a MakeFile is included in the playback folder. The compilation is tested on Ubuntu 16. (Credit and great thanks to Joao Henriques (joao [at] robots.ox.ac.uk) for this handy script!)

sudo apt-get install libopenni-dev libfreeimage-dev libglew-dev freeglut3-dev
make all

Camera pose

The trajectory.log contains blocks of 5 lines:

  • The first line in each block indicates the frame index.
  • The next 4 lines is a 4x4 matrix. This matrix transforms a point from camera space to world space.

Intrinsic matrix is in the format of [fx, 0, cx; 0, fy, cy; 0 0 1], where fx, fy, cx, cy could be found in asus.ini or kinect2.ini, respectively.

2. Category-level segmentation (beta)

https://drive.google.com/drive/folders/0B2BQi-ql8CzeUkJNbEVEQkozSmc

This includes our experimental category-level segmentation and object pose annotation:

  • Semantic labels using the 40 classes defined in NYU-D v2.
  • Object poses and better floor aligned meshes.

Such annotation will be merged to the main repository once they are stable.

This repository also includes scenes captured by both Asus Xtion PRO and Kinect v2.

ObjectNN

ObjectNN is a subset of SceneNN that has all RGB-D objects extracted and split into 20 categories. We used it for our SHREC 2017 track: RGB-D to CAD object retrieval. For more information, please see here.

License

Our datasets are free for educational and research use. We would love to hear from you if you find them useful. If you use the data, please cite

@inproceedings{scenenn-3dv16,
    author = {Binh-Son Hua and Quang-Hieu Pham and Duc Thanh Nguyen and Minh-Khoi Tran and Lap-Fai Yu and Sai-Kit Yeung},
    title = {SceneNN: A Scene Meshes Dataset with aNNotations},
    booktitle = {International Conference on 3D Vision (3DV)},
    year = {2016}
}

If you intend to use the data for commercial applications, please kindly contact us beforehand at scenenn[at]gmail.com.

scenenn's People

Contributors

songuke avatar sonhua avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.