Giter VIP home page Giter VIP logo

places_devkit's Introduction

Places365 Development Kit

Introduction

This is the documentation of the Places365 challenge 2016 development kit.

Table of contents:

  • Overview of challenge dataset
  • Challenge details
    1. Image list and annotations
    2. Submission format
    3. Evaluation routines

Please contact Bolei Zhou ([email protected]) for questions, comments, or bug reports.

Note: Please first download the image list and annotations for Places365-Standard and the image list and annotations for Place365-Challenge, and decompress the files in the data folder.

Overview of Places365-Standard Data

There are three types of image data for Places365-Standard: training data from Places365-Standard (TRAINING), validation data (VALIDATION), and test data (TEST). There is no overlap in the three sources of data: TRAINING, VALIDATION, and TEST. All three sets of data contain images of 365 categories of scenes.

             Number of images

Dataset      TRAIN      VALIDATION     TEST

Places365-Standard    1,803,460      36,500      328,500

Every image in training, validation and test sets has a single image-level label specifying the presence of one scene category

Places365-Standard statistics:

Training:

- 1,803,460 images, with between 3,068 and 5,000 per category

Validation:

- 36,500 images, with 50 images per category

Test:

- 328,500 images, with 900 images per category

Packaging details:

The 3 sets of images (training, validation and test) are available as 3 tar archives. All images are in JPEG format. We provide both the original images and images resized to 256*256 for download.

Overview of Places365-Challenge Data

There are three types of image data for this competition: training data from Places365-Challenge (TRAINING), validation data (VALIDATION), and test data (TEST). There is no overlap in the three sources of data: TRAINING, VALIDATION, and TEST. All three sets of data contain images of 365 categories of scenes. The VALIDATION and TEST are the same as the Places365-Standard. The first 5000 images (or less as it is bounded by the total image number in that category) in each category are the images from Places365-Standard train set.

             Number of images

Dataset      TRAIN      VALIDATION     TEST

Places365-Challenge    8,026,628      36,500      328,500

Every image in training, validation and test sets has a singleimage-level label specifying the presence of one scene category

Places365-Challenge statistics:

Training:

- 8,097,967 images, with between 3068 and 40,000 per category

Validation:

- 36,500 images, with 50 images per category

Test:

- 328,500 images, with 900 images per category

Packaging details:

The 3 sets of images (training, validation and test) are available as 3 tar archives. All images are in JPEG format. We provide both the original images and images resized to 256*256 for download.

Places Challenge 2016 details

The 365 scene categories used in the challenge dataset are part of the larger Places2 dataset.

All the class names and ids are available in: data/categories_places365.txt,

where each line contains the scene category name followed by its id (an integer between 0 and 364).

The difference betweee Places365-Challenge and Places365-Standard is that there are ~6.2million more extra images in Places365-challenge compared to Places365-standard. The first 5000 images (or less) per category in Places365-challenge belong to the Places365-standard.

1 Training data

Each image is considered as belonging to a particular scene category. See [1] for more details of the collection and labeling strategy.

The training images may be downloaded as a single tar archive. Within it there is a tar file for each alphabet from 'a.tar' to 'z.tar'. Note that there are 24 such files as there are no scene category names beginning with 'q' and 'x' in our database.

After untarring all of the above files, the directory structure should look similar to the following: a/abbey/00000000.jpg a/abbey/00000001.jpg ... z/zen_garden/00009067.jpg
z/zen_garden/00009068.jpg

In general, each leaf folder contains one scene category. Note that there are some categories that are fine-grained, e.g., s/swimming_pool/indoor and s/swimming_pool/outdoor. The complete list of training images and their mapping to scene category ids is available in: data/places365_train_challenge.txt

All images are in JPEG format. We also include the data/places365_train_standard.txt here, you don't need to use it.

2 Validation data

There are a total of 36,500 validation images. They are named as

  Places365_val_00000001.jpg
  Places365_val_00000002.jpg
  ...
  Places365_val_00036499.jpg
  Places365_val_00036500.jpg

There are 50 validation images for each scene category.

The classification ground truth of the validation images is in data/places365_val.txt,

where each line contains one image filename and its corresponding scene category label (from 0 to 364).

3 Test data

There are a total of 328,500 test images. The test files are named as

  Places365_test_00000001.jpg
  Places365_test_00000002.jpg
  ...
  Places365_test_00328499.jpg
  Places365_test_00328500.jpg

There are 900 test images for each scene category. The ground truth annotations will not be released.

Submission format

The submission of results on test data will consist of a text file with one line per image, in the alphabetical order of the image file names, i.e. from Places365_test_00000001.jpg to Places365_test_00328500.jpg. Each line contains up to 5 detected scenes, sorted by confidence in descending order.

The format is as follows:

<label(1)> <label(2)> <label(3)> <label(4)> <label(5)>

The predicted labels are the scene categories ( integers between 0 and 364 ). The number of labels per line must be exactly equal to 5, or it would lead to an error. The filename is the same as mentioned above, e.g., 'Places365_test_00000001.jpg' and so on.

Example file on the validation data is

evaluation/demo.val.pred.txt

Evaluation routines

The Matlab routine for evaluating the submission is

./evaluation/eval_cls.m

To see an example of using the routines, start Matlab in the 'evaluation/' folder and type demo_eval_cls;

and you will see something similar to the following output:

PLACES365 SCENE CLASSIFICATION TASK pred_file: demo.val.pred.txt ground_truth_file: ../data/places365_val.txt

guesses vs cls error

1.0000    0.9974
2.0000    0.9944
3.0000    0.9920
4.0000    0.9893
5.0000    0.9867

In this demo, we take top i ( i=1...5) predictions (and ignore the rest) from your result file and plot the error as a function of the number of guesses.

Only the error with 5 guesses will be used to determine the winner.

(The demo.val.pred.txt used here is a synthetic result.)

References

[1] B. Zhou, A. Khosla, A. Lapedriza, A. Torralba and A. Oliva Places: An Image Database for Deep Scene Understanding Arxiv, 2015 (PDF coming soon) http://places2.csail.mit.edu

places_devkit's People

Contributors

zhoubolei avatar

Watchers

 avatar  avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.