Giter VIP home page Giter VIP logo

graph_distillation's Introduction

Graph Distillation

This is the code for the paper Graph Distillation for Action Detection with Privileged Modalities presented at ECCV 2018

Please note that this is not an officially supported Google product.

In this work, we propose a method termed graph distillation that incorporates rich privileged information from a large-scale multi- modal dataset in the source domain, and improves the learning in the target domain where training data and modalities are scarce.

If you find this code useful in your research then please cite

@inproceedings{luo2018graph,
  title={Graph Distillation for Action Detection with Privileged Modalities},
  author={Luo, Zelun and Hsieh, Jun-Ting and Jiang, Lu and Niebles, Juan Carlos and Fei-Fei, Li},
  booktitle={ECCV},
  year={2018}
}

Setup

All code was developed and tested on Ubuntu 16.04 with Python 3.6 and PyTorch 0.3.1.

Pretrained Models

We can download pretrained models used in our paper running the script:

sh scripts/download_models.sh

Or alternatively you can download Cloud SDK

  1. Install Google Cloud SDK (https://cloud.google.com/sdk/install)
  2. Copy the pretrained model using the following commands:
gsutil -m cp -r gs://graph_distillation/ckpt .

Running Models

We can use the scripts in scripts/ to train models on different modalities.

Classification

See classification/run.py for descriptions of the arguments.

scripts/train_ntu_rgbd.sh trains a model for a single modality.

scripts/train_ntu_rgbd_distillation.sh trains model with graph distillation. The modality being trained is specified by the xfer_to argument, and the modalities to distill from is specified in the modalities argument.

Detection

See detection/run.py for descriptions of the arguments. Note that the visual_encoder_ckpt_path argument is the pretrained visual encoder checkpoint, which should be from training classification models.

scripts/train_pku_mmd.sh trains a model for a single modality.

scripts/train_pku_mmd_distillation.sh trains model with graph distillation. The modality being trained is specified by the xfer_to argument, and the modalities to distill from is specified in the modalities argument.

graph_distillation's People

Contributors

d1ngn1gefe1 avatar roadjiang avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar

graph_distillation's Issues

Version of depth data from NTU RGB+D

Thanks for your great work.
I've been re-producing your code to new version of pytorch.
I wonder that which version of depth data in code, specifically masked version with RoI or original version.
So, I would appreciate it if you could let me know the version.

Error in make_dataset

File "/home/neha/Downloads/graph_distillation/graph_distillation/data_pipeline/ntu_rgbd.py", line 82, in make_dataset
keep = keep[np.linspace(0, len(keep)-1, subsample).astype(int)]
IndexError: index 0 is out of bounds for axis 0 with size 0

I have run the script file for dataset. On running script file for training classification model for single modality, I am facing above error. I checked the code and found vid_names variable is empty.Kindly help.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.