Giter VIP home page Giter VIP logo

deltar's Introduction

DELTAR: Depth Estimation from a Light-weight ToF Sensor And RGB Image


DELTAR: Depth Estimation from a Light-weight ToF Sensor And RGB Image
Yijin Li, Xinyang Liu, Wenqi Dong, Han Zhou, Hujun Bao, Guofeng Zhang, Yinda Zhang, Zhaopeng Cui
ECCV 2022

Demo Video

Download Link

We provide the download link [google drive, baidu(code: 1i11)] to

  • pretrained model trained on NYU.
  • ZJUL5 dataset.
  • demo data.

Run DELTAR

Installation

conda create --name deltar --file requirements.txt

Prepare the data and pretrained model

Download from the above link, and place the data and model as below:

deltar
├── data
│   ├── demo
│   └── ZJUL5
└── weights
    └── nyu.pt

Evaluate on ZJUL5 dataset

python evaluate.py configs/test_zjuL5.txt

Run the demo

python evaluate.py configs/test_demo.txt
python scripts/make_gif.py --data_folder data/demo/room --pred_folder tmp/room

Citation

If you find this code useful for your research, please use the following BibTeX entry.

@article{deltar,
  title={DELTAR: Depth Estimation from a Light-weight ToF Sensor and RGB Image},
  author={Li Yijin and Liu Xinyang and Dong Wenqi and Zhou han and Bao Hujun and Zhang Guofeng and Zhang Yinda and Cui Zhaopeng},
  booktitle={European Conference on Computer Vision (ECCV)},
  year={2022}
}

Acknowledgements

We would like to thank the authors of Adabins, LoFTR and Twins for open-sourcing their projects.

deltar's People

Contributors

eugenelyj avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

deltar's Issues

discretize the distribution by sampling depth hypotheses

Hello professor, your work has really enlighten me. However, I have some questions. In your paper 4.1 chapter, you write "discretize the distribution by sampling depth hypotheses". Could you please tell me what does that mean or how did that? Thank you!

复现testdemo出了点问题

evaluate.py: error: unrecognized arguments: --checkpoint_path /home/user/fyy/deltar/checkpoints/nyu.pt
这条指令是导入训练的权重吗,为什么是无效的呢

Capture system for L5

Hi ,
Thanks for the great work! I'm curious how to access the histogram data from the L5 sensor? I really appreciate you can share your capture system:)

Regards.

关于数据增强

大佬您好,我看到你的数据增强部分好像只有Totensor?其余的数据增强部分是没有加吗

Different results when I train my own model & bugs

Thank you for the nice work. The evaluation and demo with your weights and configuration works like in the paper. I have installed the library versions provided in the requirements.txt

My problem now is that I want to adapt, e.g. shrink, it to fit my purpose therefore I have to train my own models. So I tried to train the entire DELTAR model as it is provided by you. I use 25k nyu images. I have tweaked the code here and there to fit my data and I changed it with the two additional remarks from below. In this setting I trained it for 50 epochs, but I have also trained for 25 epochs with the settings provided in the configs for nyu. I then get the following results for validation:
image

and training loss:
image

images from the training set looks good though:
image

but it does not generalize that well:
image

My questions now are:

Additional remarks:

  • I found that the pointNet parameters where not trained in the network before as they are not added to the parameters for the optimizer in
    def get_10x_lr_params(self):  # lr learning rate
        modules = [self.decoder, self.depth_head, self.conv_out]
  • In the nyu dataloader in "train" mode the image should be normalized t0 [0,1] (image = np.array(image, dtype=np.float32) / 255.0) before the random_crop(...) and train_preprocess(...) augmentation as it is clipped there to [0,1]

cannot make the same result as the repo example image

hello,
first, thanks for sharing your work

i attempted to reproduce the same result as the example image from readme.md
but there is a big difference in quality

i used the provided weight, data, and config
should i use different parameters to reproduce the result?

i attatched the result that i made
demo_cut
please understand that the file is compressed due to file size issue

关于模型fine-tune的问题

大佬,你的工作真的太有意义了。我有个疑问,我看了论文注意到最后是在ZJU-L5数据上面进行fine-tune,这个数据量是非常小的,总共才1000多张吧,这么小的数据量fine-tune的时候不会过拟合吗?因为我目前也做了单目深度估计模型的fine-tune,我混合了几个公开数据大概20多万张,训练了一个模型,然后加载我训练好的模型在我自己采集的2000多张数据上面fine-tune的时候过拟合了。我不清楚是由于数据量太小过拟合了,还是其他什么原因。如果可以的话,可以多说一些fine-tune的细节吗?比如一些fine-tune的示例代码。

关于tof数据问题

大佬您好,我发现训练时候模拟的tof数据是像素相对于相机平面的距离,但是真实tof传出来的数据是像素相对于相机中心的距离,您在处理tof数据的时候是将原始tof数据转换成了相对于相机平面的距离吗?

关于tof输入

您好,我想问一下tof输入是否有归一化?好像看代码依旧是真实深度值?

关于参数--simu_max_distance

感谢您的工作!
我在实验过程中发现,构建直方图的部分使用了参数--simu_max_distance=4,但是--max_depth=10。
想问一下为什么要这样设置,以及这两个参数的关系。

Setting up conda environment fails

conda create --name deltar --file requirements.txt does not work for me as it does not find a lot of packages. PackagesNotFoundError: The following packages are not available from current channels:
I installed conda on my wsl with Ubuntu-22.04.
Which conda channels do you use?

Can I retrain the same algorithm for a different resolution ToF sensor?

Hi Authors, Awesome paper, I have pretty much the question said in the subject. I have a ToF sensor which has a depth image resolution of 224x172? Can I retrain the algorithm for this sensor and use it for RGB-D mapping while this algorithm gives super resolved depth image in real time?

开源工程

大佬,你的工作太有意义了,请问你什么时候开源代码呢?

Code Release Date

Hi,
Congratulations on the amazing work! When do you plan to release the code in public?

关于选取NYUdepth数据集的获取

大佬的工作很棒!我想请教一下怎么获取论文中所述选出的数据集(可以在kaggle上获得吗),我想要进一步搭建自己的数据集以供训练,谢谢!!!

关于RGBD数据仿真TOF数据

你好,我看到您的工作训练的时候,是使用RGBD数据来仿真TOF数据,从而进行训练的。我想问一下,请问从rgbd到tof具体是怎么得到的呢?感谢您的帮助。

另外,我想问一下,您的Deltar算法可以扩展到其他TOF数据嘛?比如说3040这样的,而不是88的?需要做哪些改动呢?

谢谢,期待您的回复

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.