Giter VIP home page Giter VIP logo

Comments (23)

irolaina avatar irolaina commented on June 2, 2024

Hi,
The ground truth depth images used for training were actually raw depth images. Invalid pixels (where depth is zero) have been then excluded from training.

from fcrn-depthprediction.

seokhoonboo avatar seokhoonboo commented on June 2, 2024

sorry, I have one more question.
'relative err : 0.129' in your paper is calcualated by only testNdx datas(hole filled) in 'nyu_depth_v2_labeled.mat'(654 Image pairs)? or All test scene Images in NYU_Depth_v2_raw dataset(about 200k image pairs)?
in this git hub source code, you using only testNdx datas(hole filled) in 'nyu_depth_v2_labeled.mat'(654 Image pairs) for test.

from fcrn-depthprediction.

irolaina avatar irolaina commented on June 2, 2024

The error metrics are calculated over the official split of 654 images in the labeled subset of NYU (to fairly compare to prior works). In this case, we are using the depth maps which were filled-in using the colorization method and not the raw data. The errors should be lower when comparing to raw depth maps.

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

Hi, @seokhoonboo ! Have you successfully reimplement it in caffe? I have got some troubles, like how to set upsampling layer(since there is no direct upsampling layer like paper in caffe, instead I use deconvolution layer to replace upsampling, but it seems that the network is hard to learn anything...).

from fcrn-depthprediction.

bazilas avatar bazilas commented on June 2, 2024

@jszhujun2010 it might be easier to implement upsampling in caffe. Training this network with deconv will be problematic. However, you might want to try not to learn the parameters of the deconv and only use it as a bilinear filter.

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

Thanks, @bazilas . I'll consider to write my own unsampling layer.

from fcrn-depthprediction.

seokhoonboo avatar seokhoonboo commented on June 2, 2024

@jszhujun2010
https://github.com/matthieudelaro/caffeBVLCplus
this repository include unpooling layer, and i am trying by using it

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

Thanks, @seokhoonboo . I guess UnpoolingParameter_UnpoolMethod_MAX is what we need.

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

Hey, guys! I'm trying to train a model on Make3D dataset, here is details:

  • I resized all images to 172*230 as CNN's input(since the paper mentioned that all images are resized to 345*460 and then reduce the resolution by half), and augmented the data to 15k. I resized output data(305*55) to 172*230.
  • The CNN architecture is just as the paper's description except some minor changes. Since 4 upconv structure can not get 172*230, so I add another upconv layer(resulting 192*256) and a crop layer(as FCN segmentation).
  • I set leaning rate as 10^(-10) because larger lr will cause infinite loss. Other parameters are exactly the same as paper.

However, there is no signal of convergence after thousands of iterations. Is there anything I'm missing or wrong?

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

Hi! @iro-cp , I have one more question. Did you your fine-tune the whole network or just the part after the resnet?

from fcrn-depthprediction.

irolaina avatar irolaina commented on June 2, 2024

@jszhujun2010
Regarding fine-tuning: All layers are trained for the depth estimation task, including both the core-ResNet and the upsampling part.

To answer your previous question: When you do the data augmentation you should make sure that you do not interpolate depth values that shouldn't be interpolated (for example when the ground truth depth map includes invalid pixels). In our work, all transformations are performed with nearest neighbor interpolation.
I think the lack of convergence is related to your very low learning rate. If you had to set it this low to avoid an infinite loss, you might want to make sure that there is everything is fine with your training data for example.

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

@iro-cp
Thanks for your reply. It's still for me to train the model, the model is even hard to overfit(using two images with batch size 2, original data without transformation).
I have forked the repository and added my caffe prototxt as well as some preprocessing scripts. You guys can check it if you are interested.
Repo is here, and caffe prototxt is in caffe folder.

from fcrn-depthprediction.

cv-small-snails avatar cv-small-snails commented on June 2, 2024

Hi @jszhujun2010 Do you train your model on NYU Depth v2 dataset successfully?

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

kind of... I have trained on NYU data without data agumention and the model can fit training data well(testing result not so good though, due to overfitting).

from fcrn-depthprediction.

cv-small-snails avatar cv-small-snails commented on June 2, 2024

Hi @jszhujun2010 I have trained your model in the repository on NYU Depth v2 dataset, but I found the loss also can't converge. The Loss map can be seen here.I don't why ? Is there any question?
untitled

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

I have modified the network and I'll update it recently.

from fcrn-depthprediction.

cv-small-snails avatar cv-small-snails commented on June 2, 2024

@jszhujun2010
Thanks for your reply. I have trained your modified network on NYU Depth v2 dataset again. But I found the loss is still as shown above and can't converge.Is it normal?Is there any question?

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

Well, it's fine for me.
image
How many iterations is your setting?

from fcrn-depthprediction.

cv-small-snails avatar cv-small-snails commented on June 2, 2024

@jszhujun2010
Thanks for your reply. I have set 300000 iterations on NYU Depth v2 dataset.But the loss still can't converge.And the Loss map still didn't like the map as you show! Do you change anything of your network?I don't why ?

from fcrn-depthprediction.

jszhujun2010 avatar jszhujun2010 commented on June 2, 2024

What's your data? I only used 795 training images from official dataset(as a result, overfitting is quite obvious). I found that their training data is sampled from video, I'm too lazy to download that large file...
Another thing is that, I still cannot reimplement make3D dataset's experiment even in training(training loss can not decrease to ideal level). I guess there must be some trick in data preprocessing phase, but I still cannot figure it out... I think make3D is much more difficult to train because its depth range is too large for the network to learn.

from fcrn-depthprediction.

chrirupp avatar chrirupp commented on June 2, 2024

looks like all problems are resolved. Please open another issue if there are open questions.

from fcrn-depthprediction.

Lvhhhh avatar Lvhhhh commented on June 2, 2024

can you share the train code?

from fcrn-depthprediction.

buaaswf avatar buaaswf commented on June 2, 2024

Could you share the pre-processing code to the make3d dataset? I can not find how to augment the training dataset from 343 to 15kļ¼Œ thank you

from fcrn-depthprediction.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    šŸ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. šŸ“ŠšŸ“ˆšŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ā¤ļø Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.