Comments (12)
Hi! I am trying to use the EgoNet on a custom dataset (such as Waymo Dataset) and generate 3D bboxes from the 2D predictions generated by 2D bbox detectors such as Faster R-CNN.
A sample of KITTI predictions where 2D bboxes are output but 3D bboxes are not yet output are as follows:
Car -1 -1 0.0 158.0 386.0 220.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95 Car -1 -1 0.0 565.0 386.0 720.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95
However, when I set these 2D predictions as additional input for predicting bboxes using EgoNet, the 3D bboxes are not properly output at all and are just a bunch of zeros and minus-ones.
What should I do?
Hi, EgoNet only updates orientation and does not modify or add 3D location/dimension information. Using 2D boxes as proposals by default only updates orientations. You can use 3D box proposals as inputs instead.
from egonet.
Hi! I am trying to use the EgoNet on a custom dataset (such as Waymo Dataset) and generate 3D bboxes from the 2D predictions generated by 2D bbox detectors such as Faster R-CNN.
A sample of KITTI predictions where 2D bboxes are output but 3D bboxes are not yet output are as follows:Car -1 -1 0.0 158.0 386.0 220.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95 Car -1 -1 0.0 565.0 386.0 720.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95
However, when I set these 2D predictions as additional input for predicting bboxes using EgoNet, the 3D bboxes are not properly output at all and are just a bunch of zeros and minus-ones.
What should I do?Hi, EgoNet only updates orientation and does not modify or add 3D location/dimension information. Using 2D boxes as proposals by default only updates orientations. You can use 3D box proposals as inputs instead.
Then, what should I do if I want to generate 3D bounding boxes from 2D bounding boxes and orientation information? Are any options to config files to do so?
from egonet.
Hi! I am trying to use the EgoNet on a custom dataset (such as Waymo Dataset) and generate 3D bboxes from the 2D predictions generated by 2D bbox detectors such as Faster R-CNN.
A sample of KITTI predictions where 2D bboxes are output but 3D bboxes are not yet output are as follows:Car -1 -1 0.0 158.0 386.0 220.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95 Car -1 -1 0.0 565.0 386.0 720.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95
However, when I set these 2D predictions as additional input for predicting bboxes using EgoNet, the 3D bboxes are not properly output at all and are just a bunch of zeros and minus-ones.
What should I do?Hi, EgoNet only updates orientation and does not modify or add 3D location/dimension information. Using 2D boxes as proposals by default only updates orientations. You can use 3D box proposals as inputs instead.
Then, what should I do if I want to generate 3D bounding boxes from 2D bounding boxes and orientation information? Are any options to config files to do so?
If you mean you want to visualize the 3D boxes in this case, you can draw the 3D boxes with default sizes and the predicted orientation.
from egonet.
Hi! I am trying to use the EgoNet on a custom dataset (such as Waymo Dataset) and generate 3D bboxes from the 2D predictions generated by 2D bbox detectors such as Faster R-CNN.
A sample of KITTI predictions where 2D bboxes are output but 3D bboxes are not yet output are as follows:Car -1 -1 0.0 158.0 386.0 220.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95 Car -1 -1 0.0 565.0 386.0 720.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95
However, when I set these 2D predictions as additional input for predicting bboxes using EgoNet, the 3D bboxes are not properly output at all and are just a bunch of zeros and minus-ones.
What should I do?Hi, EgoNet only updates orientation and does not modify or add 3D location/dimension information. Using 2D boxes as proposals by default only updates orientations. You can use 3D box proposals as inputs instead.
Then, what should I do if I want to generate 3D bounding boxes from 2D bounding boxes and orientation information? Are any options to config files to do so?
If you mean you want to visualize the 3D boxes in this case, you can draw the 3D boxes with default sizes and the predicted orientation.
Precisely speaking, what I want to do is to lift the 2D bounding boxes to 3D bounding boxes without 3D proposals. Is this not possible for EgoNet?
from egonet.
Hi! I am trying to use the EgoNet on a custom dataset (such as Waymo Dataset) and generate 3D bboxes from the 2D predictions generated by 2D bbox detectors such as Faster R-CNN.
A sample of KITTI predictions where 2D bboxes are output but 3D bboxes are not yet output are as follows:Car -1 -1 0.0 158.0 386.0 220.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95 Car -1 -1 0.0 565.0 386.0 720.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95
However, when I set these 2D predictions as additional input for predicting bboxes using EgoNet, the 3D bboxes are not properly output at all and are just a bunch of zeros and minus-ones.
What should I do?Hi, EgoNet only updates orientation and does not modify or add 3D location/dimension information. Using 2D boxes as proposals by default only updates orientations. You can use 3D box proposals as inputs instead.
Then, what should I do if I want to generate 3D bounding boxes from 2D bounding boxes and orientation information? Are any options to config files to do so?
If you mean you want to visualize the 3D boxes in this case, you can draw the 3D boxes with default sizes and the predicted orientation.
Precisely speaking, what I want to do is to lift the 2D bounding boxes to 3D bounding boxes without 3D proposals. Is this not possible for EgoNet?
If you are not able to prepare 3D proposals. You can still lift 2D predictions to a 3D bounding box by:
- Run EgoNet and get 2D part coordinates, 3D dimension and orientation predictions.
- Solve a PnP problem to estimate translation that minimizes re-projection errors. You can refer to the paper "3d bounding box estimation using deep learning and geometry" for details on how this can be done.
Of course, this would not result in accurate translation estimation compared to an existing 3D proposal.
from egonet.
Hi! I am trying to use the EgoNet on a custom dataset (such as Waymo Dataset) and generate 3D bboxes from the 2D predictions generated by 2D bbox detectors such as Faster R-CNN.
A sample of KITTI predictions where 2D bboxes are output but 3D bboxes are not yet output are as follows:Car -1 -1 0.0 158.0 386.0 220.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95 Car -1 -1 0.0 565.0 386.0 720.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95
However, when I set these 2D predictions as additional input for predicting bboxes using EgoNet, the 3D bboxes are not properly output at all and are just a bunch of zeros and minus-ones.
What should I do?Hi, EgoNet only updates orientation and does not modify or add 3D location/dimension information. Using 2D boxes as proposals by default only updates orientations. You can use 3D box proposals as inputs instead.
Then, what should I do if I want to generate 3D bounding boxes from 2D bounding boxes and orientation information? Are any options to config files to do so?
If you mean you want to visualize the 3D boxes in this case, you can draw the 3D boxes with default sizes and the predicted orientation.
Precisely speaking, what I want to do is to lift the 2D bounding boxes to 3D bounding boxes without 3D proposals. Is this not possible for EgoNet?
If you are not able to prepare 3D proposals. You can still lift 2D predictions to a 3D bounding box by:
- Run EgoNet and get 2D part coordinates, 3D dimension and orientation predictions.
- Solve a PnP problem to estimate translation that minimizes re-projection errors. You can refer to the paper "3d bounding box estimation using deep learning and geometry" for details on how this can be done.
Of course, this would not result in accurate translation estimation compared to an existing 3D proposal.
So, you are suggesting that EgoNet alone cannot estimate 3D bounding boxes from 2D predictions?
from egonet.
Hi! I am trying to use the EgoNet on a custom dataset (such as Waymo Dataset) and generate 3D bboxes from the 2D predictions generated by 2D bbox detectors such as Faster R-CNN.
A sample of KITTI predictions where 2D bboxes are output but 3D bboxes are not yet output are as follows:Car -1 -1 0.0 158.0 386.0 220.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95 Car -1 -1 0.0 565.0 386.0 720.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95
However, when I set these 2D predictions as additional input for predicting bboxes using EgoNet, the 3D bboxes are not properly output at all and are just a bunch of zeros and minus-ones.
What should I do?Hi, EgoNet only updates orientation and does not modify or add 3D location/dimension information. Using 2D boxes as proposals by default only updates orientations. You can use 3D box proposals as inputs instead.
Then, what should I do if I want to generate 3D bounding boxes from 2D bounding boxes and orientation information? Are any options to config files to do so?
If you mean you want to visualize the 3D boxes in this case, you can draw the 3D boxes with default sizes and the predicted orientation.
Precisely speaking, what I want to do is to lift the 2D bounding boxes to 3D bounding boxes without 3D proposals. Is this not possible for EgoNet?
If you are not able to prepare 3D proposals. You can still lift 2D predictions to a 3D bounding box by:
- Run EgoNet and get 2D part coordinates, 3D dimension and orientation predictions.
- Solve a PnP problem to estimate translation that minimizes re-projection errors. You can refer to the paper "3d bounding box estimation using deep learning and geometry" for details on how this can be done.
Of course, this would not result in accurate translation estimation compared to an existing 3D proposal.So, you are suggesting that EgoNet alone cannot estimate 3D bounding boxes from 2D predictions?
My answer is EgoNet can estimate 3D bounding boxes from 2D predictions. The above two steps already describe what to do in this case. It seems you are not capable of understanding the two steps I mentioned. Step two is contained in some legacy code which you may refer to:
#9 (comment)
This is not an issue and you can ask further questions in the discussion part of this repo.
from egonet.
Hi! I am trying to use the EgoNet on a custom dataset (such as Waymo Dataset) and generate 3D bboxes from the 2D predictions generated by 2D bbox detectors such as Faster R-CNN.
A sample of KITTI predictions where 2D bboxes are output but 3D bboxes are not yet output are as follows:Car -1 -1 0.0 158.0 386.0 220.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95 Car -1 -1 0.0 565.0 386.0 720.0 438.0 0.0 0.0 0.0 -1.0 -1.0 -1.0 0.0 0.95
However, when I set these 2D predictions as additional input for predicting bboxes using EgoNet, the 3D bboxes are not properly output at all and are just a bunch of zeros and minus-ones.
What should I do?Hi, EgoNet only updates orientation and does not modify or add 3D location/dimension information. Using 2D boxes as proposals by default only updates orientations. You can use 3D box proposals as inputs instead.
Then, what should I do if I want to generate 3D bounding boxes from 2D bounding boxes and orientation information? Are any options to config files to do so?
If you mean you want to visualize the 3D boxes in this case, you can draw the 3D boxes with default sizes and the predicted orientation.
Precisely speaking, what I want to do is to lift the 2D bounding boxes to 3D bounding boxes without 3D proposals. Is this not possible for EgoNet?
If you are not able to prepare 3D proposals. You can still lift 2D predictions to a 3D bounding box by:
- Run EgoNet and get 2D part coordinates, 3D dimension and orientation predictions.
- Solve a PnP problem to estimate translation that minimizes re-projection errors. You can refer to the paper "3d bounding box estimation using deep learning and geometry" for details on how this can be done.
Of course, this would not result in accurate translation estimation compared to an existing 3D proposal.So, you are suggesting that EgoNet alone cannot estimate 3D bounding boxes from 2D predictions?
My answer is EgoNet can estimate 3D bounding boxes from 2D predictions. The above two steps already describe what to do in this case. It seems you are not capable of understanding the two steps I mentioned. Step two is contained in some legacy code which you may refer to: #9 (comment)
This is not an issue and you can ask further questions in the discussion part of this repo.
Alright. Thanks. Sorry if I dragged this issue too long.
from egonet.
This may be a stupid question, but can we feed in results of the pnp algorithm back into Egonet itself to obtain a refined 3D box prediction?
So if our input is a 2D image and bounding box, Egonet gives us the orientation and dimensions only as it is 2D. We then solve the PnP algorithm to get a 3D bounding box estimate, and then feed that back into Egonet to refine this prediction.
from egonet.
This may be a stupid question, but can we feed in results of the pnp algorithm back into Egonet itself to obtain a refined 3D box prediction? So if our input is a 2D image and bounding box, Egonet gives us the orientation and dimensions only as it is 2D. We then solve the PnP algorithm to get a 3D bounding box estimate, and then feed that back into Egonet to refine this prediction.
Yes. You can treat the results of the pnp algorithm as another input 3D box proposal, compute a 2D RoI from it and feed that to Ego-Net. The orientation predictions may fluctuate a bit but not too much if the 2D RoI inputs are similar.
from egonet.
Thanks. So just to confirm that after the pnp algorithm we do not use original 2D ROI, but compute this from the 3D ROI produced by pnp?
from egonet.
Thanks. So just to confirm that after the pnp algorithm we do not use original 2D ROI, but compute this from the 3D ROI produced by pnp?
Yes, During training the 2D RoIs are computed from ground truth 3D boxes.
from egonet.
Related Issues (16)
- Running the model on a single image HOT 5
- About the license for this model HOT 2
- Visualazition problem
- Use EgoNet on custom data HOT 5
- Generate 3D rectangular coordinates using 2D rectangular boxes HOT 4
- Relation between kpts_3d_pred and pose HOT 2
- Result Image HOT 1
- 关于“car_instance”中代码的请教 HOT 1
- 如何使用该模型推断自己的图片
- 如何使用该模型推断自己的图片 HOT 4
- Training Script HOT 2
- Reproduce Result in Kitti dataset HOT 1
- Reproduce results on the test split HOT 2
- No arg_max in KITTI_train_ICRs.yml file HOT 1
- Stream from camera HOT 7
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from egonet.