Comments (6)
Please refer to here. Thanks.
from lxmert.
Currently, the original CAFFE implementation (originated from Ross's py-faster-rcnn repo) could not do batch-wise inference. Because they kept the width-height ratio during inference.
A possible solution is to use detectron2/mmdetection as the backbone extractor. These libraries add the support of batch-wise inference. However, I do find that D2's batch-wise mode would decrease the score of downstream tasks because the features and RoI's are not perfectly aligned.
from lxmert.
Hi, Thanks for authors great doker environment for bu feature extraction.
However, I find it cannot use the multiply GPUs by using CUDA_VISIBLE_DEVICES=0,1,2,3 in docker?
And I check the extract.py file, it seemed that it does not support the multiply gpu extracting?Or am I missing anything?
Thanks~
Hi @Wangt-CN, can you share me /workspace/features/extract_nlvr2_image.py
? I follow this issue#79.
from lxmert.
Hi, Thanks for authors great doker environment for bu feature extraction.
However, I find it cannot use the multiply GPUs by using CUDA_VISIBLE_DEVICES=0,1,2,3 in docker?
And I check the extract.py file, it seemed that it does not support the multiply gpu extracting?Or am I missing anything?
Thanks~
Hi @Wangt-CN ,I am also interested in installing Detectron (used by VisualBERT) or Detectron2.
from lxmert.
Currently, the original CAFFE implementation (originated from Ross's py-faster-rcnn repo) could not do batch-wise inference. Because they kept the width-height ratio during inference.
A possible solution is to use detectron2/mmdetection as the backbone extractor. These libraries add the support of batch-wise inference. However, I do find that D2's batch-wise mode would decrease the score of downstream tasks because the features and RoI's are not perfectly aligned.
Hi @airsplay thanks! Seems results of these feature extraction tools are comparable and consequentially, other feature extraction tools.
By the way, Any method no need of caffe/caffe2?
although I personally can install caffe under this case of LXMERT.
from lxmert.
Currently, the original CAFFE implementation (originated from Ross's py-faster-rcnn repo) could not do batch-wise inference. Because they kept the width-height ratio during inference.
A possible solution is to use detectron2/mmdetection as the backbone extractor. These libraries add the support of batch-wise inference. However, I do find that D2's batch-wise mode would decrease the score of downstream tasks because the features and RoI's are not perfectly aligned.
OK, I think batch-wise inference can be resolved by following detectron2_mscoco_proposal_maxnms.py and I writes down codes for customized images here. I also follow issue#7 in airsplay/py-bottom-up-attention.
from lxmert.
Related Issues (20)
- binascii.Error: Incorrect padding HOT 1
- Why having a self-attn after cross_attn HOT 1
- np.all(feature == 0) is True, why
- Training is stuck
- why use tsv?
- validation score of VQA HOT 2
- the result of Lxmert on vqa2.0 does not match the paper?
- The URLs are not accessable HOT 12
- No problem
- Get predictions for batches
- Fine-tuning VQA on multiple gpus
- Pre-training mission error
- filtering raw dataset in to get tiny dataset
- how to know the object name from obj id in mscoco_imgfeat?
- how to export onnx or tensorrt model when I using lxrt.modeling.LXRTModel
- Labels
- Error in fine-tuning GQA for large datasets
- Corrupt VQA dataset from the source provided
- FileNotFoundError: [Errno 2] No such file or directory: '/hy-tmp/lxmert/data/vqa/valid.json'
- Docker extract features error
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from lxmert.