valeoai / zs3 Goto Github PK
View Code? Open in Web Editor NEWZero-Shot Semantic Segmentation
Home Page: https://arxiv.org/pdf/1906.00817.pdf
License: Other
Zero-Shot Semantic Segmentation
Home Page: https://arxiv.org/pdf/1906.00817.pdf
License: Other
Hi there,
I have read your paper and the code, brilliant idea. I have a few questions on the implementation:
Hi,
I found the training and validation splits for pascal context are not provided on its website. Could someone please tell me where can I get them?
Dear dr.maximebucher:
We would like to thank you for your enlightening paper named "Zero-Shot Semantic Segmentation". Having read your release codes elaborately, we have a kind of confusion about the step 2 training process (In train_pascal_GMMN.py, line 265). The code "loss = self.criterion(output, target)" which means the unseen class segmentation annotations will be inevitably involved in the training procedure of the classifier. In other words, the annotations of the unseen classes shouldn't be used as the supervision, since the zero-shot learning should keep the unseen classes' annotation unavailable. Therefore, we hope to know how the training process of the classifier can achieve the seen and unseen classes recognition.
Thanks again, wish for your response.
As discussed in other issues, some unseen classes GT labels are used in this code. However, in the zero-shot setting, it is unsuitable since we do not know anything about the unseen classes except its name. Using some unseen classes GT labels is more like to process a semi-supervised task rather than a zero-shot task.
Hi there,
I tried training the model from scratch using train_context, and I got fair results (31%), but the train_context_GMMN model does not go higher than 17% pixel accuracy for 2 unseen categories. I downloaded pre-trained weights for 2 and 4 unseen on the pascal-context you provided and ran eval_context. But the results are not explainable:
Seen: pixel accuracy: 4.9% mIoU: 0.5%
Unseen: pixel accuracy: 1% mIoU: 0.6%
Could you please provide us with correct pre-trained weights or shed some light on how to train/eval the model?
Hi,
When I was running eval_pascal.py, an error occured
Traceback (most recent call last):
File "eval_pascal.py", line 166, in validation
all_target = np.concatenate(all_target)
File "<__array_function__ internals>", line 6, in concatenate
ValueError: all the input array dimensions for the concatenation axis must match exactly, but along dimension 2, the array at index 0 has size 700 and the array at index 1 has size 765
I tried to output target.shape
for all elements in all_target
and the dimensions vary a lot
Hi @maximebucher ,
I have read your paper and code. Thank you for sharing! Here is one question,
It seems that in the implementation you use the unseen gt labels during the training process for finetuning.
In your setting, during the training process, the unseen labels are used, while only the unseen images are not used. Am I understanding correct?
Thanks for sharing this repo.
The idea is creative. But there is some logic error in the implementation process.
In a word, you have used the unseen label info in the training process.
has_unseen_class = len(unique_class) < 21
for VOC.unique_class
of a image in real world. You just know the unique labeled classes in that image, and the whole set of unseen classes, and the whole set of classes.for unseen_label in whole_unseen_label: target[target==unseen_label] = 255
ignore_label
parameter in CEloss
to 255
to ignore the virtual unseen label to simulate the real world application totally.Despite all of above, I still believe your idea works. Hope to see your updated result.
Hello,
I read your paper and checked the repository. However, I couldn't find the code which is responsible for training and evaluating the baseline model discussed in the experiments section which is said to replace the classification layer of DeepLabv3+ with a projection layer to project extracted visual features onto the semantic embedding space to perform cosine similarity in the projected space. In the paper, it is said that this baseline is based on the DeViSE for zero-shot image classification. Could you please point me to the location in the repository where this experiment is conducted? If this experiment is not available in this repository, could you please add the responsible code to replicate it?
Thank you in advance.
The weights of the model trained on Pascal Context are missing : https://github.com/valeoai/ZS3/releases/download/0.2/deeplab_pascal_context_02_unseen_GMMN_GC_final.pth.tar
More specifically the deeplabv3+ and GMMN with graph context weights in the case of 2 unseen classes.
Do you know where I can find them please ?
Best,
In the folder, 'zs3/embeddings/pascal/w2c/norm_embed_arr_300.pkl' , how to generate?
Hi, I am currently trying to get this to run. Is there an easy way to just try out your pretrained model on my own set of images? Appreciate any help!
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.