-
Notifications
You must be signed in to change notification settings - Fork 57
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The detail of classification. #18
Comments
Hi, If something is still unclear, just ask :) |
Thank you for explaining everything in detail. But I have a query like in your code where exactly you done this like where you convert those test features into latent features vector. One more thing is that you discard sigma and only consider mean? |
Thanks for your reply. I have a question. When you get the latent features to train the classifier, do you use the attribution information? Because in ZSL, you have 150 training (seen classes) and 50 test classes (unseen classes), you only know seen classes during training the classifier. If you do not use the attribution, how to associate seen classes and unseen classes? Thanks for your time to reply to this issue! |
how/where should we get the attributes for each dataset? |
The attributes for each dataset can be found in the dropbox folder that is linked in the readme. That folder just has to be copied in the repository, and then you can run the training script and it should work out of the box for all datasets. |
Not quite sure what the question about attribution information aims at, but I'll try to answer it: I guess with "attribution information" you just mean that we know beforehand which classes are seen and unseen? Then yes, we use attribution information. Not sure what is meant by "associate": Seen and unseen classes both live in the attribute space, and in this work also in the learned shared latent space. Thus, for a classifier in this space, they are all equally "seen". That way, they become associated. If with "associated" you mean associated with either the seen or unseen set, then this association is made by the humans that collect the data. |
Hi, that would be here: CADA-VAE-PyTorch/model/vaemodel.py Lines 310 to 323 in 26f0085
Yes, but only for the test samples. The representation that CADA-VAE learns for a given class is a multivariate Gaussian distribution with a mean and variance. By convention, when learning representations with VAEs, the mean is usually assumed to be the representation itself, if you need a single vector. For the test set, we chose to only use the mean and discard the sigma. Otherwise, we have stochastic variation in the test set, and that would negatively affect the comparability to other runs and methods. |
hi, how are the parameters of attributes trained? Did you just use the image labels of CUB, or use the pretrained bert weights? |
Hi, the model CADA-VAE is used to extract latent features. As for classification, I did not find the detail in the paper. Can you tell me more details about your setting to get the accuracy in the test of zero-shot?
The text was updated successfully, but these errors were encountered: