Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

about voken regression and voken constrastive #11

Open
lizhiustc opened this issue Nov 5, 2022 · 1 comment
Open

about voken regression and voken constrastive #11

lizhiustc opened this issue Nov 5, 2022 · 1 comment

Comments

@lizhiustc
Copy link

I have two questions.

(1) I notice that in your code

# Build Loss functions
, you design three loss function voken classification, voken regression and voken constrastive. But you only report "voken classification" in paper, maybe you find "voken regression and voken constrastive" both don't work or even harm model performance after trials? Is my guess correct ? (Because image features are far different from language embeddings. )

(2) What's the intuition that voken classification loss can improve model performance ? I suspect that different words with similar semantic will have same voken labels and voken classification loss will optimize their similarity. What is your opinion?Could you give me some intuition from your views?

@airsplay
Copy link
Owner

airsplay commented Nov 5, 2022

Copy-paste the email reply here:

  • Yes, these token losses perform similarly. We thus choose the simplest one. To me, it's classification.

  • Token label is also a strong supervision. For me, they are mostly used for distillation. Contrastive and L2-reg are more like distillation, but tokens can do the same (e.g., in the language mode distillations). Some other works to look at are: wave2vec 2.0, DINO.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants