Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about Enhanced Speaker #9

Open
ZhuFengdaaa opened this issue Sep 10, 2019 · 1 comment
Open

Questions about Enhanced Speaker #9

ZhuFengdaaa opened this issue Sep 10, 2019 · 1 comment

Comments

@ZhuFengdaaa
Copy link

You claim an enhanced version of Speaker in section 3.4.3. However, geographic information and actions are only used to calculate the weight of features in attention mechanism.

I have difficulty understanding why g,a are not used to directly calculate the context. Could you provide some works related to the motivation of this design?

@ZhuFengdaaa ZhuFengdaaa changed the title Questions about Motivation of Improvements on Speaker Questions about Enhanced Speaker Sep 10, 2019
@airsplay
Copy link
Owner

Thanks for pointing it out.

I used a trick "fused hidden state" in implementing the attention layer here:

h_tilde = torch.cat((weighted_context, h), 1)
.

Mathematically, it would "add" the information of query into the retrieved context vectors:

c   = Att(query, {key})
out = FC([query, c])

Thus, the information of g, a would be captured by the second LSTM.

I am sorry that I forget to mention it in the paper.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants