Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Issue with SAKT model #4

Open
ghodai opened this issue Feb 2, 2022 · 1 comment
Open

Issue with SAKT model #4

ghodai opened this issue Feb 2, 2022 · 1 comment

Comments

@ghodai
Copy link

ghodai commented Feb 2, 2022

I noticed a weird way of invoking the TranformerLayers in SAKT model as per the below code snippet, you are only getting the output of the last layer applied on the seq_data input, you should either combine the pool or add y as input per each call!

y = seq_data for block in self.attn_blocks: y = block(mask=1, query=q_data, key=seq_data, values=seq_data)

@THUwangcy
Copy link
Owner

Thanks for pointing out this issue! It is indeed a mistake and has been fixed in the latest commit. Meanwhile, considering that the original paper only uses a single transformer layer, our previous implementation just works the same. So the experimental results are still reliable.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants