Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Kernel] Make SelfAttention prepared for AMX_FP16; More balanced task split in Cross Attention #466

Merged
merged 2 commits into from
Jul 8, 2024

Conversation

pujiang2018
Copy link
Contributor

No description provided.

@abenmao
Copy link
Contributor

abenmao commented Jul 5, 2024

LGTM. Found the fp16 gemm kernel is not ready, so the results have not been verified yet.

@pujiang2018
Copy link
Contributor Author

LGTM. Found the fp16 gemm kernel is not ready, so the results have not been verified yet.

need to enable it in next PR with new xDNN.

@pujiang2018 pujiang2018 merged commit 69b91cf into main Jul 8, 2024
1 check failed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants