We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
任务目标: 新增decoder_attention_fuse_pass; 需求场景: decoder模型性能优化; 功能描述: 新增 decoder attention 融合的pir pass, 该pass将q k v 的计算融合成一个大算子运算(不带mask)
No response
The text was updated successfully, but these errors were encountered:
[XPU][PIR] add decoder_attention_fuse_pass (PaddlePaddle#65341)
22baf2a
您好,感谢您的pr,CI通过后联系相关同学帮忙review下
Sorry, something went wrong.
e5cc2b8
db58c29
729f1c4
c9fff7b
5ff031b
No branches or pull requests
需求描述 Feature Description
任务目标: 新增decoder_attention_fuse_pass;
需求场景: decoder模型性能优化;
功能描述: 新增 decoder attention 融合的pir pass, 该pass将q k v 的计算融合成一个大算子运算(不带mask)
替代实现 Alternatives
No response
The text was updated successfully, but these errors were encountered: