Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add pir decoder_attention_fuse_pass #65341

Open
linkk08 opened this issue Jun 20, 2024 · 1 comment
Open

add pir decoder_attention_fuse_pass #65341

linkk08 opened this issue Jun 20, 2024 · 1 comment
Labels

Comments

@linkk08
Copy link
Contributor

linkk08 commented Jun 20, 2024

需求描述 Feature Description

任务目标: 新增decoder_attention_fuse_pass;
需求场景: decoder模型性能优化;
功能描述: 新增 decoder attention 融合的pir pass, 该pass将q k v 的计算融合成一个大算子运算(不带mask)

替代实现 Alternatives

No response

@warrentdrew
Copy link
Contributor

您好,感谢您的pr,CI通过后联系相关同学帮忙review下

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants