Skip to content

Conversation

@zhink
Copy link
Contributor

@zhink zhink commented Dec 12, 2024

PR types

Others

PR changes

Others

Description

Added FLAGS to replace max_partition_size、encoder_block_shape_q、decoder_block_shape_q for append_attention OP. And the value can be adjusted for better speedup.

@paddle-bot
Copy link

paddle-bot bot commented Dec 12, 2024

Thanks for your contribution!

@codecov
Copy link

codecov bot commented Dec 12, 2024

Codecov Report

All modified and coverable lines are covered by tests ✅

Project coverage is 52.21%. Comparing base (3e15f0e) to head (87d0b78).
Report is 267 commits behind head on develop.

Additional details and impacted files
@@             Coverage Diff             @@
##           develop    #9624      +/-   ##
===========================================
- Coverage    52.45%   52.21%   -0.25%     
===========================================
  Files          721      723       +2     
  Lines       115768   114326    -1442     
===========================================
- Hits         60728    59695    -1033     
+ Misses       55040    54631     -409     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

🚀 New features to boost your workflow:
  • ❄️ Test Analytics: Detect flaky tests, report on failures, and find test suite problems.

@zhink zhink force-pushed the develop branch 2 times, most recently from ca4e474 to 476a243 Compare December 12, 2024 12:11
@zhink zhink changed the title Added FLAGS to replace max_partition_size and the value can be adjusted for better speedup Added FLAGS to replace four params and the value can be adjusted for better speedup Dec 12, 2024
@zhink zhink force-pushed the develop branch 2 times, most recently from b87d151 to ffb986b Compare December 13, 2024 02:17
@zhink zhink force-pushed the develop branch 3 times, most recently from d37765f to e8a33ff Compare December 26, 2024 07:38
Comment on lines 225 to 237
inline uint32_t get_decoder_block_shape_q() {
static const char* decoder_block_shape_q_env = std::getenv("FLAGS_flag_dec_block_shape_q");
static const uint32_t decoder_block_shape_q =
decoder_block_shape_q_env == nullptr ? 16 : std::stoi(std::string(decoder_block_shape_q_env));
return decoder_block_shape_q;
}

inline uint32_t get_encoder_block_shape_q() {
static const char* encoder_block_shape_q_env = std::getenv("FLAGS_flag_block_shape_q");
static const uint32_t encoder_block_shape_q =
encoder_block_shape_q_env == nullptr ? 64 : std::stoi(std::string(encoder_block_shape_q_env));
return encoder_block_shape_q;
}
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

flag命名名以及代码位置可以再完善一下,1. 例如FLAGS_flag_,这个小写的flag是不是多余的?建议与变量名保持一致,2. max_partition_size这个flag和这个存放在一起呗

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

好的

@carryyu
Copy link
Contributor

carryyu commented Jan 3, 2025 via email

@zhink zhink marked this pull request as draft January 3, 2025 07:38
@zhink zhink marked this pull request as ready for review January 3, 2025 07:38

**Append Attention 优化**

- `FLAGS_cascade_attention_max_partition_size`:Append Attention decoder计算时对cache_kv进行分chunk的chunk大小,batchsize为1时默认值为128,batchsize大于时512。显示设置时不区分batchsize。
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

默认值根据batchsize设置,batchsize=1时设置为128,batchsize>1时设置为512。显示设置时不再区分batchsize。

**Append Attention 优化**

- `FLAGS_cascade_attention_max_partition_size`:Append Attention decoder计算时对cache_kv进行分chunk的chunk大小,batchsize为1时默认值为128,batchsize大于时512。显示设置时不区分batchsize。
- `FLAGS_dec_block_shape_q`:Append Attention decoder计算时对q进行分块的分块大小,默认值为16。设置为16即可。
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

“设置为16即可。”这里如果没有关联指导,就删除吧。下同。

Copy link
Collaborator

@DrownFish19 DrownFish19 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@ZHUI ZHUI merged commit a52035f into PaddlePaddle:develop Jan 6, 2025
10 of 13 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants