Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: skip invokeFlattenKV_v2_ when fp16 and bf16 with CacheType::kBlock #1683

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

zhyncs
Copy link
Contributor

@zhyncs zhyncs commented May 29, 2024

Motivation and Modification

as titled

Use cases (Optional)

If this PR introduces a new feature, it is better to list some use cases here, and update the documentation.

Checklist

  1. Pre-commit or other linting tools are used to fix the potential lint issues.
  2. The modification is covered by complete unit tests. If not, please add more unit tests to ensure the correctness.
  3. If the modification has a dependency on downstream projects of a newer version, this PR should be tested with all supported versions of downstream projects.
  4. The documentation has been modified accordingly, like docstring or example tutorials.

@lzhangzz
Copy link
Collaborator

lzhangzz commented Jun 4, 2024

How about BF16, it should be the same as FP16.

@zhyncs
Copy link
Contributor Author

zhyncs commented Jun 4, 2024

How about BF16, it should be the same as FP16.

yep. I'll land the code soon.

@zhyncs zhyncs changed the title feat: skip invokeFlattenKV_v2_ when fp16 with CacheType::kBlock feat: skip invokeFlattenKV_v2_ when fp16 and bf16 with CacheType::kBlock Jun 4, 2024
@lvhan028 lvhan028 requested a review from lzhangzz June 4, 2024 06:55
@zhyncs
Copy link
Contributor Author

zhyncs commented Jun 4, 2024

Verified throughput and correctness on Llama2 13b Chat, consistent with the base.

@lzhangzz
Copy link
Collaborator

lzhangzz commented Jun 5, 2024

9% performance drop estimated for prefilling approx 200k tokens with Llama3-8B.

this PR v0.4.2
69520.48 63873.17
69465.04 63672.55
69441.23 63659.92
69397.99 63625.35
69396.67 63574.90

@zhyncs
Copy link
Contributor Author

zhyncs commented Jun 5, 2024

9% performance drop estimated for prefilling approx 200k tokens with Llama3-8B.

this PR v0.4.2
69520.48 63873.17
69465.04 63672.55
69441.23 63659.92
69397.99 63625.35
69396.67 63574.90

Ok I'll run a detailed timeline analysis later with Llama3-8B. Do you have any suggestions, such as making this feature configurable.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants