Skip to content

Conversation

@sfc-gh-zhwang
Copy link
Collaborator

@sfc-gh-zhwang sfc-gh-zhwang commented Sep 11, 2023

Get rid of the mha kvcache hacky for llama2-70b and do it in the gqa;

@sfc-gh-zhwang sfc-gh-zhwang changed the title Zhwang/llama gqa Implement kv-cache for gqa Sep 11, 2023
@sfc-gh-zhwang sfc-gh-zhwang changed the title Implement kv-cache for gqa a Sep 12, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants