bitseek - Flash Multi Head Latent Attention inspired by deepseek. #234
chetanreddyv
started this conversation in
Ideas
Replies: 2 comments 1 reply
-
Would this also apply to models like Qwen/Dots with open weights? Also, why not do "smart" PTQ with bitnet? Then again there needs to be a sharable library for this https://huggingface.co/nisten/Biggie-SmoLlm-0.15B-Base |
Beta Was this translation helpful? Give feedback.
0 replies
-
8b version when out? i think it'll be great with 8b active parameters too. any etas on expectation? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Features
Architecture Overview
I have the code for this implementation but no compute, I want to contribute this code here as open source.
Beta Was this translation helpful? Give feedback.
All reactions