-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add support for Deepseek #2692
Comments
code models using mix of expert archtecture: |
is there work on this for candle? |
@franklucky001 @bnusunny @phdbrianlee @super-fun-surf I'm working on implementing these. The PRs are not yet uploaded, but the following has been implemented
|
rad! great to hear. |
great!! |
@franklucky001 @bnusunny @phdbrianlee @super-fun-surf Both DeepSeek V2 and V3/R1 PRs have been opened:
@LaurentMazare the DeepSeek V3/R1 model is very large - do you think it would be more appropriate to implement it similar to the |
rad!. the quantized / distilled versions of R1 should work on smaller gpus. |
deepseek-ai/DeepSeek-V2-Chat
The text was updated successfully, but these errors were encountered: