Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add the support for deepseek architecture .gguf #36144

Open
zh-jp opened this issue Feb 12, 2025 · 14 comments
Open

Add the support for deepseek architecture .gguf #36144

zh-jp opened this issue Feb 12, 2025 · 14 comments
Labels
Feature request Request for a new feature

Comments

@zh-jp
Copy link

zh-jp commented Feb 12, 2025

Feature request

The current version does not support gguf under the deepseek architecture. It is hoped that the deepseek architecture will be added. [supported-model-architectures]

Motivation

In some framework based transformers (e.g. vllm) will raise error when load .gguf file of deepseek model or quantized deepseek model.

Your contribution

Is there any guidance to help users add relevant support?

@zh-jp zh-jp added the Feature request Request for a new feature label Feb 12, 2025
@Rocketknight1
Copy link
Member

cc @SunMarc @muellerzr @MekkCyber - who's the right person to ping for GGUF loading?

@MekkCyber
Copy link
Contributor

I think me and @SunMarc and @Isotr0py

@Isotr0py
Copy link
Collaborator

Seems that all deepseek-r1 gguf checkpoints are sharded, I think we should add sharded gguf support firstly.

@Isotr0py
Copy link
Collaborator

Anyway, regardless the sharded gguf weights (we can merge weights with tool from llama.cpp), will #35926 block us currently? Because the deepseek-v3 support hasn't landed yet.

@MekkCyber
Copy link
Contributor

yes deepseek v3 is still not supported for now, the pr is functional but some small adjustments are needed

@zh-jp
Copy link
Author

zh-jp commented Feb 13, 2025

Hello @MekkCyber ! Isn't the reason that the transformers don't support it because the deepseek .gguf file can't be merged?

@MekkCyber
Copy link
Contributor

Hello @zh-jp! I think it can be merged using llama-gguf-split --merge but i'm not sure if it's a good idea

@mmdbhs
Copy link

mmdbhs commented Feb 14, 2025

we need GGUF supported! please!

6 similar comments
@Jackarry188
Copy link

we need GGUF supported! please!

@floyddcn
Copy link

we need GGUF supported! please!

@htkim27
Copy link

htkim27 commented Feb 18, 2025

we need GGUF supported! please!

@fonhorst
Copy link

we need GGUF supported! please!

@shangdibufashi
Copy link

we need GGUF supported! please!

@maguih
Copy link

maguih commented Feb 27, 2025

we need GGUF supported! please!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Feature request Request for a new feature
Projects
None yet
Development

No branches or pull requests