Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Finetuning AWQ model #934

Open
marklang3 opened this issue Feb 27, 2025 · 1 comment
Open

Finetuning AWQ model #934

marklang3 opened this issue Feb 27, 2025 · 1 comment

Comments

@marklang3
Copy link

Is it possible to finetune the InternVL2_5_AWQ model? I've had no issues finetuning the base models but finetuning the AWQ version stops the model from giving any responses to questions asked. I have tried loading the model as normal, and also with the --load-in4bit and --load-in-8bit parameters and yet no matter what all the responses are blank?

@yuecao0119
Copy link
Collaborator

Hi,

The AWQ model is a quantized model whose weights have been converted to low-precision representation. Generally speaking, the AWQ model does not directly support gradient updates, that is, it cannot be fully fine-tuned.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants