Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Docs] Misleading documentation of the finetuning process #938

Open
paulpacaud opened this issue Feb 28, 2025 · 1 comment
Open

[Docs] Misleading documentation of the finetuning process #938

paulpacaud opened this issue Feb 28, 2025 · 1 comment

Comments

@paulpacaud
Copy link

paulpacaud commented Feb 28, 2025

📚 The doc issue

In the documentation "Enhancing InternVL2 on COCO Caption Using LoRA Fine-Tuning", it is written at the end to copy the config.json file from the original InternVL2-2B directory to the new merged model directory.
The original config.json does not necessarily match the new config.json, especially, the parameter max_dynamic_patch can be different between the two configs.
It means that if you finetune your model with let's say max_dynamic_patch=6 (as in the default shell scripts provided), and then you overwrite your config.json with the original model (that has max_dynamic_patch=12), you introduce a subtle error.

Suggest a potential alternative/fix

I am asking for clarifications from the authors or developers of this repository regarding this matter:

  • Why do we need to copy this config.json file while there is already one in the folder of the fine-tuned model?
@wzczc
Copy link

wzczc commented Mar 2, 2025

the same question

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants