Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The running command on Code-Feedback #6

Open
lucasliunju opened this issue Sep 3, 2024 · 3 comments
Open

The running command on Code-Feedback #6

lucasliunju opened this issue Sep 3, 2024 · 3 comments

Comments

@lucasliunju
Copy link

Hi,

Thanks for your geat work.

May I ask the training command of LoRA-GA on Code-Feedback with LLaMA2 and LLaMA3?

Thank you very much in advance!

Best

@Outsider565
Copy link
Owner

I have updated the code to incorporate the PEFT API. You can try out this new version, which should make it easier to adapt to new datasets and different models. I believe this update will streamline the process for your use case.

@lucasliunju
Copy link
Author

lucasliunju commented Sep 8, 2024

Hi @Outsider565

Thanks for your reply.

In the updated code, I haven't find the running code or command about Code-Feedback. By the way, in your previous code, I noticed you use the tokenizer of llama2 to process the data in the data.py. If I use other models, such as llama3, I would like to ask I still need to change the tokenizer to llama3?

@Outsider565
Copy link
Owner

In the legacy code, you should use the corresponding tokenizer in data.py to ensure the token limit(512 for math and 1024 for code-feedback) is correct.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants