Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it possible to run llama 3-70B and/or mixtral 8x22b through this process? #1

Open
RandomInternetPreson opened this issue Jun 26, 2024 · 1 comment

Comments

@RandomInternetPreson
Copy link

I'm running the Llama-3-Instruct-8B-SPPO-Iter3 model locally and am very impressed by the improved quality from the original model. I can't help but wonder what the results would be if this finetuning process were run on larger models.

Is it possible to run the code on these larger models, or are the smaller versions too different form their larger counterparts; requiring a rework of the training scripts?

Thank you for what you have contributed, this is great stuff!

@angelahzyuan
Copy link
Collaborator

Thank you! We've trained a slightly larger model (UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3) which achieved an LC-win rate of 53.27, using the same parameters and scripts.

As long as your GPU has sufficient VRAM, the training script should perform well. We will keep you updated as we proceed to training larger models.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants