Private 4-bit bnb Beta: 65B LLMs on a Single GPU

Thank you for your interest in our private beta!

As a beta participant, you will have early access to 4-bit quantization and paged optimizers code for efficient LLM finetuning. This code allows finetuning of the largest LLaMA models on a single GPU and is integrated with Huggingface transformers through bitsandbytes. We also support all other HuggingFace models. Please note that our code is still in development. We are looking for testers who are interested in experimenting with the code and reporting any bugs. We are particularly interested in help testing our examples and our own fine-tuning runs for new use-cases to debug edge-cases. Significant contributors will be acknowledged in the main repos and/or paper.

Your feedback and contributions will be invaluable to the release of the project!

Sign in to Google to save your progress. Learn more
Email *
Name *
Institution ("Independent" if you have no institution) *
Rate your open source skills and experience (10 is best) *
Getting started
Significant contributions to open source
Describe your experience with LLM finetuning. *
Describe your use-case or motivation for joining the private beta. *
Submit
Clear form
Never submit passwords through Google Forms.
This content is neither created nor endorsed by Google. Report Abuse - Terms of Service - Privacy Policy