๐Welcome
New to Unsloth?
Unsloth makes finetuning large language models like Llama-3, Mistral, Phi-4 and Gemma 2x faster, use 70% less memory, and with no degradation in accuracy! Our docs will guide you through training your own custom model. It covers the essentials of installing & updating Unsloth, creating datasets, running & deploying your model.
โญBeginner? Start here!๐งฌFine-tuning Guide๐กReasoning - GRPO & RL๐Unsloth Notebooks
What is finetuning and why?
Fine-tuning an LLM customizes its behavior, enhances domain knowledge, and optimizes performance for specific tasks. Finetuning is the process of updating the actual "brains" of the language model through some process called back-propagation.
By fine-tuning a pre-trained model (e.g. Llama-3.1-8B) on a specialized dataset, you can:
Update Knowledge: Introduce new domain-specific information.
Customize Behavior: Adjust the modelโs tone, personality, or response style.
Optimize for Tasks: Improve accuracy and relevance for specific use cases.
Example usecases:
Train LLM to predict if a headline impacts a company positively or negatively.
Use historical customer interactions for more accurate and custom responses.
Fine-tune LLM on legal texts for contract analysis, case law research, and compliance.
You can think of a fine-tuned model as a specialized agent designed to do specific tasks more effectively and efficiently. Fine-tuning can replicate all of RAG's capabilities, but not vice versa.
๐คFAQ + Is Fine-tuning Right For Me?How to use Unsloth?
Unsloth can be installed locally via Linux, Windows (via WSL), Kaggle, or another GPU service like Google Colab. Most use Unsloth through the interface Google Colab which provides a free GPU to train with.
๐ฅInstalling + Updating๐ ๏ธUnsloth RequirementsLast updated
Was this helpful?