๐Ÿ‘‹Welcome

New to Unsloth?

Unsloth makes finetuning large language models like Llama-3, Mistral, Phi-4 and Gemma 2x faster, use 70% less memory, and with no degradation in accuracy! Our docs will guide you through training your own custom model. It covers the essentials of installing & updating Unsloth, creating datasets, running & deploying your model.

โญBeginner? Start here!๐ŸงฌFine-tuning Guide๐Ÿ’กReasoning - GRPO & RL๐Ÿ“’Unsloth Notebooks

What is finetuning and why?

Fine-tuning an LLM customizes its behavior, enhances domain knowledge, and optimizes performance for specific tasks. Finetuning is the process of updating the actual "brains" of the language model through some process called back-propagation.

By fine-tuning a pre-trained model (e.g. Llama-3.1-8B) on a specialized dataset, you can:

  • Update Knowledge: Introduce new domain-specific information.

  • Customize Behavior: Adjust the modelโ€™s tone, personality, or response style.

  • Optimize for Tasks: Improve accuracy and relevance for specific use cases.

Example usecases:

  • Train LLM to predict if a headline impacts a company positively or negatively.

  • Use historical customer interactions for more accurate and custom responses.

  • Fine-tune LLM on legal texts for contract analysis, case law research, and compliance.

You can think of a fine-tuned model as a specialized agent designed to do specific tasks more effectively and efficiently. Fine-tuning can replicate all of RAG's capabilities, but not vice versa.

๐Ÿค”FAQ + Is Fine-tuning Right For Me?

How to use Unsloth?

Unsloth can be installed locally via Linux, Windows (via WSL), Kaggle, or another GPU service like Google Colab. Most use Unsloth through the interface Google Colab which provides a free GPU to train with.

๐Ÿ“ฅInstalling + Updating๐Ÿ› ๏ธUnsloth Requirements

Last updated

Was this helpful?