0 bookmark(s) - Sort by: Date ↓ / Title /
This document details how to run and fine-tune Gemma 3 models (1B, 4B, 12B, and 27B) using Unsloth, covering setup with Ollama and llama.cpp, and addressing potential float16 precision issues. It also highlights Unsloth's unique ability to run Gemma 3 in float16 on machines like Colab notebooks with Tesla T4 GPUs.
This article provides a comprehensive guide on fine-tuning the Llama 3.1 language model using Unsloth for efficient parameter-efficient training. It covers concepts like supervised fine-tuning, LoRA, QLoRA, and practical steps for training on a high-quality dataset.
First / Previous / Next / Last
/ Page 1 of 0