Post

Conversation

Google just dropped a new LLM! You can run it locally on just 0.5 GB RAM. Let's fine-tune this on our own data (100% locally):
David Watson 🥑
Post your reply

Google released Gemma 3 270M, a new model for hyper-efficient local AI! We'll fine-tune this model and make it very smart at playing chess and predict the next move. Tech stack: - for efficient fine-tuning. - transformers to run it locally. Let's go! 🚀
0:01 / 0:12
2️⃣ Define LoRA config We'll use LoRA for efficient fine-tuning. To do this, we use Unsloth's PEFT and specify: - The model - LoRA low-rank (r) - Layers for fine-tuning (target_modules) Check this code 👇
Image
3️⃣ Load dataset We'll fine-tune Gemma 3 to make it extremly smart at playing chess. Given a set of previous move (one move missing) & the final result it has to predict the missing move. In order to do this we're using the ChessInstruct dataset from HuggingFace. Check this👇
Image
4️⃣ Prepare dataset Next, we use a conversation style dataset to fine-tune Gemma 3. The standardize_data_formats method converts the dataset to the correct format for finetuning purposes!
Image
5️⃣ Define Trainer Here, we create a Trainer object by specifying the training config, like learning rate, model, tokenizer, and more. Check this out 👇
Image
6️⃣ Train With that done, we initiate training. The loss is generally decreasing with steps, which means the model is being fine-tuned correctly. Check this code and training logs 👇
Image
Finally, the video shows prompting the LLM before and after fine-tuning. After fine-tuning, the model is able to find the exact missing chess move instead of randomly generating some moves. Check this 👇
0:06
If you found it insightful, reshare with your network. Find me → ✔️ For more insights and tutorials on LLMs, AI Agents, and Machine Learning!
Quote
Akshay 🚀
@akshay_pachaar
Google just dropped a new LLM! You can run it locally on just 0.5 GB RAM. Let's fine-tune this on our own data (100% locally):
0.5 GB on a light weight model is completely fixated on fine tuning to your own needs, this is what we need in the open source!
Serious question from someone without deep understanding of this stuff: could that run in a browser window as long as GPU support (webasm?) is provided?
Brilliant guide to fine-tuning, Akshay. I misinterpreted 270M as 270B when I heard about the release, since hardly any powerful LLM comes with parameters in "M" these days :D. The level of performance compared to its scale is great.
Thanks for the demo / PoC. What kind of setup/bootstrap/data would I need to fine tune this for spam/cold outreach email detection?
That is awesome this Google small language model looks like it can even run on a mobile. I know they have been going this route with AI models its fantastic this Gemma update looks great, gives us control of a model that can run about anywhere!
Square profile picture
Imagine if someone made a car that needed almost no maintenance Has the lowest repair costs of any brand And can literally drive you almost anywhere with Full Self-Driving (Supervised) We did And they’re all S3XY
Wow! 🤯 Running a full LLM locally on just 0.5 GB RAM is amazing. Can’t wait to see what fine-tuning on custom data can do—fast, private, and powerful AI experiments ahead!
I enjoyed reading this. Would you recommend a similar approach for fine tuning a model for classifying text (spam/ham) or would something like Facebook’s fastText still be the recommended approach?
yeah don't run this thing nobody should be using a 270m model lmao please at the very minimum qwen 3 0.6b
Since this model is so small, is it possible to fine tune it for classification tasks like sentiment analysis and NER?
Is this really powerful for its size? And which use cases that'll be suitable for this?
Just the right topic for the timeline. Would it be possible to train such a small model for blender-specific Python? Would that even make sense? Thank you for your work.
So is there a reasonable way for me to put this on my phone and then hook it up as an assistant service and finetune to fit the purpose?
Blow to high value nvidia chips and large jumbo data centers ? Remember when deepseek was released on February 202 - What happened ??
Lego AI models. Make the finetuning simple as adding a chrome extension to a small model, and anyone can customize their own AI.
Square profile picture
Teslas have the lowest maintenance & repair costs of any brand
Rate proposed Community Notes
using these kind of small models ( Gemma 3 270M) on a well know problem ( like chess) is a great way to learn the fine tuning of LLMs.
Quote
Akshay 🚀
@akshay_pachaar
Replying to @akshay_pachaar
Google released Gemma 3 270M, a new model for hyper-efficient local AI! We'll fine-tune this model and make it very smart at playing chess and predict the next move. Tech stack: - @UnslothAI for efficient fine-tuning. - @huggingface transformers to run it locally. Let's go! 🚀