Qwen 3.5 Fine-Tuning Practical Guide — Build Your Own Model with LoRA
Complete guide to fine-tuning Qwen 3.5 with LoRA/QLoRA. From 8GB GPU QLoRA setup to Unsloth optimization, GGUF conversion, and Ollama deployment.

Qwen 3.5 Fine-Tuning Practical Guide — Build Your Own Model with LoRA
In the previous post, we covered installing and running Qwen 3.5 locally. Now let's go one step further: fine-tuning the model with your own data.
With LoRA/QLoRA, you can fine-tune Qwen 3.5 on consumer GPUs. This guide covers the entire process from data preparation to training, evaluation, and deployment.
1. Why Fine-Tune?
Qwen 3.5 is a general-purpose model. It handles most tasks well, but fine-tuning is needed when:
Related Posts

DeerFlow 2.0 Production Deployment — Docker Compose, Kubernetes, Message Gateways
Deploy DeerFlow to production with Docker Compose and Kubernetes. Connect Slack/Telegram message gateways for team access.

DeerFlow 2.0 Custom Skills + MCP + Sandbox — Building Your Own Tools and Workflows
DeerFlow's markdown-based skills system, MCP server integration, Docker/K8s sandbox, and persistent memory system with practical code examples.

DeerFlow 2.0 Multi-Agent Workflow Deep Dive — StateGraph, Plan-Execute, Human-in-the-Loop
Code-level analysis of DeerFlow's LangGraph StateGraph-based Multi-Agent Workflow. Supervisor routing, Plan-Execute pattern, and dynamic sub-agent spawning.