Zagora

QLoRA Fine-Tuning Beta

Train a lightweight QLoRA adapter on your dataset using SFT (Supervised Fine-Tuning).

LoRA Rank 8 Alpha 16 NF4 Quantization SFT
We'll notify you when your job is complete
Paste a Google Drive, Dropbox, or HuggingFace link. Always include https://
Supported Dataset Formats
Chat (recommended)
{"messages": [{"role": "user", "content": "..."}, {"role": "assistant", "content": "..."}]}
Instruction
{"instruction": "Your prompt here", "response": "Expected response"}
Pre-formatted
{"text": "Complete formatted text as a single string"}
DPO (preference tuning)
{"chosen": [{"role": "user", "content": "..."}, {"role": "assistant", "content": "..."}],
 "rejected": [...], "ref_chosen_logps": -1.5, "ref_rejected_logps": -3.2}
Format is auto-detected from your JSONL. One JSON object per line.