Configure Your Model

Choose a VLM architecture and tune training settings for your use case in Datature Vi.

Before You Start
  • A workflow open in the workflow canvas
  • A configured dataset with train-test split set
  • A rough idea of your accuracy and speed requirements

Model configuration in Datature Vi has two parts: picking the right architecture for your task, then tuning the training and inference settings. Both happen inside the workflow canvas when you create a workflow.

For Qwen2.5-VL and Qwen3-VL, fine-tuning starts from instruction-tuned *-Instruct checkpoints only (base weights are not used); Qwen3.5 uses the post-trained Qwen3.5-{size} release, not -Base, and other architectures use the single published lines summarized on Model architectures.

1

Add a model to your workflow

Add a model to your workflow

In the workflow canvas, click the plus button below the Dataset nodes. Select a model architecture from the list. The model node appears on the canvas with default settings applied.

You should see
The workflow canvas Model node showing a selected architecture and parameter size with all settings configured

Your architecture selection is complete when the Model node on the workflow canvas displays your chosen architecture and size.

What to configure

Quick architecture guide

Use the table below to narrow down your choice before reading the full architecture comparison.

Quick architecture guide

Goal
Recommended model
Best overall performance
Qwen3.5 9B
Maximum accuracy, no resource limit
Qwen3.5 35B-A3B or Qwen3-VL 32B
Multilingual or long-context tasks
Qwen3.5 9B
Fast inference, edge deployment
NVILA-Lite 2B or Qwen3.5 0.8B
Complex multi-step reasoning
Cosmos-Reason1 7B
Physical world and spatial reasoning
Cosmos-Reason2 8B
Fine-grained visual understanding
InternVL3.5 8B
Proven general-purpose tasks (mature ecosystem)
Qwen2.5-VL 7B

Start with a smaller model (Qwen3.5 4B or Qwen3-VL 4B) to validate your dataset and pipeline with fast iteration cycles. Once your approach is working, scale up to Qwen3.5 9B or larger for production-grade accuracy.

Default vs. custom settings

Datature Vi's default settings work well for most first training runs. Start with defaults and adjust only if you see a specific issue: training not converging, outputs too short, or GPU memory errors.

The settings that move the needle most are:

  1. Model size: Larger models generally perform better but cost more to train
  2. Epochs: Too few means underfitting; too many means overfitting
  3. Learning rate: Affects how fast and how stably the model adapts to your data
  4. Training mode: LoRA trains faster and uses less memory; full fine-tuning can reach higher accuracy

Next steps

QLoRA Training Guide

Understand quantized LoRA, configure NF4 or FP4, and estimate VRAM for your model.

Full SFT Training Guide

Hardware requirements, multi-GPU setup, and when full fine-tuning is worth the cost.

Start A Training Run

Select GPU hardware and launch your configured workflow.