Note: Please do not remove the questions. Answer beside them.
- Did you update?
pip install --upgrade unsloth unsloth_zoo
I am running a Docker image with tag 2026.4.2-pt2.9.0-vllm-0.16.0-cu12.8-studio-release-v0.1.35-beta
Colab or Kaggle or local / cloud
Local
- Number GPUs used, use
nvidia-smi
A single H100 GPU
- Which notebook? Please link!
Unsloth Studio
- Which Unsloth version, TRL version, transformers version, PyTorch version?
Unsloth 2026.4.2, TRL 0.23.1, Transformers 4.57.1, PyTorch 2.9.1+cu128
- Which trainer?
SFTTrainer, GRPOTrainer etc
I assume it is SFT by default, but there is no visualization in Studio.
I tried my custom .parquet dataset with images and text columns, as well as unsloth/LaTeX_OCR example, but both of them produce the same Text model is not compatible with a multimodal dataset. Switch to a vision model or choose a text-only dataset. on unsloth/Qwen3.5-4B and unsloth/gemma-4-E2B-it even though they both have vision capabilities. However, unsloth/Qwen3-VL-4B-Instruct worked fine and was successfully trained on the dataset.
Note: Please do not remove the questions. Answer beside them.
pip install --upgrade unsloth unsloth_zooI am running a Docker image with tag 2026.4.2-pt2.9.0-vllm-0.16.0-cu12.8-studio-release-v0.1.35-beta
ColaborKaggleor local / cloudLocal
nvidia-smiA single H100 GPU
Unsloth Studio
Unsloth 2026.4.2, TRL 0.23.1, Transformers 4.57.1, PyTorch 2.9.1+cu128
SFTTrainer,GRPOTraineretcI assume it is SFT by default, but there is no visualization in Studio.
I tried my custom .parquet dataset with images and text columns, as well as unsloth/LaTeX_OCR example, but both of them produce the same
Text model is not compatible with a multimodal dataset. Switch to a vision model or choose a text-only dataset.on unsloth/Qwen3.5-4B and unsloth/gemma-4-E2B-it even though they both have vision capabilities. However, unsloth/Qwen3-VL-4B-Instruct worked fine and was successfully trained on the dataset.