Catastrophic Forgetting
Sequential tuning on Harvard-FairVLMed then PathVQA shows only about a 5% drop for MedQwen, versus much larger drops for standard LoRA and MoELoRA.
MedQwen reports strong performance across medical VQA, report generation, zero-shot classification, and continual learning.
| Model | VQA-RAD | SLAKE | PathVQA | OMVQA | Avg. |
|---|---|---|---|---|---|
| Qwen-2.5-VL 7B | 61.8 / 27.2 | 64.7 / 36.7 | 60.5 / 33.4 | 60.8 | 49.3 |
| HealthGPT-L14 | 74.5 / 54.5 | 71.9 / 56.2 | 75.2 / 42.1 | 67.2 | 63.1 |
| MedQwen | 78.8 / 59.6 | 75.3 / 59.9 | 84.2 / 49.1 | 70.6 | 68.2 |
Sequential tuning on Harvard-FairVLMed then PathVQA shows only about a 5% drop for MedQwen, versus much larger drops for standard LoRA and MoELoRA.
MedQwen converges faster than LoRA-MoE baselines and narrows the gap to full fine-tuning as rank increases.
MedQwen answers medical questions across multiple modalities and generates medical reports for chest X-ray images.
On nine radiology benchmarks, MedQwen reaches 58.83 average accuracy and achieves about 95.31% of full FT MoE performance while using 339× fewer trainable parameters.
MedQwen improves over prior methods on MIMIC-CXR and IU-Xray, with strong gains in F1-RadGraph, BLEU-1, ROUGE, and CheXbert.
@article{nejati2026medqwen,
title = {Sparse Spectral LoRA: Routed Experts for Medical VLMs},
author = {Omid Nejati Manzari and Hojat Asgariandehkordi and Taha Koleilat and Yiming Xiao and Hassan Rivaz},
journal = {arXiv preprint},
year = {2026}
}