Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
14 changes: 8 additions & 6 deletions docs/source/models/supported_models.md
Original file line number Diff line number Diff line change
Expand Up @@ -1082,12 +1082,6 @@ You can enable it by passing `--mm-processor-kwargs '{"do_pan_and_scan": True}'`
:::

:::{warning}
### AllenAI Molmo-7B-D-0924 (multi-modal)

⚠️ Accuracy Note: For improved output quality (especially in object localization tasks), we recommend using the pinned dependency versions listed in [`requirements/molmo.txt`](https://github.com/vllm-project/vllm/blob/main/requirements/molmo.txt).
These versions match the environment that achieved consistent results on both A10 and L40 GPUs.
_Note: This setup currently works with `vllm==0.7.0`._

Both V0 and V1 support `Gemma3ForConditionalGeneration` for text-only inputs.
However, there are differences in how they handle text + image inputs:

Expand All @@ -1113,6 +1107,14 @@ This limitation exists because the model's mixed attention pattern (bidirectiona
To use `TIGER-Lab/Mantis-8B-siglip-llama3`, you have to pass `--hf_overrides '{"architectures": ["MantisForConditionalGeneration"]}'` when running vLLM.
:::

:::{warning}
AllenAI Molmo-7B-D-0924 (multi-modal)

⚠️ Accuracy Note: For improved output quality (especially in object localization tasks), we recommend using the pinned dependency versions listed in [`requirements/molmo.txt`](https://github.com/vllm-project/vllm/blob/main/requirements/molmo.txt).
These versions match the environment that achieved consistent results on both A10 and L40 GPUs.
_Note: This setup currently works with `vllm==0.7.0`._
:::

:::{note}
The official `openbmb/MiniCPM-V-2` doesn't work yet, so we need to use a fork (`HwwwH/MiniCPM-V-2`) for now.
For more details, please see: <gh-pr:4087#issuecomment-2250397630>
Expand Down