29 lines
610 B
Markdown
Raw Permalink Normal View History

(multi-modality)=
# Multi-Modality
2024-06-03 13:56:41 +08:00
vLLM provides experimental support for multi-modal models through the {mod}`vllm.multimodal` package.
Multi-modal inputs can be passed alongside text and token prompts to [supported models](#supported-mm-models)
via the `multi_modal_data` field in {class}`vllm.inputs.PromptType`.
2024-06-03 13:56:41 +08:00
Looking to add your own multi-modal model? Please follow the instructions listed [here](#supports-multimodal).
## Module Contents
2024-06-03 13:56:41 +08:00
```{eval-rst}
.. autodata:: vllm.multimodal.MULTIMODAL_REGISTRY
```
2024-06-03 13:56:41 +08:00
## Submodules
:::{toctree}
:maxdepth: 1
inputs
parse
processing
profiling
registry
:::