diff --git a/docs/source/models/supported_models.rst b/docs/source/models/supported_models.rst index 7a9c87f406c66..223c68b40766e 100644 --- a/docs/source/models/supported_models.rst +++ b/docs/source/models/supported_models.rst @@ -234,10 +234,10 @@ Multimodal Language Models - Image - :code:`openbmb/MiniCPM-V-2` (see note), :code:`openbmb/MiniCPM-Llama3-V-2_5`, :code:`openbmb/MiniCPM-V-2_6`, etc. - - * - :code: `UltravoxModel` + * - :code:`UltravoxModel` - Ultravox - Audio - - :code: `fixie-ai/ultravox-v0_3` + - :code:`fixie-ai/ultravox-v0_3` - .. note:: diff --git a/examples/offline_inference_audio_language.py b/examples/offline_inference_audio_language.py index 7b886f8e2001a..56ce8646c20c9 100644 --- a/examples/offline_inference_audio_language.py +++ b/examples/offline_inference_audio_language.py @@ -1,6 +1,6 @@ """ This example shows how to use vLLM for running offline inference -with the correct prompt format on vision language models. +with the correct prompt format on audio language models. For most models, the prompt format should follow corresponding examples on HuggingFace model repository.