diff --git a/docs/source/models/supported_models.rst b/docs/source/models/supported_models.rst index 545e4182..e21cdd65 100644 --- a/docs/source/models/supported_models.rst +++ b/docs/source/models/supported_models.rst @@ -73,6 +73,10 @@ If your model uses one of the above model architectures, you can seamlessly run Otherwise, please refer to :ref:`Adding a New Model ` for instructions on how to implement support for your model. Alternatively, you can raise an issue on our `GitHub `_ project. +.. note:: + Currently, the ROCm version of vLLM does not support Mixtral. + Additionally, it only supports Mistral for context lengths up to 4096. + .. tip:: The easiest way to check if your model is supported is to run the program below: @@ -84,12 +88,17 @@ Alternatively, you can raise an issue on our `GitHub `_ instead of HuggingFace Hub, set an environment variable: .. code-block:: shell $ export VLLM_USE_MODELSCOPE=True + And use with :code:`trust_remote_code=True`. + .. code-block:: python from vllm import LLM @@ -97,5 +106,3 @@ Alternatively, you can raise an issue on our `GitHub