vllm/docs/source/models/supported_models.rst

41 lines
1.4 KiB
ReStructuredText
Raw Normal View History

2023-06-02 22:35:17 -07:00
.. _supported_models:
Supported Models
================
vLLM supports a variety of generative Transformer models in `HuggingFace Transformers <https://huggingface.co/models>`_.
2023-06-17 03:07:40 -07:00
The following is the list of model architectures that are currently supported by vLLM.
2023-06-02 22:35:17 -07:00
Alongside each architecture, we include some popular models that use it.
.. list-table::
:widths: 25 75
:header-rows: 1
* - Architecture
- Models
* - :code:`GPT2LMHeadModel`
- GPT-2
* - :code:`GPTNeoXForCausalLM`
- GPT-NeoX, Pythia, OpenAssistant, Dolly V2, StableLM
* - :code:`LlamaForCausalLM`
- LLaMA, Vicuna, Alpaca, Koala, Guanaco
2023-06-02 22:35:17 -07:00
* - :code:`OPTForCausalLM`
- OPT, OPT-IML
2023-06-17 03:07:40 -07:00
If your model uses one of the above model architectures, you can seamlessly run your model with vLLM.
2023-06-02 22:35:17 -07:00
Otherwise, please refer to :ref:`Adding a New Model <adding_a_new_model>` for instructions on how to implement support for your model.
2023-06-17 03:07:40 -07:00
Alternatively, you can raise an issue on our `GitHub <https://github.com/WoosukKwon/vllm/issues>`_ project.
2023-06-02 22:35:17 -07:00
.. tip::
The easiest way to check if your model is supported is to run the program below:
.. code-block:: python
2023-06-17 03:07:40 -07:00
from vllm import LLM
2023-06-02 22:35:17 -07:00
llm = LLM(model=...) # Name or path of your model
output = llm.generate("Hello, my name is")
print(output)
2023-06-17 03:07:40 -07:00
If vLLM successfully generates text, it indicates that your model is supported.