2023-06-17 03:07:40 -07:00
|
|
|
Welcome to vLLM!
|
|
|
|
================
|
2023-05-22 17:02:44 -07:00
|
|
|
|
2023-06-18 01:26:12 +08:00
|
|
|
vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLM).
|
|
|
|
|
2023-05-22 17:02:44 -07:00
|
|
|
Documentation
|
|
|
|
-------------
|
|
|
|
|
|
|
|
.. toctree::
|
|
|
|
:maxdepth: 1
|
|
|
|
:caption: Getting Started
|
|
|
|
|
|
|
|
getting_started/installation
|
|
|
|
getting_started/quickstart
|
2023-06-02 22:35:17 -07:00
|
|
|
|
|
|
|
.. toctree::
|
|
|
|
:maxdepth: 1
|
|
|
|
:caption: Models
|
|
|
|
|
|
|
|
models/supported_models
|
|
|
|
models/adding_model
|