This website requires JavaScript.
Explore
Help
Register
Sign In
20231088
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
1
Packages
Projects
Releases
Wiki
Activity
vllm
/
docs
/
source
/
models
History
Roger Wang
4bb98f2190
[Misc] Update config loading for Qwen2-VL and remove Granite (
#8837
)
2024-09-26 07:45:30 -07:00
..
adding_model.rst
[Doc][CI/Build] Update docs and tests to use
vllm serve
(
#6431
)
2024-07-17 07:43:21 +00:00
enabling_multimodal_inputs.rst
[VLM][Core] Support profiling with multiple multi-modal inputs per prompt (
#7126
)
2024-08-14 17:55:42 +00:00
engine_args.rst
[Doc][CI/Build] Update docs and tests to use
vllm serve
(
#6431
)
2024-07-17 07:43:21 +00:00
lora.rst
[Core] Support Lora lineage and base model metadata management (
#6315
)
2024-09-20 06:20:56 +00:00
performance.rst
[Scheduler] Warning upon preemption and Swapping (
#4647
)
2024-05-13 23:50:44 +09:00
spec_decode.rst
[Documentation][Spec Decode] Add documentation about lossless guarantees in Speculative Decoding in vLLM (
#7962
)
2024-09-05 16:25:29 -04:00
supported_models.rst
[Misc] Update config loading for Qwen2-VL and remove Granite (
#8837
)
2024-09-26 07:45:30 -07:00
vlm.rst
Revert "rename PromptInputs and inputs with backward compatibility (
#8760
) (
#8810
)
2024-09-25 10:36:26 -07:00