This website requires JavaScript.
Explore
Help
Register
Sign In
20231088
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
1
Packages
Projects
Releases
Wiki
Activity
vllm
/
docs
/
source
/
models
History
Cyrus Leung
ae96ef8fbd
[VLM] Calculate maximum number of multi-modal tokens by model (
#6121
)
2024-07-04 16:37:23 -07:00
..
adding_model.rst
[Core] Registry for processing model inputs (
#5214
)
2024-06-28 12:09:56 +00:00
engine_args.rst
Don't show default value for flags in
EngineArgs
(
#4223
)
2024-04-21 09:15:28 -07:00
lora.rst
[Model] Add base class for LoRA-supported models (
#5018
)
2024-06-27 16:03:04 +08:00
performance.rst
[Scheduler] Warning upon preemption and Swapping (
#4647
)
2024-05-13 23:50:44 +09:00
spec_decode.rst
[Docs] [Spec decode] Fix docs error in code example (
#5427
)
2024-06-11 10:31:56 -07:00
supported_models.rst
[Model] Jamba support (
#4115
)
2024-07-02 23:11:29 +00:00
vlm.rst
[VLM] Calculate maximum number of multi-modal tokens by model (
#6121
)
2024-07-04 16:37:23 -07:00