This website requires JavaScript.
Explore
Help
Register
Sign In
20231088
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
1
Packages
Projects
Releases
Wiki
Activity
vllm
/
vllm
/
model_executor
History
JFDuan
0d93f15694
Accelerate LLaMA model loading (
#234
)
2023-08-30 01:00:13 -07:00
..
layers
use flash-attn via xformers (
#877
)
2023-08-29 21:52:13 -07:00
models
Accelerate LLaMA model loading (
#234
)
2023-08-30 01:00:13 -07:00
parallel_utils
Add Falcon support (new) (
#592
)
2023-08-02 14:04:39 -07:00
__init__.py
[Quality] Add code formatter and linter (
#326
)
2023-07-03 11:31:55 -07:00
input_metadata.py
Add support for BLOOM (
#331
)
2023-07-03 13:12:35 -07:00
model_loader.py
Add support for aquila (
#663
)
2023-08-22 00:13:36 -07:00
utils.py
Change the name to vLLM (
#150
)
2023-06-17 03:07:40 -07:00
weight_utils.py
Accelerate LLaMA model loading (
#234
)
2023-08-30 01:00:13 -07:00