From c1026311b59446d1ada5f950ddbdbe0bb21943b0 Mon Sep 17 00:00:00 2001 From: Zhuohan Li Date: Mon, 18 Sep 2023 12:23:35 -0700 Subject: [PATCH] [Community] Add vLLM Discord server (#1086) --- README.md | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index ab500728..5e05f1dc 100644 --- a/README.md +++ b/README.md @@ -10,13 +10,14 @@ Easy, fast, and cheap LLM serving for everyone

-| Documentation | Blog | Paper | Discussions | +| Documentation | Blog | Paper | Discord |

--- *Latest News* 🔥 +- [2023/09] We created our [Discord server](https://discord.gg/jz7wjKhh6g)! Join us to discuss vLLM and LLM serving! We will also post the latest announcements and updates there. - [2023/09] We released our [PagedAttention paper](https://arxiv.org/abs/2309.06180) on arXiv! - [2023/08] We would like to express our sincere gratitude to [Andreessen Horowitz](https://a16z.com/2023/08/30/supporting-the-open-source-ai-community/) (a16z) for providing a generous grant to support the open-source development and research of vLLM. - [2023/07] Added support for LLaMA-2! You can run and serve 7B/13B/70B LLaMA-2s on vLLM with a single command! @@ -111,7 +112,7 @@ Please check out [CONTRIBUTING.md](./CONTRIBUTING.md) for how to get involved. If you use vLLM for your research, please cite our [paper](https://arxiv.org/abs/2309.06180): ```bibtex @inproceedings{kwon2023efficient, - title={Efficient Memory Management for Large Language Model Serving with PagedAttention}, + title={Efficient Memory Management for Large Language Model Serving with PagedAttention}, author={Woosuk Kwon and Zhuohan Li and Siyuan Zhuang and Ying Sheng and Lianmin Zheng and Cody Hao Yu and Joseph E. Gonzalez and Hao Zhang and Ion Stoica}, booktitle={Proceedings of the ACM SIGOPS 29th Symposium on Operating Systems Principles}, year={2023}