| | Bitwise Consistent On-Policy Reinforcement Learning with VLLM and TorchTitan (vllm.ai) |
| 1 point by brrrrrm 29 days ago | past |
|
| | vLLM TPU: A New Unified Backend Supporting PyTorch and JAX on TPU (vllm.ai) |
| 1 point by pykello 55 days ago | past |
|
| | VLLM TPU: A New Unified Back End Supporting PyTorch and Jax on TPU (vllm.ai) |
| 1 point by alphabetting 56 days ago | past |
|
| | Inside vLLM: Anatomy of a High-Throughput LLM Inference System (vllm.ai) |
| 2 points by matt_d 3 months ago | past |
|
| | vLLM with torch.compile: Efficient LLM inference on PyTorch (vllm.ai) |
| 1 point by matt_d 3 months ago | past |
|
| | VLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention (vllm.ai) |
| 20 points by jxmorris12 5 months ago | past | 5 comments |
|
| | VLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention (2023) (vllm.ai) |
| 3 points by telotortium 8 months ago | past |
|
| | vLLM V1: A Major Upgrade to vLLM's Core Architecture (vllm.ai) |
| 2 points by ozgune 10 months ago | past |
|
| | vLLM V1: A Major Upgrade to vLLM's Core Architecture (vllm.ai) |
| 5 points by xmo 10 months ago | past |
|
| | VLLM 2024 Retrospective and 2025 Vision (vllm.ai) |
| 1 point by shenli3514 10 months ago | past |
|
| | Installing and Developing VLLM with Ease (vllm.ai) |
| 1 point by brethil 11 months ago | past |
|
| | vLLM v0.6.0: 2.7x Throughput Improvement and 5x Latency Reduction (vllm.ai) |
| 3 points by xmo on Sept 5, 2024 | past |
|
| | VLLM automatic prefix / prompt caching (vllm.ai) |
| 2 points by danielhanchen on Aug 25, 2024 | past | 1 comment |
|
| | VLLM hosts local LLMs easily (vllm.ai) |
| 2 points by myprotegeai on Aug 4, 2024 | past |
|
| | Llama 3.1 Support in VLLM (vllm.ai) |
| 2 points by e12e on July 23, 2024 | past |
|
| | vLLM (vllm.ai) |
| 2 points by jonbaer on April 24, 2024 | past |
|
| | VLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention (vllm.ai) |
| 2 points by udev4096 on Jan 7, 2024 | past |
|
| | Notes on VLLM v.s. DeepSpeed-FastGen (vllm.ai) |
| 3 points by Palmik on Nov 15, 2023 | past |
|
| | vLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention (vllm.ai) |
| 295 points by wskwon on June 20, 2023 | past | 42 comments |
|