AI Models & Infra
vLLM: Easy, Fast, and Cheap LLM Serving for Everyone
Speakers
Presentation Slides
Presentation Video
vLLM is a fast and easy-to-use library for LLM inference and serving. In this talk, I will briefly introduce the evolution of the vLLM project, the open-source community behind it, and highlight some features that are interesting to many users.