vLLM Blog

vLLM is a fast and easy-to-use library for LLM inference and serving.

Serving Geospatial, Vision, and Beyond: Enabling Multimodal Output Processing in vLLM

Christian Pinto (IBM Research Europe - Dublin), Michele Gazzetti (IBM Research Europe - Dublin), Michael Johnston (IBM Research Europe - Dublin), Maximilien Philippe Marie de Bayser (IBM Research - Brazil)

Introduction Until recently, generative AI infrastructure has been tightly coupled with autoregressive text generation models that produce output token-by-token, typically in the form of natural...