A completions API request with an empty prompt will crash the vllm API server. The impact is limited based on what model is being served. * Serving gpt2 is affected. * Most models are not affected, as vllm will prepend tokens to the prompt, avoiding the problematic code. https://github.com/vllm-project/vllm/commit/e25fee57c2e69161bd261f5986dc5aeb198bbd42