diff --git a/vllm/entrypoints/openai/serving_completion.py b/vllm/entrypoints/openai/serving_completion.py index cd5385afa875f..34f1200753f8d 100644 --- a/vllm/entrypoints/openai/serving_completion.py +++ b/vllm/entrypoints/openai/serving_completion.py @@ -110,7 +110,6 @@ async def create_completion( )) for i, prompt_inputs in enumerate(prompts): - print(guided_decode_logits_processor) sampling_params = request.to_sampling_params( tokenizer, guided_decode_logits_processor,