You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The OpenAI API (and by extension vLLM's completions functionality) supports configuring the number of logprobs-per-token to return at the granularity of each request, via the logprobs argument. However, benchmarks/benchmark_serving.py currently does not configure the logprobs argument when generating requests, nor does benchmarks/benchmark_serving.py have a --logprobs CLI argument. This is an issue because it is desirable to benchmark the impact of different --logprobs settings on vLLM performance.
So the issue is that (1) benchmarks/benchmark_serving.py should support a --logprobs argument, and (2) the value of the --logprobs CLI argument should configure the underlying logprobs argument to the completion requests generated during benchmarking.
shows how to configure OpenAI API requests with the logprobs argument set.
Additional context
No response
Before submitting a new issue...
Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
The text was updated successfully, but these errors were encountered:
🚀 The feature, motivation and pitch
The OpenAI API (and by extension vLLM's completions functionality) supports configuring the number of logprobs-per-token to return at the granularity of each request, via the
logprobs
argument. However,benchmarks/benchmark_serving.py
currently does not configure thelogprobs
argument when generating requests, nor doesbenchmarks/benchmark_serving.py
have a--logprobs
CLI argument. This is an issue because it is desirable to benchmark the impact of different--logprobs
settings on vLLM performance.So the issue is that (1)
benchmarks/benchmark_serving.py
should support a--logprobs
argument, and (2) the value of the--logprobs
CLI argument should configure the underlyinglogprobs
argument to the completion requests generated during benchmarking.Alternatives
In
tests/utils.py
, the functionshows how to configure OpenAI API requests with the logprobs argument set.
Additional context
No response
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: