Fix: Echo without asking for new tokens or logprobs in OpenAI Completions API#2995
Fix: Echo without asking for new tokens or logprobs in OpenAI Completions API#2995matheper wants to merge 10 commits intovllm-project:mainfrom
Conversation
|
Sorry for the long wait! From a quick search, it appears that since then, a related PR (#3441) has been merged, but it does not fully address the cases mentioned in this PR. If you're still interested, update your code and I'll help get your changes merged. |
|
This pull request has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this pull request should remain open. Thank you! |
|
This pull request has merge conflicts that must be resolved before it can be |
Hi,
I've noticed that a recent modification to the OpenAI entrypoint has made these two use cases inoperable, which were previously supported by vLLM:
This PR adds these use cases back by adding the prompt to the engine.generate, resolving an issue of referencing logprobs when they are non-existent, and fixes the usage calculation. Closes #2706.
Here are the error logs of both use cases prior to this PR: