-
Notifications
You must be signed in to change notification settings - Fork 73
[Metrics] Handle vLLM streaming response in streaming server #518
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
JeffLuoo
commented
Mar 17, 2025
- Update streaming integration test when the response includes usage, the DONE message is returned together with the last message. The end of stream contains empty message.
✅ Deploy Preview for gateway-api-inference-extension ready!
To edit notification comments on pull requests, go to your Netlify site configuration. |
cacc9c8
to
d7b1ff6
Compare
- Update streaming integration test when the response includes usage, the DONE message is returned together with the last message. The end of stream contains empty message.
d7b1ff6
to
dba4a1b
Compare
/approve |
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: ahg-g, JeffLuoo The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
…tes-sigs#518) - Update streaming integration test when the response includes usage, the DONE message is returned together with the last message. The end of stream contains empty message.