Skip to content

[ML] Append all data to Chat Completion buffer (#127658) #128136

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 2 commits into from
May 19, 2025

Conversation

prwhelan
Copy link
Member

Backport #127658

Moved the Chat Completion buffer into the
StreamingUnifiedChatCompletionResults so that all Chat Completion
responses can benefit from it. Chat Completions is meant to adhere to
OpenAI as much as possible, and OpenAI only sends one response chunk at
a time. All implementations of Chat Completions will now buffer.

This fixes a bug where more than two chunks in a single item would be
dropped, instead they are all added to the buffer.

This fixes a bug where onComplete would omit trailing items in the
buffer.
@prwhelan prwhelan added >bug :ml Machine learning backport Team:ML Meta label for the ML team auto-merge-without-approval Automatically merge pull request when CI checks pass (NB doesn't wait for reviews!) v8.19.0 labels May 19, 2025
@prwhelan prwhelan marked this pull request as ready for review May 19, 2025 16:24
@elasticsearchmachine elasticsearchmachine merged commit 8580981 into elastic:8.19 May 19, 2025
15 checks passed
@prwhelan prwhelan deleted the backport/8.19/127658 branch May 19, 2025 17:44
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
auto-merge-without-approval Automatically merge pull request when CI checks pass (NB doesn't wait for reviews!) backport >bug :ml Machine learning Team:ML Meta label for the ML team v8.19.0
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants