Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add a choice of how to end streaming from callback: STOP or CANCEL #1476

Merged
merged 7 commits into from
Feb 9, 2025

Conversation

sbalandi
Copy link
Contributor

@sbalandi sbalandi commented Jan 3, 2025

No description provided.

@github-actions github-actions bot added category: visual language Visual language pipeline category: continuous batching Continuous batching category: LLM LLM pipeline (stateful, static) category: speculative decoding Speculative decoding category: GenAI C++ API Changes in GenAI C++ public headers no-match-files category: prompt lookup labels Jan 3, 2025
@sbalandi
Copy link
Contributor Author

sbalandi commented Jan 3, 2025

TODO: add CANCEL for ContinuousBatching

@ilya-lavrenov ilya-lavrenov added this to the 2025.0 milestone Jan 4, 2025
@ilya-lavrenov ilya-lavrenov self-assigned this Jan 6, 2025
@sbalandi sbalandi force-pushed the callback branch 5 times, most recently from 454cdd9 to 1592ed0 Compare January 8, 2025 19:38
@github-actions github-actions bot added category: Python API Python API for GenAI category: samples GenAI samples labels Jan 8, 2025
@sbalandi sbalandi force-pushed the callback branch 3 times, most recently from 10a755b to d18fe16 Compare January 8, 2025 22:19
@sbalandi
Copy link
Contributor Author

sbalandi commented Jan 8, 2025

TODO: add CANCEL for ContinuousBatching

done

@sbalandi sbalandi marked this pull request as ready for review January 8, 2025 22:43
@sbalandi sbalandi force-pushed the callback branch 3 times, most recently from 2758f6b to 03ca3ce Compare January 9, 2025 21:56
Copy link
Contributor

@ilya-lavrenov ilya-lavrenov left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please, add tests for new functionality.

@andrei-kochin andrei-kochin modified the milestones: 2025.0, 2025.1 Jan 13, 2025
@sbalandi sbalandi force-pushed the callback branch 3 times, most recently from fb83926 to e75b81a Compare February 5, 2025 19:15
@sbalandi
Copy link
Contributor Author

sbalandi commented Feb 6, 2025

  • I was wrong about tests for StreamerBase with exactly bool return value. There are tests with put() with None to return. I have added with bool and with status.
  • It was added two tests: test_callback_terminate_by_bool and test_callback_terminate_by_status, which checks that output generated seq stops on exactly the token that the user requested via the streamer, but there is one extra token in generation sequence now. I would remove it in another PR, if it's required (need to be discussed)

@sbalandi sbalandi force-pushed the callback branch 2 times, most recently from 62c1460 to 0ceac52 Compare February 7, 2025 10:58
@ilya-lavrenov
Copy link
Contributor

Please, resolve conflicts.

@ilya-lavrenov ilya-lavrenov added this pull request to the merge queue Feb 9, 2025
@github-merge-queue github-merge-queue bot removed this pull request from the merge queue due to failed status checks Feb 9, 2025
@ilya-lavrenov ilya-lavrenov added this pull request to the merge queue Feb 9, 2025
Merged via the queue into openvinotoolkit:master with commit f93b92b Feb 9, 2025
62 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
category: continuous batching Continuous batching category: GenAI C++ API Changes in GenAI C++ public headers category: LLM LLM pipeline (stateful, static) category: prompt lookup category: Python API Python API for GenAI category: samples GenAI samples category: speculative decoding Speculative decoding category: visual language Visual language pipeline category: whisper Whisper pipeline no-match-files
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants