Incorporating current summary length in token count calculation for the update_running_summary function #4670
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Background
This PR addresses an oversight in my previous PR #4652, which introduced a new batch processing approach within the update_running_summary function. The goal of this approach is to prevent the total token count from exceeding the model's maximum limit when new events are being processed. While this method accounted for the token length of new events and the summarization prompt, it failed to consider the length of the current summary.
Changes
In the modified version, a summary_tlength variable has been introduced in the update_running_summary function. This variable holds the token length of the current summary, ensuring the total token count of three components: 1) the current summary, 2) the new events, and 3) the summarization prompt, does not surpass the maximum token limit imposed by the model.
Test Plan
Reran the tests and passed.
PR Quality Checklist