Skip to content

Commit f95e4fa

Browse files
committed
[fix] The cached token content is incorrectly truncated bug
1 parent fe01d73 commit f95e4fa

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

  • projects/llm_framework/main_llm/src/runner

projects/llm_framework/main_llm/src/runner/LLM.hpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -580,7 +580,7 @@ class LLM
580580
if (_attr.runing_callback)
581581
{
582582
cached_token.push_back(max_index);
583-
if (cached_token.size() >= 3)
583+
if (cached_token.size() >= 5)
584584
{
585585
float t_cost_ms = t_cost.cost();
586586
float token_per_sec = token_ids.size() / (t_cost_ms / 1000);

0 commit comments

Comments
 (0)