From 484707ad29e173ee940e1d480d6408b7b7d2168a Mon Sep 17 00:00:00 2001 From: Zander Chase <130414180+vowelparrot@users.noreply.github.com> Date: Mon, 1 May 2023 10:36:54 -0700 Subject: [PATCH] Add incremental messages token count (#3890) --- langchain/experimental/autonomous_agents/autogpt/prompt.py | 1 + 1 file changed, 1 insertion(+) diff --git a/langchain/experimental/autonomous_agents/autogpt/prompt.py b/langchain/experimental/autonomous_agents/autogpt/prompt.py index 909e9b0c..d1f3a9b7 100644 --- a/langchain/experimental/autonomous_agents/autogpt/prompt.py +++ b/langchain/experimental/autonomous_agents/autogpt/prompt.py @@ -70,6 +70,7 @@ class AutoGPTPrompt(BaseChatPromptTemplate, BaseModel): if used_tokens + message_tokens > self.send_token_limit - 1000: break historical_messages = [message] + historical_messages + used_tokens += message_tokens input_message = HumanMessage(content=kwargs["user_input"]) messages: List[BaseMessage] = [base_prompt, time_prompt, memory_message] messages += historical_messages