From 7c6009b76f04628b1617cec07c7d0bb766ca1009 Mon Sep 17 00:00:00 2001 From: Mattt394 <36588863+Mattt394@users.noreply.github.com> Date: Thu, 15 Feb 2024 10:20:10 +1300 Subject: [PATCH] experimental[patch]: Fixed typos in SmartLLMChain ideation and critique prompts (#11507) Noticed and fixed a few typos in the SmartLLMChain default ideation and critique prompts --------- Co-authored-by: Harrison Chase --- .../langchain_experimental/smart_llm/base.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/libs/experimental/langchain_experimental/smart_llm/base.py b/libs/experimental/langchain_experimental/smart_llm/base.py index d9d8929cb0..d69fa20801 100644 --- a/libs/experimental/langchain_experimental/smart_llm/base.py +++ b/libs/experimental/langchain_experimental/smart_llm/base.py @@ -215,7 +215,7 @@ class SmartLLMChain(Chain): HumanMessagePromptTemplate, "You are a researcher tasked with investigating the " f"{self.n_ideas} response options provided. List the flaws and " - "faulty logic of each answer options. Let'w work this out in a step" + "faulty logic of each answer option. Let's work this out in a step" " by step way to be sure we have all the errors:", ), ] @@ -229,10 +229,10 @@ class SmartLLMChain(Chain): HumanMessagePromptTemplate, "You are a resolver tasked with 1) finding which of " f"the {self.n_ideas} answer options the researcher thought was " - "best,2) improving that answer and 3) printing the answer in full. " - "Don't output anything for step 1 or 2, only the full answer in 3. " - "Let's work this out in a step by step way to be sure we have " - "the right answer:", + "best, 2) improving that answer and 3) printing the answer in " + "full. Don't output anything for step 1 or 2, only the full " + "answer in 3. Let's work this out in a step by step way to " + "be sure we have the right answer:", ), ] )