mirror of
https://github.com/hwchase17/langchain
synced 2024-11-08 07:10:35 +00:00
ebf998acb6
Co-authored-by: Harrison Chase <hw.chase.17@gmail.com> Co-authored-by: Lance Martin <lance@langchain.dev> Co-authored-by: Jacob Lee <jacoblee93@gmail.com>
50 lines
1.6 KiB
Python
50 lines
1.6 KiB
Python
from langchain.prompts import ChatPromptTemplate, MessagesPlaceholder
|
|
from langchain.schema import AgentAction, AgentFinish
|
|
|
|
template = """You are a helpful assistant. Help the user answer any questions.
|
|
|
|
You have access to the following tools:
|
|
|
|
{tools}
|
|
|
|
In order to use a tool, you can use <tool></tool> and <tool_input></tool_input> tags. You will then get back a response in the form <observation></observation>
|
|
For example, if you have a tool called 'search' that could run a google search, in order to search for the weather in SF you would respond:
|
|
|
|
<tool>search</tool><tool_input>weather in SF</tool_input>
|
|
<observation>64 degrees</observation>
|
|
|
|
When you are done, you can respond as normal to the user.
|
|
|
|
Example 1:
|
|
|
|
Human: Hi!
|
|
|
|
Assistant: Hi! How are you?
|
|
|
|
Human: What is the weather in SF?
|
|
Assistant: <tool>search</tool><tool_input>weather in SF</tool_input>
|
|
<observation>64 degrees</observation>
|
|
It is 64 degress in SF
|
|
|
|
|
|
Begin!"""
|
|
|
|
conversational_prompt = ChatPromptTemplate.from_messages([
|
|
("system", template),
|
|
MessagesPlaceholder(variable_name="chat_history"),
|
|
("user", "{question}"),
|
|
("ai", "{agent_scratchpad}")
|
|
])
|
|
|
|
|
|
def parse_output(message):
|
|
text = message.content
|
|
if "</tool>" in text:
|
|
tool, tool_input = text.split("</tool>")
|
|
_tool = tool.split("<tool>")[1]
|
|
_tool_input = tool_input.split("<tool_input>")[1]
|
|
if "</tool_input>" in _tool_input:
|
|
_tool_input = _tool_input.split("</tool_input>")[0]
|
|
return AgentAction(tool=_tool, tool_input=_tool_input, log=text)
|
|
else:
|
|
return AgentFinish(return_values={"output": text}, log=text) |