mirror of
https://github.com/hwchase17/langchain
synced 2024-11-18 09:25:54 +00:00
Cypher search: Check if generated Cypher is provided in backticks (#5541)
# Check if generated Cypher code is wrapped in backticks Some LLMs like the VertexAI like to explain how they generated the Cypher statement and wrap the actual code in three backticks: ![Screenshot from 2023-06-01 08-08-23](https://github.com/hwchase17/langchain/assets/19948365/1d8eecb3-d26c-4882-8f5b-6a9bc7e93690) I have observed a similar pattern with OpenAI chat models in a conversational settings, where multiple user and assistant message are provided to the LLM to generate Cypher statements, where then the LLM wants to maybe apologize for previous steps or explain its thoughts. Interestingly, both OpenAI and VertexAI wrap the code in three backticks if they are doing any explaining or apologizing. Checking if the generated cypher is wrapped in backticks seems like a low-hanging fruit to expand the cypher search to other LLMs and conversational settings.
This commit is contained in:
parent
1a9ac3b1f9
commit
a0ea6f6b6b
@ -1,6 +1,7 @@
|
||||
"""Question answering over a graph."""
|
||||
from __future__ import annotations
|
||||
|
||||
import re
|
||||
from typing import Any, Dict, List, Optional
|
||||
|
||||
from pydantic import Field
|
||||
@ -14,6 +15,16 @@ from langchain.graphs.neo4j_graph import Neo4jGraph
|
||||
from langchain.prompts.base import BasePromptTemplate
|
||||
|
||||
|
||||
def extract_cypher(text: str) -> str:
|
||||
# The pattern to find Cypher code enclosed in triple backticks
|
||||
pattern = r"```(.*?)```"
|
||||
|
||||
# Find all matches in the input text
|
||||
matches = re.findall(pattern, text, re.DOTALL)
|
||||
|
||||
return matches[0] if matches else text
|
||||
|
||||
|
||||
class GraphCypherQAChain(Chain):
|
||||
"""Chain for question-answering against a graph by generating Cypher statements."""
|
||||
|
||||
@ -73,6 +84,9 @@ class GraphCypherQAChain(Chain):
|
||||
{"question": question, "schema": self.graph.get_schema}, callbacks=callbacks
|
||||
)
|
||||
|
||||
# Extract Cypher code if it is wrapped in backticks
|
||||
generated_cypher = extract_cypher(generated_cypher)
|
||||
|
||||
_run_manager.on_text("Generated Cypher:", end="\n", verbose=self.verbose)
|
||||
_run_manager.on_text(
|
||||
generated_cypher, color="green", end="\n", verbose=self.verbose
|
||||
|
@ -50,9 +50,10 @@ CYPHER_GENERATION_PROMPT = PromptTemplate(
|
||||
)
|
||||
|
||||
CYPHER_QA_TEMPLATE = """You are an assistant that helps to form nice and human understandable answers.
|
||||
The information part contains the provided information that you can use to construct an answer.
|
||||
The information part contains the provided information that you must use to construct an answer.
|
||||
The provided information is authorative, you must never doubt it or try to use your internal knowledge to correct it.
|
||||
Make it sound like the information are coming from an AI assistant, but don't add any information.
|
||||
Make the answer sound as a response to the question. Do not mention that you based the result on the given information.
|
||||
If the provided information is empty, say that you don't know the answer.
|
||||
Information:
|
||||
{context}
|
||||
|
||||
|
15
tests/unit_tests/chains/test_graph_qa.py
Normal file
15
tests/unit_tests/chains/test_graph_qa.py
Normal file
@ -0,0 +1,15 @@
|
||||
from langchain.chains.graph_qa.cypher import extract_cypher
|
||||
|
||||
|
||||
def test_no_backticks() -> None:
|
||||
"""Test if there are no backticks, so the original text should be returned."""
|
||||
query = "MATCH (n) RETURN n"
|
||||
output = extract_cypher(query)
|
||||
assert output == query
|
||||
|
||||
|
||||
def test_backticks() -> None:
|
||||
"""Test if there are backticks. Query from within backticks should be returned."""
|
||||
query = "You can use the following query: ```MATCH (n) RETURN n```"
|
||||
output = extract_cypher(query)
|
||||
assert output == "MATCH (n) RETURN n"
|
Loading…
Reference in New Issue
Block a user