mirror of
https://github.com/hwchase17/langchain
synced 2024-11-06 03:20:49 +00:00
ed58eeb9c5
Moved the following modules to new package langchain-community in a backwards compatible fashion: ``` mv langchain/langchain/adapters community/langchain_community mv langchain/langchain/callbacks community/langchain_community/callbacks mv langchain/langchain/chat_loaders community/langchain_community mv langchain/langchain/chat_models community/langchain_community mv langchain/langchain/document_loaders community/langchain_community mv langchain/langchain/docstore community/langchain_community mv langchain/langchain/document_transformers community/langchain_community mv langchain/langchain/embeddings community/langchain_community mv langchain/langchain/graphs community/langchain_community mv langchain/langchain/llms community/langchain_community mv langchain/langchain/memory/chat_message_histories community/langchain_community mv langchain/langchain/retrievers community/langchain_community mv langchain/langchain/storage community/langchain_community mv langchain/langchain/tools community/langchain_community mv langchain/langchain/utilities community/langchain_community mv langchain/langchain/vectorstores community/langchain_community mv langchain/langchain/agents/agent_toolkits community/langchain_community mv langchain/langchain/cache.py community/langchain_community mv langchain/langchain/adapters community/langchain_community mv langchain/langchain/callbacks community/langchain_community/callbacks mv langchain/langchain/chat_loaders community/langchain_community mv langchain/langchain/chat_models community/langchain_community mv langchain/langchain/document_loaders community/langchain_community mv langchain/langchain/docstore community/langchain_community mv langchain/langchain/document_transformers community/langchain_community mv langchain/langchain/embeddings community/langchain_community mv langchain/langchain/graphs community/langchain_community mv langchain/langchain/llms community/langchain_community mv langchain/langchain/memory/chat_message_histories community/langchain_community mv langchain/langchain/retrievers community/langchain_community mv langchain/langchain/storage community/langchain_community mv langchain/langchain/tools community/langchain_community mv langchain/langchain/utilities community/langchain_community mv langchain/langchain/vectorstores community/langchain_community mv langchain/langchain/agents/agent_toolkits community/langchain_community mv langchain/langchain/cache.py community/langchain_community ``` Moved the following to core ``` mv langchain/langchain/utils/json_schema.py core/langchain_core/utils mv langchain/langchain/utils/html.py core/langchain_core/utils mv langchain/langchain/utils/strings.py core/langchain_core/utils cat langchain/langchain/utils/env.py >> core/langchain_core/utils/env.py rm langchain/langchain/utils/env.py ``` See .scripts/community_split/script_integrations.sh for all changes
87 lines
3.1 KiB
Python
87 lines
3.1 KiB
Python
from typing import Any, Dict, List, cast
|
|
|
|
from langchain_core.callbacks import CallbackManagerForRetrieverRun
|
|
from langchain_core.documents import Document
|
|
from langchain_core.pydantic_v1 import Field
|
|
from langchain_core.retrievers import BaseRetriever
|
|
|
|
|
|
class LlamaIndexRetriever(BaseRetriever):
|
|
"""`LlamaIndex` retriever.
|
|
|
|
It is used for the question-answering with sources over
|
|
an LlamaIndex data structure."""
|
|
|
|
index: Any
|
|
"""LlamaIndex index to query."""
|
|
query_kwargs: Dict = Field(default_factory=dict)
|
|
"""Keyword arguments to pass to the query method."""
|
|
|
|
def _get_relevant_documents(
|
|
self, query: str, *, run_manager: CallbackManagerForRetrieverRun
|
|
) -> List[Document]:
|
|
"""Get documents relevant for a query."""
|
|
try:
|
|
from llama_index.indices.base import BaseGPTIndex
|
|
from llama_index.response.schema import Response
|
|
except ImportError:
|
|
raise ImportError(
|
|
"You need to install `pip install llama-index` to use this retriever."
|
|
)
|
|
index = cast(BaseGPTIndex, self.index)
|
|
|
|
response = index.query(query, response_mode="no_text", **self.query_kwargs)
|
|
response = cast(Response, response)
|
|
# parse source nodes
|
|
docs = []
|
|
for source_node in response.source_nodes:
|
|
metadata = source_node.extra_info or {}
|
|
docs.append(
|
|
Document(page_content=source_node.source_text, metadata=metadata)
|
|
)
|
|
return docs
|
|
|
|
|
|
class LlamaIndexGraphRetriever(BaseRetriever):
|
|
"""`LlamaIndex` graph data structure retriever.
|
|
|
|
It is used for question-answering with sources over an LlamaIndex
|
|
graph data structure."""
|
|
|
|
graph: Any
|
|
"""LlamaIndex graph to query."""
|
|
query_configs: List[Dict] = Field(default_factory=list)
|
|
"""List of query configs to pass to the query method."""
|
|
|
|
def _get_relevant_documents(
|
|
self, query: str, *, run_manager: CallbackManagerForRetrieverRun
|
|
) -> List[Document]:
|
|
"""Get documents relevant for a query."""
|
|
try:
|
|
from llama_index.composability.graph import (
|
|
QUERY_CONFIG_TYPE,
|
|
ComposableGraph,
|
|
)
|
|
from llama_index.response.schema import Response
|
|
except ImportError:
|
|
raise ImportError(
|
|
"You need to install `pip install llama-index` to use this retriever."
|
|
)
|
|
graph = cast(ComposableGraph, self.graph)
|
|
|
|
# for now, inject response_mode="no_text" into query configs
|
|
for query_config in self.query_configs:
|
|
query_config["response_mode"] = "no_text"
|
|
query_configs = cast(List[QUERY_CONFIG_TYPE], self.query_configs)
|
|
response = graph.query(query, query_configs=query_configs)
|
|
response = cast(Response, response)
|
|
|
|
# parse source nodes
|
|
docs = []
|
|
for source_node in response.source_nodes:
|
|
metadata = source_node.extra_info or {}
|
|
docs.append(
|
|
Document(page_content=source_node.source_text, metadata=metadata)
|
|
)
|
|
return docs
|