mirror of
https://github.com/hwchase17/langchain
synced 2024-11-06 03:20:49 +00:00
fa5d49f2c1
ran ```bash g grep -l "langchain.vectorstores" | xargs -L 1 sed -i '' "s/langchain\.vectorstores/langchain_community.vectorstores/g" g grep -l "langchain.document_loaders" | xargs -L 1 sed -i '' "s/langchain\.document_loaders/langchain_community.document_loaders/g" g grep -l "langchain.chat_loaders" | xargs -L 1 sed -i '' "s/langchain\.chat_loaders/langchain_community.chat_loaders/g" g grep -l "langchain.document_transformers" | xargs -L 1 sed -i '' "s/langchain\.document_transformers/langchain_community.document_transformers/g" g grep -l "langchain\.graphs" | xargs -L 1 sed -i '' "s/langchain\.graphs/langchain_community.graphs/g" g grep -l "langchain\.memory\.chat_message_histories" | xargs -L 1 sed -i '' "s/langchain\.memory\.chat_message_histories/langchain_community.chat_message_histories/g" gco master libs/langchain/tests/unit_tests/*/test_imports.py gco master libs/langchain/tests/unit_tests/**/test_public_api.py ```
132 lines
4.0 KiB
Python
132 lines
4.0 KiB
Python
import base64
|
|
import io
|
|
from pathlib import Path
|
|
|
|
from langchain.pydantic_v1 import BaseModel
|
|
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
|
from langchain.storage import LocalFileStore
|
|
from langchain_community.chat_models import ChatOllama
|
|
from langchain_community.embeddings import OllamaEmbeddings
|
|
from langchain_community.vectorstores import Chroma
|
|
from langchain_core.documents import Document
|
|
from langchain_core.messages import HumanMessage
|
|
from langchain_core.output_parsers import StrOutputParser
|
|
from langchain_core.runnables import RunnableLambda, RunnablePassthrough
|
|
from PIL import Image
|
|
|
|
|
|
def resize_base64_image(base64_string, size=(128, 128)):
|
|
"""
|
|
Resize an image encoded as a Base64 string.
|
|
|
|
:param base64_string: A Base64 encoded string of the image to be resized.
|
|
:param size: A tuple representing the new size (width, height) for the image.
|
|
:return: A Base64 encoded string of the resized image.
|
|
"""
|
|
img_data = base64.b64decode(base64_string)
|
|
img = Image.open(io.BytesIO(img_data))
|
|
resized_img = img.resize(size, Image.LANCZOS)
|
|
buffered = io.BytesIO()
|
|
resized_img.save(buffered, format=img.format)
|
|
return base64.b64encode(buffered.getvalue()).decode("utf-8")
|
|
|
|
|
|
def get_resized_images(docs):
|
|
"""
|
|
Resize images from base64-encoded strings.
|
|
|
|
:param docs: A list of base64-encoded image to be resized.
|
|
:return: Dict containing a list of resized base64-encoded strings.
|
|
"""
|
|
b64_images = []
|
|
for doc in docs:
|
|
if isinstance(doc, Document):
|
|
doc = doc.page_content
|
|
# Optional: re-size image
|
|
# resized_image = resize_base64_image(doc, size=(1280, 720))
|
|
b64_images.append(doc)
|
|
return {"images": b64_images}
|
|
|
|
|
|
def img_prompt_func(data_dict, num_images=1):
|
|
"""
|
|
Ollama prompt for image analysis.
|
|
|
|
:param data_dict: A dict with images and a user-provided question.
|
|
:param num_images: Number of images to include in the prompt.
|
|
:return: A list containing message objects for each image and the text prompt.
|
|
"""
|
|
messages = []
|
|
if data_dict["context"]["images"]:
|
|
for image in data_dict["context"]["images"][:num_images]:
|
|
image_message = {
|
|
"type": "image_url",
|
|
"image_url": f"data:image/jpeg;base64,{image}",
|
|
}
|
|
messages.append(image_message)
|
|
text_message = {
|
|
"type": "text",
|
|
"text": (
|
|
"You are a helpful assistant that gives a description of food pictures.\n"
|
|
"Give a detailed summary of the image.\n"
|
|
),
|
|
}
|
|
messages.append(text_message)
|
|
return [HumanMessage(content=messages)]
|
|
|
|
|
|
def multi_modal_rag_chain(retriever):
|
|
"""
|
|
Multi-modal RAG chain,
|
|
|
|
:param retriever: A function that retrieves the necessary context for the model.
|
|
:return: A chain of functions representing the multi-modal RAG process.
|
|
"""
|
|
# Initialize the multi-modal Large Language Model with specific parameters
|
|
model = ChatOllama(model="bakllava", temperature=0)
|
|
|
|
# Define the RAG pipeline
|
|
chain = (
|
|
{
|
|
"context": retriever | RunnableLambda(get_resized_images),
|
|
"question": RunnablePassthrough(),
|
|
}
|
|
| RunnableLambda(img_prompt_func)
|
|
| model
|
|
| StrOutputParser()
|
|
)
|
|
|
|
return chain
|
|
|
|
|
|
# Load chroma
|
|
vectorstore_mvr = Chroma(
|
|
collection_name="image_summaries",
|
|
persist_directory=str(Path(__file__).parent.parent / "chroma_db_multi_modal"),
|
|
embedding_function=OllamaEmbeddings(model="llama2:7b"),
|
|
)
|
|
|
|
# Load file store
|
|
store = LocalFileStore(
|
|
str(Path(__file__).parent.parent / "multi_vector_retriever_metadata")
|
|
)
|
|
id_key = "doc_id"
|
|
|
|
# Create the multi-vector retriever
|
|
retriever = MultiVectorRetriever(
|
|
vectorstore=vectorstore_mvr,
|
|
byte_store=store,
|
|
id_key=id_key,
|
|
)
|
|
|
|
# Create RAG chain
|
|
chain = multi_modal_rag_chain(retriever)
|
|
|
|
|
|
# Add typing for input
|
|
class Question(BaseModel):
|
|
__root__: str
|
|
|
|
|
|
chain = chain.with_types(input_type=Question)
|