forked from Archives/langchain
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1.5 KiB
1.5 KiB
Momento
This page covers how to use the Momento ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Momento wrappers.
Installation and Setup
- Sign up for a free account here and get an auth token
- Install the Momento Python SDK with
pip install momento
Wrappers
Cache
The Cache wrapper allows for Momento to be used as a serverless, distributed, low-latency cache for LLM prompts and responses.
Standard Cache
The standard cache is the go-to use case for Momento users in any environment.
Import the cache as follows:
from langchain.cache import MomentoCache
And set up like so:
from datetime import timedelta
from momento import CacheClient, Configurations, CredentialProvider
import langchain
# Instantiate the Momento client
cache_client = CacheClient(
Configurations.Laptop.v1(),
CredentialProvider.from_environment_variable("MOMENTO_AUTH_TOKEN"),
default_ttl=timedelta(days=1))
# Choose a Momento cache name of your choice
cache_name = "langchain"
# Instantiate the LLM cache
langchain.llm_cache = MomentoCache(cache_client, cache_name)
Memory
Momento can be used as a distributed memory store for LLMs.
Chat Message History Memory
See this notebook for a walkthrough of how to use Momento as a memory store for chat message history.