|
|
|
@ -1710,14 +1710,12 @@
|
|
|
|
|
"metadata": {},
|
|
|
|
|
"outputs": [],
|
|
|
|
|
"source": [
|
|
|
|
|
"from elasticsearch import Elasticsearch\n",
|
|
|
|
|
"from langchain.globals import set_llm_cache\n",
|
|
|
|
|
"from langchain_elasticsearch import ElasticsearchCache\n",
|
|
|
|
|
"\n",
|
|
|
|
|
"es_client = Elasticsearch(hosts=\"http://localhost:9200\")\n",
|
|
|
|
|
"set_llm_cache(\n",
|
|
|
|
|
" ElasticsearchCache(\n",
|
|
|
|
|
" es_connection=es_client,\n",
|
|
|
|
|
" es_url=\"http://localhost:9200\",\n",
|
|
|
|
|
" index_name=\"llm-chat-cache\",\n",
|
|
|
|
|
" metadata={\"project\": \"my_chatgpt_project\"},\n",
|
|
|
|
|
" )\n",
|
|
|
|
@ -1761,7 +1759,6 @@
|
|
|
|
|
"import json\n",
|
|
|
|
|
"from typing import Any, Dict, List\n",
|
|
|
|
|
"\n",
|
|
|
|
|
"from elasticsearch import Elasticsearch\n",
|
|
|
|
|
"from langchain.globals import set_llm_cache\n",
|
|
|
|
|
"from langchain_core.caches import RETURN_VAL_TYPE\n",
|
|
|
|
|
"from langchain_elasticsearch import ElasticsearchCache\n",
|
|
|
|
@ -1792,9 +1789,10 @@
|
|
|
|
|
" ]\n",
|
|
|
|
|
"\n",
|
|
|
|
|
"\n",
|
|
|
|
|
"es_client = Elasticsearch(hosts=\"http://localhost:9200\")\n",
|
|
|
|
|
"set_llm_cache(\n",
|
|
|
|
|
" SearchableElasticsearchCache(es_connection=es_client, index_name=\"llm-chat-cache\")\n",
|
|
|
|
|
" SearchableElasticsearchCache(\n",
|
|
|
|
|
" es_url=\"http://localhost:9200\", index_name=\"llm-chat-cache\"\n",
|
|
|
|
|
" )\n",
|
|
|
|
|
")"
|
|
|
|
|
]
|
|
|
|
|
},
|
|
|
|
|