2023-09-26 09:03:22 +00:00
|
|
|
import os
|
2024-03-28 13:43:10 +00:00
|
|
|
import uuid
|
2024-04-09 15:56:07 +00:00
|
|
|
import shutil
|
2023-09-26 09:03:22 +00:00
|
|
|
from flask import Blueprint, request, jsonify
|
2024-04-08 11:56:27 +00:00
|
|
|
from urllib.parse import urlparse
|
2023-09-26 09:03:22 +00:00
|
|
|
import requests
|
|
|
|
from pymongo import MongoClient
|
|
|
|
from bson.objectid import ObjectId
|
|
|
|
from werkzeug.utils import secure_filename
|
2023-09-27 15:25:57 +00:00
|
|
|
|
2024-02-13 20:41:36 +00:00
|
|
|
from application.api.user.tasks import ingest, ingest_remote
|
2023-09-26 09:03:22 +00:00
|
|
|
|
|
|
|
from application.core.settings import settings
|
2023-09-29 16:17:48 +00:00
|
|
|
from application.vectorstore.vector_creator import VectorCreator
|
|
|
|
|
2023-09-26 09:03:22 +00:00
|
|
|
mongo = MongoClient(settings.MONGO_URI)
|
|
|
|
db = mongo["docsgpt"]
|
|
|
|
conversations_collection = db["conversations"]
|
|
|
|
vectors_collection = db["vectors"]
|
2023-11-14 01:16:06 +00:00
|
|
|
prompts_collection = db["prompts"]
|
2023-11-22 12:16:37 +00:00
|
|
|
feedback_collection = db["feedback"]
|
2024-03-28 13:43:10 +00:00
|
|
|
api_key_collection = db["api_keys"]
|
2023-09-26 09:03:22 +00:00
|
|
|
user = Blueprint('user', __name__)
|
|
|
|
|
2023-09-27 15:25:57 +00:00
|
|
|
current_dir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
|
|
|
|
2023-09-26 09:03:22 +00:00
|
|
|
@user.route("/api/delete_conversation", methods=["POST"])
|
|
|
|
def delete_conversation():
|
|
|
|
# deletes a conversation from the database
|
|
|
|
conversation_id = request.args.get("id")
|
|
|
|
# write to mongodb
|
|
|
|
conversations_collection.delete_one(
|
|
|
|
{
|
|
|
|
"_id": ObjectId(conversation_id),
|
|
|
|
}
|
|
|
|
)
|
|
|
|
|
|
|
|
return {"status": "ok"}
|
|
|
|
|
2024-04-17 22:51:29 +00:00
|
|
|
@user.route("/api/delete_all_conversations", methods=["POST"])
|
|
|
|
def delete_all_conversations():
|
|
|
|
conversations_collection.delete_many({})
|
|
|
|
return {"status": "ok"}
|
|
|
|
|
2023-09-26 09:03:22 +00:00
|
|
|
@user.route("/api/get_conversations", methods=["get"])
|
|
|
|
def get_conversations():
|
|
|
|
# provides a list of conversations
|
2024-03-01 15:42:22 +00:00
|
|
|
conversations = conversations_collection.find().sort("date", -1).limit(30)
|
2023-09-26 09:03:22 +00:00
|
|
|
list_conversations = []
|
|
|
|
for conversation in conversations:
|
|
|
|
list_conversations.append({"id": str(conversation["_id"]), "name": conversation["name"]})
|
|
|
|
|
|
|
|
#list_conversations = [{"id": "default", "name": "default"}, {"id": "jeff", "name": "jeff"}]
|
|
|
|
|
|
|
|
return jsonify(list_conversations)
|
|
|
|
|
|
|
|
|
|
|
|
@user.route("/api/get_single_conversation", methods=["get"])
|
|
|
|
def get_single_conversation():
|
|
|
|
# provides data for a conversation
|
|
|
|
conversation_id = request.args.get("id")
|
|
|
|
conversation = conversations_collection.find_one({"_id": ObjectId(conversation_id)})
|
|
|
|
return jsonify(conversation['queries'])
|
|
|
|
|
2023-10-09 04:55:38 +00:00
|
|
|
@user.route("/api/update_conversation_name", methods=["POST"])
|
|
|
|
def update_conversation_name():
|
|
|
|
# update data for a conversation
|
|
|
|
data = request.get_json()
|
|
|
|
id = data["id"]
|
|
|
|
name = data["name"]
|
2023-10-09 10:46:20 +00:00
|
|
|
conversations_collection.update_one({"_id": ObjectId(id)},{"$set":{"name":name}})
|
|
|
|
return {"status": "ok"}
|
2023-10-09 04:55:38 +00:00
|
|
|
|
2023-09-26 09:03:22 +00:00
|
|
|
|
|
|
|
@user.route("/api/feedback", methods=["POST"])
|
|
|
|
def api_feedback():
|
|
|
|
data = request.get_json()
|
|
|
|
question = data["question"]
|
|
|
|
answer = data["answer"]
|
|
|
|
feedback = data["feedback"]
|
|
|
|
|
2023-11-22 12:16:37 +00:00
|
|
|
|
|
|
|
feedback_collection.insert_one(
|
|
|
|
{
|
|
|
|
"question": question,
|
|
|
|
"answer": answer,
|
|
|
|
"feedback": feedback,
|
|
|
|
}
|
2023-09-26 09:03:22 +00:00
|
|
|
)
|
2023-11-22 12:16:37 +00:00
|
|
|
return {"status": "ok"}
|
2023-09-26 09:03:22 +00:00
|
|
|
|
2023-10-15 21:16:48 +00:00
|
|
|
@user.route("/api/delete_by_ids", methods=["get"])
|
|
|
|
def delete_by_ids():
|
|
|
|
"""Delete by ID. These are the IDs in the vectorstore"""
|
|
|
|
|
|
|
|
ids = request.args.get("path")
|
|
|
|
if not ids:
|
|
|
|
return {"status": "error"}
|
|
|
|
|
|
|
|
if settings.VECTOR_STORE == "faiss":
|
|
|
|
result = vectors_collection.delete_index(ids=ids)
|
|
|
|
if result:
|
|
|
|
return {"status": "ok"}
|
|
|
|
return {"status": "error"}
|
2023-09-26 09:03:22 +00:00
|
|
|
|
|
|
|
@user.route("/api/delete_old", methods=["get"])
|
|
|
|
def delete_old():
|
|
|
|
"""Delete old indexes."""
|
|
|
|
import shutil
|
|
|
|
|
|
|
|
path = request.args.get("path")
|
|
|
|
dirs = path.split("/")
|
|
|
|
dirs_clean = []
|
2023-10-04 12:57:40 +00:00
|
|
|
for i in range(0, len(dirs)):
|
2023-09-26 09:03:22 +00:00
|
|
|
dirs_clean.append(secure_filename(dirs[i]))
|
|
|
|
# check that path strats with indexes or vectors
|
2023-10-04 12:57:40 +00:00
|
|
|
|
|
|
|
if dirs_clean[0] not in ["indexes", "vectors"]:
|
2023-09-26 09:03:22 +00:00
|
|
|
return {"status": "error"}
|
2023-10-04 12:57:40 +00:00
|
|
|
path_clean = "/".join(dirs_clean)
|
|
|
|
vectors_collection.delete_one({"name": dirs_clean[-1], 'user': dirs_clean[-2]})
|
2023-09-29 16:17:48 +00:00
|
|
|
if settings.VECTOR_STORE == "faiss":
|
|
|
|
try:
|
|
|
|
shutil.rmtree(os.path.join(current_dir, path_clean))
|
|
|
|
except FileNotFoundError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
vetorstore = VectorCreator.create_vectorstore(
|
|
|
|
settings.VECTOR_STORE, path=os.path.join(current_dir, path_clean)
|
|
|
|
)
|
|
|
|
vetorstore.delete_index()
|
|
|
|
|
2023-09-26 09:03:22 +00:00
|
|
|
return {"status": "ok"}
|
|
|
|
|
|
|
|
@user.route("/api/upload", methods=["POST"])
|
|
|
|
def upload_file():
|
|
|
|
"""Upload a file to get vectorized and indexed."""
|
|
|
|
if "user" not in request.form:
|
|
|
|
return {"status": "no user"}
|
|
|
|
user = secure_filename(request.form["user"])
|
|
|
|
if "name" not in request.form:
|
|
|
|
return {"status": "no name"}
|
|
|
|
job_name = secure_filename(request.form["name"])
|
|
|
|
# check if the post request has the file part
|
2024-04-09 15:56:07 +00:00
|
|
|
files = request.files.getlist("file")
|
|
|
|
|
|
|
|
if not files or all(file.filename == '' for file in files):
|
2023-09-26 09:03:22 +00:00
|
|
|
return {"status": "no file name"}
|
|
|
|
|
2024-04-09 15:56:07 +00:00
|
|
|
# Directory where files will be saved
|
|
|
|
save_dir = os.path.join(current_dir, settings.UPLOAD_FOLDER, user, job_name)
|
|
|
|
os.makedirs(save_dir, exist_ok=True)
|
|
|
|
|
|
|
|
if len(files) > 1:
|
|
|
|
# Multiple files; prepare them for zip
|
|
|
|
temp_dir = os.path.join(save_dir, "temp")
|
|
|
|
os.makedirs(temp_dir, exist_ok=True)
|
|
|
|
|
|
|
|
for file in files:
|
|
|
|
filename = secure_filename(file.filename)
|
|
|
|
file.save(os.path.join(temp_dir, filename))
|
|
|
|
|
|
|
|
# Use shutil.make_archive to zip the temp directory
|
|
|
|
zip_path = shutil.make_archive(base_name=os.path.join(save_dir, job_name), format='zip', root_dir=temp_dir)
|
|
|
|
final_filename = os.path.basename(zip_path)
|
|
|
|
|
|
|
|
# Clean up the temporary directory after zipping
|
|
|
|
shutil.rmtree(temp_dir)
|
2023-09-26 09:03:22 +00:00
|
|
|
else:
|
2024-04-09 15:56:07 +00:00
|
|
|
# Single file
|
|
|
|
file = files[0]
|
|
|
|
final_filename = secure_filename(file.filename)
|
|
|
|
file_path = os.path.join(save_dir, final_filename)
|
|
|
|
file.save(file_path)
|
|
|
|
|
|
|
|
# Call ingest with the single file or zipped file
|
|
|
|
task = ingest.delay(settings.UPLOAD_FOLDER, [".rst", ".md", ".pdf", ".txt", ".docx",
|
|
|
|
".csv", ".epub", ".html", ".mdx"],
|
|
|
|
job_name, final_filename, user)
|
|
|
|
|
|
|
|
return {"status": "ok", "task_id": task.id}
|
2024-02-13 20:41:36 +00:00
|
|
|
|
|
|
|
@user.route("/api/remote", methods=["POST"])
|
|
|
|
def upload_remote():
|
|
|
|
"""Upload a remote source to get vectorized and indexed."""
|
|
|
|
if "user" not in request.form:
|
|
|
|
return {"status": "no user"}
|
|
|
|
user = secure_filename(request.form["user"])
|
|
|
|
if "source" not in request.form:
|
|
|
|
return {"status": "no source"}
|
|
|
|
source = secure_filename(request.form["source"])
|
|
|
|
if "name" not in request.form:
|
|
|
|
return {"status": "no name"}
|
|
|
|
job_name = secure_filename(request.form["name"])
|
|
|
|
# check if the post request has the file part
|
|
|
|
if "data" not in request.form:
|
|
|
|
print("No data")
|
|
|
|
return {"status": "no data"}
|
|
|
|
source_data = request.form["data"]
|
|
|
|
|
|
|
|
if source_data:
|
|
|
|
task = ingest_remote.delay(source_data=source_data, job_name=job_name, user=user, loader=source)
|
|
|
|
# task id
|
|
|
|
task_id = task.id
|
|
|
|
return {"status": "ok", "task_id": task_id}
|
|
|
|
else:
|
|
|
|
return {"status": "error"}
|
2023-09-26 09:03:22 +00:00
|
|
|
|
|
|
|
@user.route("/api/task_status", methods=["GET"])
|
|
|
|
def task_status():
|
|
|
|
"""Get celery job status."""
|
|
|
|
task_id = request.args.get("task_id")
|
2023-10-01 19:05:13 +00:00
|
|
|
from application.celery import celery
|
|
|
|
task = celery.AsyncResult(task_id)
|
2023-09-26 09:03:22 +00:00
|
|
|
task_meta = task.info
|
|
|
|
return {"status": task.status, "result": task_meta}
|
|
|
|
|
|
|
|
|
|
|
|
@user.route("/api/combine", methods=["GET"])
|
|
|
|
def combined_json():
|
|
|
|
user = "local"
|
|
|
|
"""Provide json file with combined available indexes."""
|
|
|
|
# get json from https://d3dg1063dc54p9.cloudfront.net/combined.json
|
|
|
|
|
|
|
|
data = [
|
|
|
|
{
|
|
|
|
"name": "default",
|
|
|
|
"language": "default",
|
|
|
|
"version": "",
|
|
|
|
"description": "default",
|
|
|
|
"fullName": "default",
|
|
|
|
"date": "default",
|
|
|
|
"docLink": "default",
|
|
|
|
"model": settings.EMBEDDINGS_NAME,
|
2023-11-14 01:16:06 +00:00
|
|
|
"location": "remote",
|
2023-09-26 09:03:22 +00:00
|
|
|
}
|
|
|
|
]
|
|
|
|
# structure: name, language, version, description, fullName, date, docLink
|
|
|
|
# append data from vectors_collection
|
|
|
|
for index in vectors_collection.find({"user": user}):
|
|
|
|
data.append(
|
|
|
|
{
|
|
|
|
"name": index["name"],
|
|
|
|
"language": index["language"],
|
|
|
|
"version": "",
|
|
|
|
"description": index["name"],
|
|
|
|
"fullName": index["name"],
|
|
|
|
"date": index["date"],
|
|
|
|
"docLink": index["location"],
|
|
|
|
"model": settings.EMBEDDINGS_NAME,
|
|
|
|
"location": "local",
|
|
|
|
}
|
|
|
|
)
|
2023-09-29 16:17:48 +00:00
|
|
|
if settings.VECTOR_STORE == "faiss":
|
|
|
|
data_remote = requests.get("https://d3dg1063dc54p9.cloudfront.net/combined.json").json()
|
|
|
|
for index in data_remote:
|
|
|
|
index["location"] = "remote"
|
|
|
|
data.append(index)
|
2024-04-09 15:38:42 +00:00
|
|
|
if 'duckduck_search' in settings.RETRIEVERS_ENABLED:
|
|
|
|
data.append(
|
|
|
|
{
|
|
|
|
"name": "DuckDuckGo Search",
|
|
|
|
"language": "en",
|
|
|
|
"version": "",
|
|
|
|
"description": "duckduck_search",
|
|
|
|
"fullName": "DuckDuckGo Search",
|
|
|
|
"date": "duckduck_search",
|
|
|
|
"docLink": "duckduck_search",
|
|
|
|
"model": settings.EMBEDDINGS_NAME,
|
|
|
|
"location": "custom",
|
|
|
|
}
|
|
|
|
)
|
2024-04-09 16:11:09 +00:00
|
|
|
if 'brave_search' in settings.RETRIEVERS_ENABLED:
|
|
|
|
data.append(
|
|
|
|
{
|
|
|
|
"name": "Brave Search",
|
|
|
|
"language": "en",
|
|
|
|
"version": "",
|
|
|
|
"description": "brave_search",
|
|
|
|
"fullName": "Brave Search",
|
|
|
|
"date": "brave_search",
|
|
|
|
"docLink": "brave_search",
|
|
|
|
"model": settings.EMBEDDINGS_NAME,
|
|
|
|
"location": "custom",
|
|
|
|
}
|
|
|
|
)
|
2023-09-26 09:03:22 +00:00
|
|
|
|
|
|
|
return jsonify(data)
|
|
|
|
|
|
|
|
|
|
|
|
@user.route("/api/docs_check", methods=["POST"])
|
|
|
|
def check_docs():
|
|
|
|
# check if docs exist in a vectorstore folder
|
|
|
|
data = request.get_json()
|
|
|
|
# split docs on / and take first part
|
|
|
|
if data["docs"].split("/")[0] == "local":
|
|
|
|
return {"status": "exists"}
|
2024-04-08 12:06:05 +00:00
|
|
|
vectorstore = "vectors/" + secure_filename(data["docs"])
|
2023-09-26 09:03:22 +00:00
|
|
|
base_path = "https://raw.githubusercontent.com/arc53/DocsHUB/main/"
|
|
|
|
if os.path.exists(vectorstore) or data["docs"] == "default":
|
|
|
|
return {"status": "exists"}
|
|
|
|
else:
|
2024-04-08 11:56:27 +00:00
|
|
|
file_url = urlparse(base_path + vectorstore + "index.faiss")
|
|
|
|
|
2024-04-09 16:25:08 +00:00
|
|
|
if (
|
|
|
|
file_url.scheme in ['https'] and
|
|
|
|
file_url.netloc == 'raw.githubusercontent.com' and
|
|
|
|
file_url.path.startswith('/arc53/DocsHUB/main/')
|
|
|
|
):
|
2024-04-08 11:56:27 +00:00
|
|
|
r = requests.get(file_url.geturl())
|
|
|
|
if r.status_code != 200:
|
|
|
|
return {"status": "null"}
|
|
|
|
else:
|
|
|
|
if not os.path.exists(vectorstore):
|
|
|
|
os.makedirs(vectorstore)
|
|
|
|
with open(vectorstore + "index.faiss", "wb") as f:
|
|
|
|
f.write(r.content)
|
|
|
|
|
|
|
|
r = requests.get(base_path + vectorstore + "index.pkl")
|
|
|
|
with open(vectorstore + "index.pkl", "wb") as f:
|
|
|
|
f.write(r.content)
|
2023-09-26 09:03:22 +00:00
|
|
|
else:
|
2024-04-08 11:56:27 +00:00
|
|
|
return {"status": "null"}
|
2023-09-26 09:03:22 +00:00
|
|
|
|
|
|
|
return {"status": "loaded"}
|
|
|
|
|
2023-11-14 01:16:06 +00:00
|
|
|
@user.route("/api/create_prompt", methods=["POST"])
|
|
|
|
def create_prompt():
|
|
|
|
data = request.get_json()
|
2023-11-22 23:55:41 +00:00
|
|
|
content = data["content"]
|
2023-11-14 01:16:06 +00:00
|
|
|
name = data["name"]
|
2023-11-22 23:55:41 +00:00
|
|
|
if name == "":
|
|
|
|
return {"status": "error"}
|
2023-11-14 01:16:06 +00:00
|
|
|
user = "local"
|
2023-11-22 23:55:41 +00:00
|
|
|
resp = prompts_collection.insert_one(
|
2023-11-14 01:16:06 +00:00
|
|
|
{
|
|
|
|
"name": name,
|
2023-11-22 23:55:41 +00:00
|
|
|
"content": content,
|
2023-11-14 01:16:06 +00:00
|
|
|
"user": user,
|
|
|
|
}
|
|
|
|
)
|
2023-11-22 23:55:41 +00:00
|
|
|
new_id = str(resp.inserted_id)
|
2023-11-23 00:09:17 +00:00
|
|
|
return {"id": new_id}
|
2023-11-14 01:16:06 +00:00
|
|
|
|
|
|
|
@user.route("/api/get_prompts", methods=["GET"])
|
|
|
|
def get_prompts():
|
|
|
|
user = "local"
|
|
|
|
prompts = prompts_collection.find({"user": user})
|
|
|
|
list_prompts = []
|
|
|
|
list_prompts.append({"id": "default", "name": "default", "type": "public"})
|
|
|
|
list_prompts.append({"id": "creative", "name": "creative", "type": "public"})
|
2023-11-24 15:17:48 +00:00
|
|
|
list_prompts.append({"id": "strict", "name": "strict", "type": "public"})
|
2023-11-14 01:16:06 +00:00
|
|
|
for prompt in prompts:
|
2023-11-22 23:55:41 +00:00
|
|
|
list_prompts.append({"id": str(prompt["_id"]), "name": prompt["name"], "type": "private"})
|
2023-11-14 01:16:06 +00:00
|
|
|
|
|
|
|
return jsonify(list_prompts)
|
|
|
|
|
|
|
|
@user.route("/api/get_single_prompt", methods=["GET"])
|
|
|
|
def get_single_prompt():
|
|
|
|
prompt_id = request.args.get("id")
|
2023-11-22 23:55:41 +00:00
|
|
|
if prompt_id == 'default':
|
|
|
|
with open(os.path.join(current_dir, "prompts", "chat_combine_default.txt"), "r") as f:
|
|
|
|
chat_combine_template = f.read()
|
|
|
|
return jsonify({"content": chat_combine_template})
|
|
|
|
elif prompt_id == 'creative':
|
|
|
|
with open(os.path.join(current_dir, "prompts", "chat_combine_creative.txt"), "r") as f:
|
|
|
|
chat_reduce_creative = f.read()
|
|
|
|
return jsonify({"content": chat_reduce_creative})
|
|
|
|
elif prompt_id == 'strict':
|
|
|
|
with open(os.path.join(current_dir, "prompts", "chat_combine_strict.txt"), "r") as f:
|
|
|
|
chat_reduce_strict = f.read()
|
|
|
|
return jsonify({"content": chat_reduce_strict})
|
|
|
|
|
|
|
|
|
2023-11-14 01:16:06 +00:00
|
|
|
prompt = prompts_collection.find_one({"_id": ObjectId(prompt_id)})
|
2023-11-22 23:55:41 +00:00
|
|
|
return jsonify({"content": prompt["content"]})
|
2023-11-14 01:16:06 +00:00
|
|
|
|
|
|
|
@user.route("/api/delete_prompt", methods=["POST"])
|
|
|
|
def delete_prompt():
|
2023-11-22 23:55:41 +00:00
|
|
|
data = request.get_json()
|
|
|
|
id = data["id"]
|
2023-11-14 01:16:06 +00:00
|
|
|
prompts_collection.delete_one(
|
|
|
|
{
|
2023-11-22 23:55:41 +00:00
|
|
|
"_id": ObjectId(id),
|
2023-11-14 01:16:06 +00:00
|
|
|
}
|
|
|
|
)
|
|
|
|
return {"status": "ok"}
|
|
|
|
|
2023-11-22 23:55:41 +00:00
|
|
|
@user.route("/api/update_prompt", methods=["POST"])
|
2023-11-14 01:16:06 +00:00
|
|
|
def update_prompt_name():
|
|
|
|
data = request.get_json()
|
|
|
|
id = data["id"]
|
|
|
|
name = data["name"]
|
2023-11-22 23:55:41 +00:00
|
|
|
content = data["content"]
|
|
|
|
# check if name is null
|
|
|
|
if name == "":
|
|
|
|
return {"status": "error"}
|
|
|
|
prompts_collection.update_one({"_id": ObjectId(id)},{"$set":{"name":name, "content": content}})
|
2023-11-14 01:16:06 +00:00
|
|
|
return {"status": "ok"}
|
|
|
|
|
2023-09-26 09:03:22 +00:00
|
|
|
|
|
|
|
|
2024-03-28 13:43:10 +00:00
|
|
|
@user.route("/api/get_api_keys", methods=["GET"])
|
|
|
|
def get_api_keys():
|
|
|
|
user = "local"
|
|
|
|
keys = api_key_collection.find({"user": user})
|
|
|
|
list_keys = []
|
|
|
|
for key in keys:
|
2024-04-04 11:38:23 +00:00
|
|
|
list_keys.append({
|
|
|
|
"id": str(key["_id"]),
|
|
|
|
"name": key["name"],
|
|
|
|
"key": key["key"][:4] + "..." + key["key"][-4:],
|
|
|
|
"source": key["source"],
|
|
|
|
"prompt_id": key["prompt_id"],
|
|
|
|
"chunks": key["chunks"]
|
|
|
|
})
|
2024-03-28 13:43:10 +00:00
|
|
|
return jsonify(list_keys)
|
|
|
|
|
|
|
|
@user.route("/api/create_api_key", methods=["POST"])
|
|
|
|
def create_api_key():
|
|
|
|
data = request.get_json()
|
|
|
|
name = data["name"]
|
|
|
|
source = data["source"]
|
2024-04-04 11:38:23 +00:00
|
|
|
prompt_id = data["prompt_id"]
|
|
|
|
chunks = data["chunks"]
|
2024-03-28 13:43:10 +00:00
|
|
|
key = str(uuid.uuid4())
|
|
|
|
user = "local"
|
|
|
|
resp = api_key_collection.insert_one(
|
|
|
|
{
|
|
|
|
"name": name,
|
|
|
|
"key": key,
|
|
|
|
"source": source,
|
|
|
|
"user": user,
|
2024-04-04 11:38:23 +00:00
|
|
|
"prompt_id": prompt_id,
|
|
|
|
"chunks": chunks
|
2024-03-28 13:43:10 +00:00
|
|
|
}
|
|
|
|
)
|
|
|
|
new_id = str(resp.inserted_id)
|
|
|
|
return {"id": new_id, "key": key}
|
2023-09-26 09:03:22 +00:00
|
|
|
|
2024-03-28 13:43:10 +00:00
|
|
|
@user.route("/api/delete_api_key", methods=["POST"])
|
|
|
|
def delete_api_key():
|
|
|
|
data = request.get_json()
|
|
|
|
id = data["id"]
|
|
|
|
api_key_collection.delete_one(
|
|
|
|
{
|
|
|
|
"_id": ObjectId(id),
|
|
|
|
}
|
|
|
|
)
|
|
|
|
return {"status": "ok"}
|
2023-09-26 09:03:22 +00:00
|
|
|
|