|
|
|
@ -65,12 +65,15 @@ def get_stream_prompt(query, prompt_file, isQuestion, info_file=None):
|
|
|
|
|
openai.api_key = API_KEY
|
|
|
|
|
if isQuestion:
|
|
|
|
|
data = util.read_json(INFO.BRAIN_DATA)
|
|
|
|
|
result = GPT.gpt_tools.search_chunks(query, data, count=1)
|
|
|
|
|
my_info = util.read_file(info_file)
|
|
|
|
|
prompt = util.read_file(prompt_file)
|
|
|
|
|
prompt = prompt.replace('<<INFO>>', result[0]['content'])
|
|
|
|
|
prompt = prompt.replace('<<QS>>', query)
|
|
|
|
|
prompt = prompt.replace('<<MY-INFO>>', my_info)
|
|
|
|
|
if data:
|
|
|
|
|
result = GPT.gpt_tools.search_chunks(query, data, count=1)
|
|
|
|
|
my_info = util.read_file(info_file)
|
|
|
|
|
prompt = util.read_file(prompt_file)
|
|
|
|
|
prompt = prompt.replace('<<INFO>>', result[0]['content'])
|
|
|
|
|
prompt = prompt.replace('<<QS>>', query)
|
|
|
|
|
prompt = prompt.replace('<<MY-INFO>>', my_info)
|
|
|
|
|
else:
|
|
|
|
|
prompt = ''
|
|
|
|
|
else:
|
|
|
|
|
chunk = textwrap.wrap(query, 10000)[0]
|
|
|
|
|
prompt = util.read_file(prompt_file).replace('<<DATA>>', chunk)
|
|
|
|
|