|
|
|
@ -172,6 +172,7 @@
|
|
|
|
|
]
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
"attachments": {},
|
|
|
|
|
"cell_type": "markdown",
|
|
|
|
|
"id": "f2e47962",
|
|
|
|
|
"metadata": {},
|
|
|
|
@ -338,7 +339,7 @@
|
|
|
|
|
"\n",
|
|
|
|
|
"\n",
|
|
|
|
|
"def extract_chunk(content, template_prompt):\n",
|
|
|
|
|
" \"\"\"This function applies a prompt to some input content. In this case it returns a summarize chunk of text\"\"\"\n",
|
|
|
|
|
" \"\"\"This function applies a prompt to some input content. In this case it returns a summarized chunk of text\"\"\"\n",
|
|
|
|
|
" prompt = template_prompt + content\n",
|
|
|
|
|
" response = openai.ChatCompletion.create(\n",
|
|
|
|
|
" model=GPT_MODEL, messages=[{\"role\": \"user\", \"content\": prompt}], temperature=0\n",
|
|
|
|
@ -480,6 +481,7 @@
|
|
|
|
|
]
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
"attachments": {},
|
|
|
|
|
"cell_type": "markdown",
|
|
|
|
|
"id": "dab07e98",
|
|
|
|
|
"metadata": {},
|
|
|
|
@ -656,6 +658,7 @@
|
|
|
|
|
]
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
"attachments": {},
|
|
|
|
|
"cell_type": "markdown",
|
|
|
|
|
"id": "dd3e7868",
|
|
|
|
|
"metadata": {},
|
|
|
|
|