diff --git a/examples/How_to_call_functions_for_knowledge_retrieval.ipynb b/examples/How_to_call_functions_for_knowledge_retrieval.ipynb index b7554c99..2383f135 100644 --- a/examples/How_to_call_functions_for_knowledge_retrieval.ipynb +++ b/examples/How_to_call_functions_for_knowledge_retrieval.ipynb @@ -172,6 +172,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "f2e47962", "metadata": {}, @@ -338,7 +339,7 @@ "\n", "\n", "def extract_chunk(content, template_prompt):\n", - " \"\"\"This function applies a prompt to some input content. In this case it returns a summarize chunk of text\"\"\"\n", + " \"\"\"This function applies a prompt to some input content. In this case it returns a summarized chunk of text\"\"\"\n", " prompt = template_prompt + content\n", " response = openai.ChatCompletion.create(\n", " model=GPT_MODEL, messages=[{\"role\": \"user\", \"content\": prompt}], temperature=0\n", @@ -480,6 +481,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "dab07e98", "metadata": {}, @@ -656,6 +658,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "dd3e7868", "metadata": {},