Merge pull request #395 from openai/ted/minor-wording-updates-to-qa-notebook

polishes wording in a few places
This commit is contained in:
Ted Sanders 2023-05-05 13:07:20 -07:00 committed by GitHub
commit 5da484282e

View File

@ -83,7 +83,7 @@
"\n",
"Specifically, this notebook demonstrates the following procedure:\n",
"\n",
"1. Prepare search data (once)\n",
"1. Prepare search data (once per document)\n",
" 1. Collect: We'll download a few hundred Wikipedia articles about the 2022 Olympics\n",
" 2. Chunk: Documents are split into short, mostly self-contained sections to be embedded\n",
" 3. Embed: Each section is embedded with the OpenAI API\n",
@ -97,7 +97,7 @@
"\n",
"### Costs\n",
"\n",
"Because GPT is more expensive than embeddings search, a system with a high volume of queries will have its costs dominated by step 3.\n",
"Because GPT is more expensive than embeddings search, a system with a decent volume of queries will have its costs dominated by step 3.\n",
"\n",
"- For `gpt-3.5-turbo` using ~1,000 tokens per query, it costs ~$0.002 per query, or ~500 queries per dollar (as of Apr 2023)\n",
"- For `gpt-4`, again assuming ~1,000 tokens per query, it costs ~$0.03 per query, or ~30 queries per dollar (as of Apr 2023)\n",
@ -574,7 +574,7 @@
"source": [
"Thanks to the Wikipedia article included in the input message, GPT answers correctly.\n",
"\n",
"In this particular case, GPT was intelligent enough to realize that the original question was underspecified, as there were three curling gold medals, not just one.\n",
"In this particular case, GPT was intelligent enough to realize that the original question was underspecified, as there were three curling gold medal events, not just one.\n",
"\n",
"Of course, this example partly relied on human intelligence. We knew the question was about curling, so we inserted a Wikipedia article on curling.\n",
"\n",
@ -591,7 +591,7 @@
"\n",
"To save you the time & expense, we've prepared a pre-embedded dataset of a few hundred Wikipedia articles about the 2022 Winter Olympics.\n",
"\n",
"To see how we constructed this dataset, or to modify it, see [Embedding Wikipedia articles for search](Embedding_Wikipedia_articles_for_search.ipynb)."
"To see how we constructed this dataset, or to modify it yourself, see [Embedding Wikipedia articles for search](Embedding_Wikipedia_articles_for_search.ipynb)."
]
},
{
@ -1011,7 +1011,7 @@
"source": [
"Despite `gpt-3.5-turbo` having no knowledge of the 2022 Winter Olympics, our search system was able to retrieve reference text for the model to read, allowing it to correctly list the gold medal winners in the Men's and Women's tournaments.\n",
"\n",
"However, it still wasn't quite perfect - the model failed to list the gold medal winners from the Mixed doubles event."
"However, it still wasn't quite perfectthe model failed to list the gold medal winners from the Mixed doubles event."
]
},
{