From a0281f5acb9c5f294039b297f419524e0a95b0ca Mon Sep 17 00:00:00 2001 From: Moonsik Kang <2792+deepblue@users.noreply.github.com> Date: Thu, 25 May 2023 19:18:31 -0700 Subject: [PATCH] Fixed typo: 'ouput' to 'output' in all documentation (#5272) # Fixed typo: 'ouput' to 'output' in all documentation In this instance, the typo 'ouput' was amended to 'output' in all occurrences within the documentation. There are no dependencies required for this change. --- docs/modules/memory/types/buffer.ipynb | 4 ++-- docs/modules/memory/types/buffer_window.ipynb | 8 ++++---- docs/modules/memory/types/entity_summary_memory.ipynb | 4 ++-- docs/modules/memory/types/kg.ipynb | 8 ++++---- docs/modules/memory/types/summary.ipynb | 4 ++-- docs/modules/memory/types/token_buffer.ipynb | 8 ++++---- docs/use_cases/agents/multi_modal_output_agent.ipynb | 6 +++--- docs/use_cases/multi_modal/image_agent.ipynb | 6 +++--- 8 files changed, 24 insertions(+), 24 deletions(-) diff --git a/docs/modules/memory/types/buffer.ipynb b/docs/modules/memory/types/buffer.ipynb index e310a627..19a12e1e 100644 --- a/docs/modules/memory/types/buffer.ipynb +++ b/docs/modules/memory/types/buffer.ipynb @@ -30,7 +30,7 @@ "outputs": [], "source": [ "memory = ConversationBufferMemory()\n", - "memory.save_context({\"input\": \"hi\"}, {\"ouput\": \"whats up\"})" + "memory.save_context({\"input\": \"hi\"}, {\"output\": \"whats up\"})" ] }, { @@ -70,7 +70,7 @@ "outputs": [], "source": [ "memory = ConversationBufferMemory(return_messages=True)\n", - "memory.save_context({\"input\": \"hi\"}, {\"ouput\": \"whats up\"})" + "memory.save_context({\"input\": \"hi\"}, {\"output\": \"whats up\"})" ] }, { diff --git a/docs/modules/memory/types/buffer_window.ipynb b/docs/modules/memory/types/buffer_window.ipynb index a653665a..b2d15c0b 100644 --- a/docs/modules/memory/types/buffer_window.ipynb +++ b/docs/modules/memory/types/buffer_window.ipynb @@ -30,8 +30,8 @@ "outputs": [], "source": [ "memory = ConversationBufferWindowMemory( k=1)\n", - "memory.save_context({\"input\": \"hi\"}, {\"ouput\": \"whats up\"})\n", - "memory.save_context({\"input\": \"not much you\"}, {\"ouput\": \"not much\"})" + "memory.save_context({\"input\": \"hi\"}, {\"output\": \"whats up\"})\n", + "memory.save_context({\"input\": \"not much you\"}, {\"output\": \"not much\"})" ] }, { @@ -71,8 +71,8 @@ "outputs": [], "source": [ "memory = ConversationBufferWindowMemory( k=1, return_messages=True)\n", - "memory.save_context({\"input\": \"hi\"}, {\"ouput\": \"whats up\"})\n", - "memory.save_context({\"input\": \"not much you\"}, {\"ouput\": \"not much\"})" + "memory.save_context({\"input\": \"hi\"}, {\"output\": \"whats up\"})\n", + "memory.save_context({\"input\": \"not much you\"}, {\"output\": \"not much\"})" ] }, { diff --git a/docs/modules/memory/types/entity_summary_memory.ipynb b/docs/modules/memory/types/entity_summary_memory.ipynb index 6d63d7ae..b3d024a2 100644 --- a/docs/modules/memory/types/entity_summary_memory.ipynb +++ b/docs/modules/memory/types/entity_summary_memory.ipynb @@ -35,7 +35,7 @@ "memory.load_memory_variables(_input)\n", "memory.save_context(\n", " _input,\n", - " {\"ouput\": \" That sounds like a great project! What kind of project are they working on?\"}\n", + " {\"output\": \" That sounds like a great project! What kind of project are they working on?\"}\n", ")" ] }, @@ -73,7 +73,7 @@ "memory.load_memory_variables(_input)\n", "memory.save_context(\n", " _input,\n", - " {\"ouput\": \" That sounds like a great project! What kind of project are they working on?\"}\n", + " {\"output\": \" That sounds like a great project! What kind of project are they working on?\"}\n", ")" ] }, diff --git a/docs/modules/memory/types/kg.ipynb b/docs/modules/memory/types/kg.ipynb index c6739d45..143e14d9 100644 --- a/docs/modules/memory/types/kg.ipynb +++ b/docs/modules/memory/types/kg.ipynb @@ -32,8 +32,8 @@ "source": [ "llm = OpenAI(temperature=0)\n", "memory = ConversationKGMemory(llm=llm)\n", - "memory.save_context({\"input\": \"say hi to sam\"}, {\"ouput\": \"who is sam\"})\n", - "memory.save_context({\"input\": \"sam is a friend\"}, {\"ouput\": \"okay\"})" + "memory.save_context({\"input\": \"say hi to sam\"}, {\"output\": \"who is sam\"})\n", + "memory.save_context({\"input\": \"sam is a friend\"}, {\"output\": \"okay\"})" ] }, { @@ -73,8 +73,8 @@ "outputs": [], "source": [ "memory = ConversationKGMemory(llm=llm, return_messages=True)\n", - "memory.save_context({\"input\": \"say hi to sam\"}, {\"ouput\": \"who is sam\"})\n", - "memory.save_context({\"input\": \"sam is a friend\"}, {\"ouput\": \"okay\"})" + "memory.save_context({\"input\": \"say hi to sam\"}, {\"output\": \"who is sam\"})\n", + "memory.save_context({\"input\": \"sam is a friend\"}, {\"output\": \"okay\"})" ] }, { diff --git a/docs/modules/memory/types/summary.ipynb b/docs/modules/memory/types/summary.ipynb index b2dcbb9b..77d2a50a 100644 --- a/docs/modules/memory/types/summary.ipynb +++ b/docs/modules/memory/types/summary.ipynb @@ -30,7 +30,7 @@ "outputs": [], "source": [ "memory = ConversationSummaryMemory(llm=OpenAI(temperature=0))\n", - "memory.save_context({\"input\": \"hi\"}, {\"ouput\": \"whats up\"})" + "memory.save_context({\"input\": \"hi\"}, {\"output\": \"whats up\"})" ] }, { @@ -70,7 +70,7 @@ "outputs": [], "source": [ "memory = ConversationSummaryMemory(llm=OpenAI(temperature=0), return_messages=True)\n", - "memory.save_context({\"input\": \"hi\"}, {\"ouput\": \"whats up\"})" + "memory.save_context({\"input\": \"hi\"}, {\"output\": \"whats up\"})" ] }, { diff --git a/docs/modules/memory/types/token_buffer.ipynb b/docs/modules/memory/types/token_buffer.ipynb index 9cf371f1..a5fd6c0d 100644 --- a/docs/modules/memory/types/token_buffer.ipynb +++ b/docs/modules/memory/types/token_buffer.ipynb @@ -32,8 +32,8 @@ "outputs": [], "source": [ "memory = ConversationTokenBufferMemory(llm=llm, max_token_limit=10)\n", - "memory.save_context({\"input\": \"hi\"}, {\"ouput\": \"whats up\"})\n", - "memory.save_context({\"input\": \"not much you\"}, {\"ouput\": \"not much\"})" + "memory.save_context({\"input\": \"hi\"}, {\"output\": \"whats up\"})\n", + "memory.save_context({\"input\": \"not much you\"}, {\"output\": \"not much\"})" ] }, { @@ -73,8 +73,8 @@ "outputs": [], "source": [ "memory = ConversationTokenBufferMemory(llm=llm, max_token_limit=10, return_messages=True)\n", - "memory.save_context({\"input\": \"hi\"}, {\"ouput\": \"whats up\"})\n", - "memory.save_context({\"input\": \"not much you\"}, {\"ouput\": \"not much\"})" + "memory.save_context({\"input\": \"hi\"}, {\"output\": \"whats up\"})\n", + "memory.save_context({\"input\": \"not much you\"}, {\"output\": \"not much\"})" ] }, { diff --git a/docs/use_cases/agents/multi_modal_output_agent.ipynb b/docs/use_cases/agents/multi_modal_output_agent.ipynb index 35102050..6e7da84c 100644 --- a/docs/use_cases/agents/multi_modal_output_agent.ipynb +++ b/docs/use_cases/agents/multi_modal_output_agent.ipynb @@ -125,7 +125,7 @@ "metadata": {}, "outputs": [], "source": [ - "def show_ouput(output):\n", + "def show_output(output):\n", " \"\"\"Display the multi-modal output from the agent.\"\"\"\n", " UUID_PATTERN = re.compile(\n", " r\"([0-9A-Za-z]{8}-[0-9A-Za-z]{4}-[0-9A-Za-z]{4}-[0-9A-Za-z]{4}-[0-9A-Za-z]{12})\"\n", @@ -168,7 +168,7 @@ } ], "source": [ - "show_ouput(output)" + "show_output(output)" ] }, { @@ -266,7 +266,7 @@ } ], "source": [ - "show_ouput(output)" + "show_output(output)" ] }, { diff --git a/docs/use_cases/multi_modal/image_agent.ipynb b/docs/use_cases/multi_modal/image_agent.ipynb index ca856ede..5ed4223d 100644 --- a/docs/use_cases/multi_modal/image_agent.ipynb +++ b/docs/use_cases/multi_modal/image_agent.ipynb @@ -125,7 +125,7 @@ "metadata": {}, "outputs": [], "source": [ - "def show_ouput(output):\n", + "def show_output(output):\n", " \"\"\"Display the multi-modal output from the agent.\"\"\"\n", " UUID_PATTERN = re.compile(\n", " r\"([0-9A-Za-z]{8}-[0-9A-Za-z]{4}-[0-9A-Za-z]{4}-[0-9A-Za-z]{4}-[0-9A-Za-z]{12})\"\n", @@ -168,7 +168,7 @@ } ], "source": [ - "show_ouput(output)" + "show_output(output)" ] }, { @@ -221,7 +221,7 @@ "metadata": {}, "outputs": [], "source": [ - "show_ouput(output)" + "show_output(output)" ] }, {