added new techniques
parent
d27d18ff1b
commit
1e4208e419
Binary file not shown.
After Width: | Height: | Size: 194 KiB |
Binary file not shown.
After Width: | Height: | Size: 211 KiB |
Binary file not shown.
After Width: | Height: | Size: 110 KiB |
Binary file not shown.
After Width: | Height: | Size: 82 KiB |
Binary file not shown.
After Width: | Height: | Size: 66 KiB |
Binary file not shown.
After Width: | Height: | Size: 224 KiB |
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,25 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
import { Callout, FileTree } from 'nextra-theme-docs'
|
||||
import {Screenshot} from 'components/screenshot'
|
||||
import ART from '../../img/ART.png'
|
||||
import ART2 from '../../img/ART2.png'
|
||||
|
||||
Combining CoT prompting and tools in an interleaved manner has shown to be a strong and robust approach to address many tasks with LLMs. These approaches typically require hand-crafting task-specific demonstrations and carefully scripted interleaving of model generations with tool use. [Paranjape et al., (2023)](https://arxiv.org/abs/2303.09014) propose a new framework that uses a frozen LLM to automatically generate intermediate reasoning steps as a program.
|
||||
|
||||
ART works as follows:
|
||||
- given a new task, it select demonstrations of multi-step reasoning and tool use from a task library
|
||||
- at test time, it pauses generation whenever external tools are called, and integrate their output before resuming generation
|
||||
|
||||
ART encourages the model to generalize from demonstrations to decompose a new task and
|
||||
use tools in appropriate places, in a zero-shot fashion. In addition, ART is extensible as it also enables humans to fix mistakes in the reasoning steps or add new tools by simply updating the task and tool libraries. The process is demonstrated below:
|
||||
|
||||
<Screenshot src={ART} alt="ART" />
|
||||
Image Source: [Paranjape et al., (2023)](https://arxiv.org/abs/2303.09014)
|
||||
|
||||
ART substantially improves over few-shot prompting and automatic CoT on unseen tasks in the BigBench and MMLU benchmarks, and exceeds performance of hand-crafted CoT prompts when human feedback is incorporated.
|
||||
|
||||
Below is a table demonstrating ART's performance on BigBench and MMLU tasks:
|
||||
|
||||
<Screenshot src={ART2} alt="ART2" />
|
||||
Image Source: [Paranjape et al., (2023)](https://arxiv.org/abs/2303.09014)
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Automatic Reasoning and Tool-use (ART)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,28 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
import { Callout, FileTree } from 'nextra-theme-docs'
|
||||
import {Screenshot} from 'components/screenshot'
|
||||
import TOT from '../../img/TOT.png'
|
||||
import TOT2 from '../../img/TOT2.png'
|
||||
import TOT3 from '../../img/TOT3.png'
|
||||
|
||||
For complex tasks that require exploration or strategic lookahead, traditional or simple prompting techniques fall short. [Yao et el. (2023)](https://arxiv.org/abs/2305.10601) recently proposed Tree of Thoughts (ToT), a framework that generalizes over chain-of-thought prompting and encourages exploration over thoughts that serve as intermediate steps for general problem solving with language models.
|
||||
|
||||
ToT maintains a tree of thoughts, where thoughts represent coherent language sequences that serve as intermediate steps toward solving a problem. This approach enables an LM to self-evaluate the progress intermediate thoughts make towards solving a problem through a deliberate reasoning process. The LM ability to generate and evaluate thoughts is then combined with search algorithms (e.g., breath-first search and depth-first search) to enable systematic exploration of thoughts with lookahead and backtracking.
|
||||
|
||||
The ToT framework is illustrated below:
|
||||
|
||||
<Screenshot src={TOT} alt="TOT" />
|
||||
Image Source: [Yao et el. (2023)](https://arxiv.org/abs/2305.10601)
|
||||
|
||||
When using ToT, different tasks requires defining the number of candidates and the number of thoughts/steps. For instance, as demonstrated in the paper, Game of 24 is used as a mathematical reasoning task which requires decomposing the thoughts into 3 steps, each involving an intermediate equation. At each step, the best b=5 candidates are kept.
|
||||
|
||||
To perform BFS in ToT for the Game of 24 task, the LM is prompted to evaluate each thought candidate as "sure/maybe/impossible" with regard to reaching 24. As stated by the authors, "the aim is to promote correct partial solutions that can be verdicted within few lookahead trials, and eliminate impossible partial solutions based on "too big/small" commonsense, and keep the rest "maybe"". Values are sampled 3 times for each thought. The process is illustrated below:
|
||||
|
||||
<Screenshot src={TOT2} alt="TOT2" />
|
||||
Image Source: [Yao et el. (2023)](https://arxiv.org/abs/2305.10601)
|
||||
|
||||
From the results reported in the figure below, ToT substantially outperforms the other prompting methods:
|
||||
|
||||
<Screenshot src={TOT2} alt="TOT2" />
|
||||
Image Source: [Yao et el. (2023)](https://arxiv.org/abs/2305.10601)
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
@ -0,0 +1,3 @@
|
||||
# Tree of Thoughts (ToT)
|
||||
|
||||
This page needs a translation! Feel free to contribute a translation by clicking the `Edit this page` button on the right side.
|
Loading…
Reference in New Issue