mirror of
https://github.com/dair-ai/Prompt-Engineering-Guide
synced 2024-11-10 01:13:36 +00:00
78 lines
2.5 KiB
Plaintext
78 lines
2.5 KiB
Plaintext
|
# Open Domain Question Answering with LLMs
|
|||
|
|
|||
|
import { Tabs, Tab } from 'nextra/components'
|
|||
|
import {Callout} from 'nextra/components'
|
|||
|
|
|||
|
## Background
|
|||
|
The following prompt tests an LLM's capabilities to answer open-domain questions which involves answering factual questions without any evidence provided.
|
|||
|
|
|||
|
<Callout type="warning" emoji="⚠️">
|
|||
|
Note that due to the challenging nature of the task, LLMs are likely to hallucinate when they have no knowledge regarding the question.
|
|||
|
</Callout>
|
|||
|
|
|||
|
## Prompt
|
|||
|
```markdown
|
|||
|
In this conversation between a human and the AI, the AI is helpful and friendly, and when it does not know the answer it says "I don’t know".
|
|||
|
|
|||
|
AI: Hi, how can I help you?
|
|||
|
Human: Can I get McDonalds at the SeaTac airport?
|
|||
|
```
|
|||
|
|
|||
|
## Code / API
|
|||
|
|
|||
|
<Tabs items={['GPT-4 (OpenAI)', 'Mixtral MoE 8x7B Instruct (Fireworks)']}>
|
|||
|
<Tab>
|
|||
|
|
|||
|
```python
|
|||
|
from openai import OpenAI
|
|||
|
client = OpenAI()
|
|||
|
|
|||
|
response = client.chat.completions.create(
|
|||
|
model="gpt-4",
|
|||
|
messages=[
|
|||
|
{
|
|||
|
"role": "user",
|
|||
|
"content": "In this conversation between a human and the AI, the AI is helpful and friendly, and when it does not know the answer it says \"I don’t know\".\n\nAI: Hi, how can I help you?\nHuman: Can I get McDonalds at the SeaTac airport?"
|
|||
|
}
|
|||
|
],
|
|||
|
temperature=1,
|
|||
|
max_tokens=250,
|
|||
|
top_p=1,
|
|||
|
frequency_penalty=0,
|
|||
|
presence_penalty=0
|
|||
|
)
|
|||
|
```
|
|||
|
</Tab>
|
|||
|
|
|||
|
<Tab>
|
|||
|
```python
|
|||
|
import fireworks.client
|
|||
|
fireworks.client.api_key = "<FIREWORKS_API_KEY>"
|
|||
|
completion = fireworks.client.ChatCompletion.create(
|
|||
|
model="accounts/fireworks/models/mixtral-8x7b-instruct",
|
|||
|
messages=[
|
|||
|
{
|
|||
|
"role": "user",
|
|||
|
"content": "In this conversation between a human and the AI, the AI is helpful and friendly, and when it does not know the answer it says \"I don’t know\".\n\nAI: Hi, how can I help you?\nHuman: Can I get McDonalds at the SeaTac airport?",
|
|||
|
}
|
|||
|
],
|
|||
|
stop=["<|im_start|>","<|im_end|>","<|endoftext|>"],
|
|||
|
stream=True,
|
|||
|
n=1,
|
|||
|
top_p=1,
|
|||
|
top_k=40,
|
|||
|
presence_penalty=0,
|
|||
|
frequency_penalty=0,
|
|||
|
prompt_truncate_len=1024,
|
|||
|
context_length_exceeded_behavior="truncate",
|
|||
|
temperature=0.9,
|
|||
|
max_tokens=4000
|
|||
|
)
|
|||
|
```
|
|||
|
</Tab>
|
|||
|
|
|||
|
|
|||
|
</Tabs>
|
|||
|
|
|||
|
## Reference
|
|||
|
- [Sparks of Artificial General Intelligence: Early experiments with GPT-4](https://arxiv.org/abs/2303.12712) (13 April 2023)
|