mirror of
https://github.com/hwchase17/langchain
synced 2024-10-31 15:20:26 +00:00
87e502c6bc
Co-authored-by: jacoblee93 <jacoblee93@gmail.com> Co-authored-by: Harrison Chase <hw.chase.17@gmail.com>
82 lines
2.5 KiB
Plaintext
82 lines
2.5 KiB
Plaintext
# OpenAI
|
|
|
|
>[OpenAI](https://en.wikipedia.org/wiki/OpenAI) is American artificial intelligence (AI) research laboratory
|
|
> consisting of the non-profit `OpenAI Incorporated`
|
|
> and its for-profit subsidiary corporation `OpenAI Limited Partnership`.
|
|
> `OpenAI` conducts AI research with the declared intention of promoting and developing a friendly AI.
|
|
> `OpenAI` systems run on an `Azure`-based supercomputing platform from `Microsoft`.
|
|
|
|
>The [OpenAI API](https://platform.openai.com/docs/models) is powered by a diverse set of models with different capabilities and price points.
|
|
>
|
|
>[ChatGPT](https://chat.openai.com) is the Artificial Intelligence (AI) chatbot developed by `OpenAI`.
|
|
|
|
## Installation and Setup
|
|
- Install the Python SDK with
|
|
```bash
|
|
pip install openai
|
|
```
|
|
- Get an OpenAI api key and set it as an environment variable (`OPENAI_API_KEY`)
|
|
- If you want to use OpenAI's tokenizer (only available for Python 3.9+), install it
|
|
```bash
|
|
pip install tiktoken
|
|
```
|
|
|
|
|
|
## LLM
|
|
|
|
```python
|
|
from langchain.llms import OpenAI
|
|
```
|
|
|
|
If you are using a model hosted on `Azure`, you should use different wrapper for that:
|
|
```python
|
|
from langchain.llms import AzureOpenAI
|
|
```
|
|
For a more detailed walkthrough of the `Azure` wrapper, see [this notebook](/docs/modules/model_io/models/llms/integrations/azure_openai_example.html)
|
|
|
|
|
|
|
|
## Text Embedding Model
|
|
|
|
```python
|
|
from langchain.embeddings import OpenAIEmbeddings
|
|
```
|
|
For a more detailed walkthrough of this, see [this notebook](/docs/modules/data_connection/text_embedding/integrations/openai.html)
|
|
|
|
|
|
## Tokenizer
|
|
|
|
There are several places you can use the `tiktoken` tokenizer. By default, it is used to count tokens
|
|
for OpenAI LLMs.
|
|
|
|
You can also use it to count tokens when splitting documents with
|
|
```python
|
|
from langchain.text_splitter import CharacterTextSplitter
|
|
CharacterTextSplitter.from_tiktoken_encoder(...)
|
|
```
|
|
For a more detailed walkthrough of this, see [this notebook](/docs/modules/data_connection/document_transformers/text_splitters/tiktoken.html)
|
|
|
|
## Chain
|
|
|
|
See a [usage example](/docs/modules/chains/additional/moderation.html).
|
|
|
|
```python
|
|
from langchain.chains import OpenAIModerationChain
|
|
```
|
|
|
|
## Document Loader
|
|
|
|
See a [usage example](/docs/modules/data_connection/document_loaders/integrations/chatgpt_loader.html).
|
|
|
|
```python
|
|
from langchain.document_loaders.chatgpt import ChatGPTLoader
|
|
```
|
|
|
|
## Retriever
|
|
|
|
See a [usage example](/docs/modules/data_connection/retrievers/integrations/chatgpt-plugin.html).
|
|
|
|
```python
|
|
from langchain.retrievers import ChatGPTPluginRetriever
|
|
```
|