mirror of
https://github.com/xtekky/gpt4free.git
synced 2024-11-05 00:01:00 +00:00
b4b74c991b
Add auth header to GeminiPro provider
149 lines
4.6 KiB
Python
149 lines
4.6 KiB
Python
from __future__ import annotations
|
|
|
|
import uuid
|
|
|
|
from aiohttp import ClientSession, BaseConnector
|
|
|
|
from ..typing import AsyncResult, Messages
|
|
from .base_provider import AsyncGeneratorProvider, ProviderModelMixin
|
|
from .helper import get_connector
|
|
|
|
models = {
|
|
"gpt-4": {
|
|
"id": "gpt-4",
|
|
"name": "GPT-4",
|
|
"maxLength": 24000,
|
|
"tokenLimit": 8000,
|
|
},
|
|
"gpt-4-0613": {
|
|
"id": "gpt-4-0613",
|
|
"name": "GPT-4",
|
|
"maxLength": 32000,
|
|
"tokenLimit": 8000,
|
|
},
|
|
"gpt-3.5-turbo": {
|
|
"id": "gpt-3.5-turbo",
|
|
"name": "GPT-3.5-Turbo",
|
|
"maxLength": 48000,
|
|
"tokenLimit": 14000,
|
|
"context": "16K",
|
|
},
|
|
"gpt-3.5-turbo-16k": {
|
|
"id": "gpt-3.5-turbo-16k",
|
|
"name": "GPT-3.5-16k",
|
|
"maxLength": 48000,
|
|
"tokenLimit": 16000,
|
|
},
|
|
"gpt-4-1106-preview": {
|
|
"id": "gpt-4-1106-preview",
|
|
"name": "GPT-4-Turbo",
|
|
"maxLength": 260000,
|
|
"tokenLimit": 126000,
|
|
"context": "128K",
|
|
},
|
|
"gpt-4-plus": {
|
|
"id": "gpt-4-plus",
|
|
"name": "GPT-4-Plus",
|
|
"maxLength": 130000,
|
|
"tokenLimit": 31000,
|
|
"context": "32K",
|
|
},
|
|
"gemini-pro": {
|
|
"id": "gemini-pro",
|
|
"name": "Gemini-Pro",
|
|
"maxLength": 120000,
|
|
"tokenLimit": 30000,
|
|
"context": "32K",
|
|
},
|
|
"claude-2": {
|
|
"id": "claude-2",
|
|
"name": "Claude-2-200k",
|
|
"maxLength": 800000,
|
|
"tokenLimit": 200000,
|
|
"context": "200K",
|
|
},
|
|
"claude-instant-1": {
|
|
"id": "claude-instant-1",
|
|
"name": "Claude-instant-1",
|
|
"maxLength": 400000,
|
|
"tokenLimit": 100000,
|
|
"context": "100K",
|
|
}
|
|
}
|
|
|
|
class Liaobots(AsyncGeneratorProvider, ProviderModelMixin):
|
|
url = "https://liaobots.site"
|
|
working = True
|
|
supports_message_history = True
|
|
supports_gpt_35_turbo = True
|
|
supports_gpt_4 = True
|
|
default_model = "gpt-3.5-turbo"
|
|
models = list(models)
|
|
model_aliases = {
|
|
"claude-v2": "claude-2"
|
|
}
|
|
_auth_code = None
|
|
_cookie_jar = None
|
|
|
|
@classmethod
|
|
async def create_async_generator(
|
|
cls,
|
|
model: str,
|
|
messages: Messages,
|
|
auth: str = None,
|
|
proxy: str = None,
|
|
connector: BaseConnector = None,
|
|
**kwargs
|
|
) -> AsyncResult:
|
|
headers = {
|
|
"authority": "liaobots.com",
|
|
"content-type": "application/json",
|
|
"origin": cls.url,
|
|
"referer": f"{cls.url}/",
|
|
"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36",
|
|
}
|
|
async with ClientSession(
|
|
headers=headers,
|
|
cookie_jar=cls._cookie_jar,
|
|
connector=get_connector(connector, proxy)
|
|
) as session:
|
|
cls._auth_code = auth if isinstance(auth, str) else cls._auth_code
|
|
if not cls._auth_code:
|
|
async with session.post(
|
|
"https://liaobots.work/recaptcha/api/login",
|
|
proxy=proxy,
|
|
data={"token": "abcdefghijklmnopqrst"},
|
|
verify_ssl=False
|
|
) as response:
|
|
response.raise_for_status()
|
|
async with session.post(
|
|
"https://liaobots.work/api/user",
|
|
proxy=proxy,
|
|
json={"authcode": ""},
|
|
verify_ssl=False
|
|
) as response:
|
|
response.raise_for_status()
|
|
cls._auth_code = (await response.json(content_type=None))["authCode"]
|
|
cls._cookie_jar = session.cookie_jar
|
|
|
|
data = {
|
|
"conversationId": str(uuid.uuid4()),
|
|
"model": models[cls.get_model(model)],
|
|
"messages": messages,
|
|
"key": "",
|
|
"prompt": kwargs.get("system_message", "You are ChatGPT, a large language model trained by OpenAI. Follow the user's instructions carefully."),
|
|
}
|
|
async with session.post(
|
|
"https://liaobots.work/api/chat",
|
|
proxy=proxy,
|
|
json=data,
|
|
headers={"x-auth-code": cls._auth_code},
|
|
verify_ssl=False
|
|
) as response:
|
|
response.raise_for_status()
|
|
async for chunk in response.content.iter_any():
|
|
if b"<html coupert-item=" in chunk:
|
|
raise RuntimeError("Invalid session")
|
|
if chunk:
|
|
yield chunk.decode()
|