gpt4free/g4f/Provider/Liaobots.py

162 lines
5.0 KiB
Python
Raw Normal View History

from __future__ import annotations
import uuid
2024-01-23 22:48:11 +00:00
from aiohttp import ClientSession, BaseConnector
2023-07-28 10:07:17 +00:00
2023-10-09 08:22:17 +00:00
from ..typing import AsyncResult, Messages
from .base_provider import AsyncGeneratorProvider, ProviderModelMixin
2024-01-23 23:46:35 +00:00
from .helper import get_connector
from ..requests import raise_for_status
2023-07-28 10:07:17 +00:00
models = {
"gpt-3.5-turbo": {
"id": "gpt-3.5-turbo",
2023-12-29 01:16:47 +00:00
"name": "GPT-3.5-Turbo",
"maxLength": 48000,
"tokenLimit": 14000,
"context": "16K",
},
2024-03-27 07:03:09 +00:00
"gpt-4-turbo": {
"id": "gpt-4-turbo-preview",
2023-12-29 01:16:47 +00:00
"name": "GPT-4-Turbo",
"maxLength": 260000,
"tokenLimit": 126000,
"context": "128K",
},
2024-03-26 16:18:37 +00:00
"gpt-4": {
2023-12-29 01:16:47 +00:00
"id": "gpt-4-plus",
"name": "GPT-4-Plus",
"maxLength": 130000,
"tokenLimit": 31000,
"context": "32K",
},
"gpt-4-0613": {
"id": "gpt-4-0613",
"name": "GPT-4-0613",
"maxLength": 60000,
"tokenLimit": 15000,
"context": "16K",
},
2023-12-29 01:16:47 +00:00
"gemini-pro": {
"id": "gemini-pro",
"name": "Gemini-Pro",
"maxLength": 120000,
"tokenLimit": 30000,
"context": "32K",
},
"claude-3-opus-20240229": {
"id": "claude-3-opus-20240229",
"name": "Claude-3-Opus",
2023-12-29 01:16:47 +00:00
"maxLength": 800000,
"tokenLimit": 200000,
"context": "200K",
},
"claude-3-sonnet-20240229": {
"id": "claude-3-sonnet-20240229",
"name": "Claude-3-Sonnet",
"maxLength": 800000,
"tokenLimit": 200000,
"context": "200K",
},
"claude-2.1": {
"id": "claude-2.1",
"name": "Claude-2.1-200k",
"maxLength": 800000,
"tokenLimit": 200000,
"context": "200K",
},
"claude-2.0": {
"id": "claude-2.0",
"name": "Claude-2.0-100k",
"maxLength": 400000,
"tokenLimit": 100000,
"context": "100K",
},
2023-12-29 01:16:47 +00:00
"claude-instant-1": {
"id": "claude-instant-1",
"name": "Claude-instant-1",
"maxLength": 400000,
"tokenLimit": 100000,
"context": "100K",
}
}
2023-07-28 10:07:17 +00:00
class Liaobots(AsyncGeneratorProvider, ProviderModelMixin):
url = "https://liaobots.site"
working = True
2023-10-27 20:59:14 +00:00
supports_message_history = True
supports_system_message = True
supports_gpt_35_turbo = True
supports_gpt_4 = True
default_model = "gpt-3.5-turbo"
models = list(models)
model_aliases = {
"claude-v2": "claude-2"
}
_auth_code = None
2023-12-31 21:59:24 +00:00
_cookie_jar = None
2023-07-28 10:07:17 +00:00
@classmethod
async def create_async_generator(
cls,
2023-07-28 10:07:17 +00:00
model: str,
2023-10-09 08:22:17 +00:00
messages: Messages,
auth: str = None,
proxy: str = None,
2024-01-23 22:48:11 +00:00
connector: BaseConnector = None,
**kwargs
2023-10-09 08:22:17 +00:00
) -> AsyncResult:
2023-07-28 10:07:17 +00:00
headers = {
"authority": "liaobots.com",
"content-type": "application/json",
"origin": cls.url,
"referer": f"{cls.url}/",
"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36",
2023-07-28 10:07:17 +00:00
}
async with ClientSession(
2023-12-31 21:59:24 +00:00
headers=headers,
2024-01-23 22:48:11 +00:00
cookie_jar=cls._cookie_jar,
2024-02-26 10:25:07 +00:00
connector=get_connector(connector, proxy, True)
) as session:
cls._auth_code = auth if isinstance(auth, str) else cls._auth_code
if not cls._auth_code:
async with session.post(
"https://liaobots.work/recaptcha/api/login",
proxy=proxy,
data={"token": "abcdefghijklmnopqrst"},
verify_ssl=False
) as response:
await raise_for_status(response)
async with session.post(
"https://liaobots.work/api/user",
proxy=proxy,
json={"authcode": ""},
verify_ssl=False
) as response:
await raise_for_status(response)
cls._auth_code = (await response.json(content_type=None))["authCode"]
2023-12-31 21:59:24 +00:00
cls._cookie_jar = session.cookie_jar
data = {
"conversationId": str(uuid.uuid4()),
"model": models[cls.get_model(model)],
"messages": messages,
"key": "",
2023-11-16 18:46:25 +00:00
"prompt": kwargs.get("system_message", "You are ChatGPT, a large language model trained by OpenAI. Follow the user's instructions carefully."),
}
async with session.post(
"https://liaobots.work/api/chat",
proxy=proxy,
json=data,
headers={"x-auth-code": cls._auth_code},
verify_ssl=False
) as response:
await raise_for_status(response)
2023-12-31 21:59:24 +00:00
async for chunk in response.content.iter_any():
if b"<html coupert-item=" in chunk:
raise RuntimeError("Invalid session")
if chunk:
yield chunk.decode()