2023-09-03 08:26:26 +00:00
|
|
|
from __future__ import annotations
|
|
|
|
|
2023-12-06 08:35:36 +00:00
|
|
|
import uuid, json, asyncio, os
|
2023-11-11 09:14:39 +00:00
|
|
|
from py_arkose_generator.arkose import get_values_for_request
|
2023-11-19 04:36:04 +00:00
|
|
|
from asyncstdlib.itertools import tee
|
|
|
|
from async_property import async_cached_property
|
2023-12-06 08:35:36 +00:00
|
|
|
from selenium.webdriver.common.by import By
|
|
|
|
from selenium.webdriver.support.ui import WebDriverWait
|
|
|
|
from selenium.webdriver.support import expected_conditions as EC
|
|
|
|
|
2023-10-04 05:20:51 +00:00
|
|
|
from ..base_provider import AsyncGeneratorProvider
|
2023-12-06 08:35:36 +00:00
|
|
|
from ..helper import get_event_loop, format_prompt
|
2023-12-02 04:40:07 +00:00
|
|
|
from ...webdriver import get_browser
|
2023-10-09 11:33:20 +00:00
|
|
|
from ...typing import AsyncResult, Messages
|
2023-10-04 05:20:51 +00:00
|
|
|
from ...requests import StreamSession
|
2023-11-19 04:36:04 +00:00
|
|
|
|
|
|
|
models = {
|
|
|
|
"gpt-3.5": "text-davinci-002-render-sha",
|
|
|
|
"gpt-3.5-turbo": "text-davinci-002-render-sha",
|
|
|
|
"gpt-4": "gpt-4",
|
|
|
|
"gpt-4-gizmo": "gpt-4-gizmo"
|
|
|
|
}
|
2023-09-03 08:26:26 +00:00
|
|
|
|
2023-10-03 20:12:56 +00:00
|
|
|
class OpenaiChat(AsyncGeneratorProvider):
|
2023-08-27 15:37:44 +00:00
|
|
|
url = "https://chat.openai.com"
|
2023-09-05 15:27:24 +00:00
|
|
|
working = True
|
2023-11-19 04:36:04 +00:00
|
|
|
needs_auth = True
|
2023-08-25 04:41:32 +00:00
|
|
|
supports_gpt_35_turbo = True
|
2023-11-19 04:36:04 +00:00
|
|
|
supports_gpt_4 = True
|
|
|
|
_access_token: str = None
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
async def create(
|
|
|
|
cls,
|
|
|
|
prompt: str = None,
|
|
|
|
model: str = "",
|
|
|
|
messages: Messages = [],
|
|
|
|
history_disabled: bool = False,
|
|
|
|
action: str = "next",
|
|
|
|
conversation_id: str = None,
|
|
|
|
parent_id: str = None,
|
|
|
|
**kwargs
|
|
|
|
) -> Response:
|
|
|
|
if prompt:
|
2023-11-20 13:00:40 +00:00
|
|
|
messages.append({
|
|
|
|
"role": "user",
|
|
|
|
"content": prompt
|
|
|
|
})
|
2023-11-19 04:36:04 +00:00
|
|
|
generator = cls.create_async_generator(
|
|
|
|
model,
|
|
|
|
messages,
|
|
|
|
history_disabled=history_disabled,
|
|
|
|
action=action,
|
|
|
|
conversation_id=conversation_id,
|
|
|
|
parent_id=parent_id,
|
|
|
|
response_fields=True,
|
|
|
|
**kwargs
|
|
|
|
)
|
|
|
|
return Response(
|
|
|
|
generator,
|
2023-11-20 13:00:40 +00:00
|
|
|
await anext(generator),
|
2023-11-19 04:36:04 +00:00
|
|
|
action,
|
|
|
|
messages,
|
|
|
|
kwargs
|
|
|
|
)
|
2023-08-25 04:41:32 +00:00
|
|
|
|
|
|
|
@classmethod
|
2023-10-03 20:12:56 +00:00
|
|
|
async def create_async_generator(
|
2023-08-25 04:41:32 +00:00
|
|
|
cls,
|
|
|
|
model: str,
|
2023-10-09 11:33:20 +00:00
|
|
|
messages: Messages,
|
2023-08-25 04:41:32 +00:00
|
|
|
proxy: str = None,
|
2023-10-09 11:33:20 +00:00
|
|
|
timeout: int = 120,
|
2023-09-05 15:27:24 +00:00
|
|
|
access_token: str = None,
|
2023-11-12 09:12:05 +00:00
|
|
|
auto_continue: bool = False,
|
2023-11-19 04:36:04 +00:00
|
|
|
history_disabled: bool = True,
|
|
|
|
action: str = "next",
|
|
|
|
conversation_id: str = None,
|
|
|
|
parent_id: str = None,
|
|
|
|
response_fields: bool = False,
|
2023-10-09 11:33:20 +00:00
|
|
|
**kwargs
|
|
|
|
) -> AsyncResult:
|
2023-11-19 04:36:04 +00:00
|
|
|
if not model:
|
|
|
|
model = "gpt-3.5"
|
|
|
|
elif model not in models:
|
|
|
|
raise ValueError(f"Model are not supported: {model}")
|
|
|
|
if not parent_id:
|
|
|
|
parent_id = str(uuid.uuid4())
|
2023-09-05 15:27:24 +00:00
|
|
|
if not access_token:
|
2023-12-06 08:35:36 +00:00
|
|
|
access_token = cls._access_token
|
|
|
|
if not access_token:
|
|
|
|
login_url = os.environ.get("G4F_LOGIN_URL")
|
|
|
|
if login_url:
|
|
|
|
yield f"Please login: [ChatGPT]({login_url})\n\n"
|
|
|
|
access_token = cls._access_token = await cls.browse_access_token(proxy)
|
2023-09-05 15:27:24 +00:00
|
|
|
headers = {
|
|
|
|
"Accept": "text/event-stream",
|
|
|
|
"Authorization": f"Bearer {access_token}",
|
|
|
|
}
|
2023-11-19 04:36:04 +00:00
|
|
|
async with StreamSession(
|
|
|
|
proxies={"https": proxy},
|
|
|
|
impersonate="chrome110",
|
|
|
|
headers=headers,
|
|
|
|
timeout=timeout
|
|
|
|
) as session:
|
|
|
|
end_turn = EndTurn()
|
2023-11-20 13:00:40 +00:00
|
|
|
while not end_turn.is_end:
|
|
|
|
data = {
|
|
|
|
"action": action,
|
|
|
|
"arkose_token": await get_arkose_token(proxy, timeout),
|
|
|
|
"conversation_id": conversation_id,
|
|
|
|
"parent_message_id": parent_id,
|
|
|
|
"model": models[model],
|
|
|
|
"history_and_training_disabled": history_disabled and not auto_continue,
|
|
|
|
}
|
|
|
|
if action != "continue":
|
2023-12-06 08:35:36 +00:00
|
|
|
prompt = format_prompt(messages) if not conversation_id else messages[-1]["content"]
|
2023-11-20 13:00:40 +00:00
|
|
|
data["messages"] = [{
|
|
|
|
"id": str(uuid.uuid4()),
|
|
|
|
"author": {"role": "user"},
|
2023-12-06 08:35:36 +00:00
|
|
|
"content": {"content_type": "text", "parts": [prompt]},
|
2023-11-20 13:00:40 +00:00
|
|
|
}]
|
2023-11-12 09:12:05 +00:00
|
|
|
async with session.post(f"{cls.url}/backend-api/conversation", json=data) as response:
|
|
|
|
try:
|
|
|
|
response.raise_for_status()
|
|
|
|
except:
|
2023-11-19 04:36:04 +00:00
|
|
|
raise RuntimeError(f"Error {response.status_code}: {await response.text()}")
|
|
|
|
last_message = 0
|
2023-11-12 09:12:05 +00:00
|
|
|
async for line in response.iter_lines():
|
2023-11-20 13:00:40 +00:00
|
|
|
if not line.startswith(b"data: "):
|
|
|
|
continue
|
|
|
|
line = line[6:]
|
|
|
|
if line == b"[DONE]":
|
|
|
|
break
|
|
|
|
try:
|
|
|
|
line = json.loads(line)
|
|
|
|
except:
|
|
|
|
continue
|
|
|
|
if "message" not in line:
|
|
|
|
continue
|
|
|
|
if "error" in line and line["error"]:
|
|
|
|
raise RuntimeError(line["error"])
|
|
|
|
if "message_type" not in line["message"]["metadata"]:
|
|
|
|
continue
|
|
|
|
if line["message"]["author"]["role"] != "assistant":
|
|
|
|
continue
|
|
|
|
if line["message"]["metadata"]["message_type"] in ("next", "continue", "variant"):
|
|
|
|
conversation_id = line["conversation_id"]
|
|
|
|
parent_id = line["message"]["id"]
|
|
|
|
if response_fields:
|
|
|
|
response_fields = False
|
|
|
|
yield ResponseFields(conversation_id, parent_id, end_turn)
|
|
|
|
new_message = line["message"]["content"]["parts"][0]
|
|
|
|
yield new_message[last_message:]
|
|
|
|
last_message = len(new_message)
|
|
|
|
if "finish_details" in line["message"]["metadata"]:
|
|
|
|
if line["message"]["metadata"]["finish_details"]["type"] == "stop":
|
|
|
|
end_turn.end()
|
|
|
|
if not auto_continue:
|
|
|
|
break
|
|
|
|
action = "continue"
|
2023-11-12 09:12:05 +00:00
|
|
|
await asyncio.sleep(5)
|
2023-10-03 20:12:56 +00:00
|
|
|
|
|
|
|
@classmethod
|
2023-11-19 04:36:04 +00:00
|
|
|
async def browse_access_token(cls, proxy: str = None) -> str:
|
2023-10-22 21:53:18 +00:00
|
|
|
def browse() -> str:
|
2023-12-06 08:35:36 +00:00
|
|
|
driver = get_browser(proxy=proxy)
|
2023-10-22 21:53:18 +00:00
|
|
|
try:
|
2023-11-19 04:36:04 +00:00
|
|
|
driver.get(f"{cls.url}/")
|
2023-10-22 21:53:18 +00:00
|
|
|
WebDriverWait(driver, 1200).until(
|
|
|
|
EC.presence_of_element_located((By.ID, "prompt-textarea"))
|
|
|
|
)
|
|
|
|
javascript = "return (await (await fetch('/api/auth/session')).json())['accessToken']"
|
|
|
|
return driver.execute_script(javascript)
|
|
|
|
finally:
|
|
|
|
driver.quit()
|
|
|
|
loop = get_event_loop()
|
|
|
|
return await loop.run_in_executor(
|
|
|
|
None,
|
|
|
|
browse
|
|
|
|
)
|
2023-10-28 05:21:00 +00:00
|
|
|
|
2023-11-19 04:36:04 +00:00
|
|
|
async def get_arkose_token(proxy: str = None, timeout: int = None) -> str:
|
2023-10-28 05:21:00 +00:00
|
|
|
config = {
|
|
|
|
"pkey": "3D86FBBA-9D22-402A-B512-3420086BA6CC",
|
|
|
|
"surl": "https://tcr9i.chat.openai.com",
|
|
|
|
"headers": {
|
|
|
|
"User-Agent": 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/107.0.0.0 Safari/537.36'
|
|
|
|
},
|
|
|
|
"site": "https://chat.openai.com",
|
|
|
|
}
|
2023-11-11 09:14:39 +00:00
|
|
|
args_for_request = get_values_for_request(config)
|
|
|
|
async with StreamSession(
|
|
|
|
proxies={"https": proxy},
|
|
|
|
impersonate="chrome107",
|
2023-11-19 04:36:04 +00:00
|
|
|
timeout=timeout
|
2023-11-11 09:14:39 +00:00
|
|
|
) as session:
|
|
|
|
async with session.post(**args_for_request) as response:
|
|
|
|
response.raise_for_status()
|
|
|
|
decoded_json = await response.json()
|
|
|
|
if "token" in decoded_json:
|
|
|
|
return decoded_json["token"]
|
2023-11-19 04:36:04 +00:00
|
|
|
raise RuntimeError(f"Response: {decoded_json}")
|
|
|
|
|
|
|
|
class EndTurn():
|
|
|
|
def __init__(self):
|
|
|
|
self.is_end = False
|
|
|
|
|
|
|
|
def end(self):
|
|
|
|
self.is_end = True
|
|
|
|
|
|
|
|
class ResponseFields():
|
|
|
|
def __init__(
|
|
|
|
self,
|
|
|
|
conversation_id: str,
|
|
|
|
message_id: str,
|
|
|
|
end_turn: EndTurn
|
|
|
|
):
|
|
|
|
self.conversation_id = conversation_id
|
|
|
|
self.message_id = message_id
|
|
|
|
self._end_turn = end_turn
|
|
|
|
|
|
|
|
class Response():
|
|
|
|
def __init__(
|
|
|
|
self,
|
|
|
|
generator: AsyncResult,
|
|
|
|
fields: ResponseFields,
|
|
|
|
action: str,
|
|
|
|
messages: Messages,
|
|
|
|
options: dict
|
|
|
|
):
|
|
|
|
self.aiter, self.copy = tee(generator)
|
|
|
|
self.fields = fields
|
|
|
|
self.action = action
|
|
|
|
self._messages = messages
|
|
|
|
self._options = options
|
|
|
|
|
|
|
|
def __aiter__(self):
|
|
|
|
return self.aiter
|
|
|
|
|
|
|
|
@async_cached_property
|
|
|
|
async def message(self) -> str:
|
|
|
|
return "".join([chunk async for chunk in self.copy])
|
|
|
|
|
|
|
|
async def next(self, prompt: str, **kwargs) -> Response:
|
|
|
|
return await OpenaiChat.create(
|
|
|
|
**self._options,
|
|
|
|
prompt=prompt,
|
|
|
|
messages=await self.messages,
|
|
|
|
action="next",
|
|
|
|
conversation_id=self.fields.conversation_id,
|
|
|
|
parent_id=self.fields.message_id,
|
|
|
|
**kwargs
|
|
|
|
)
|
|
|
|
|
|
|
|
async def do_continue(self, **kwargs) -> Response:
|
|
|
|
if self.end_turn:
|
|
|
|
raise RuntimeError("Can't continue message. Message already finished.")
|
|
|
|
return await OpenaiChat.create(
|
|
|
|
**self._options,
|
|
|
|
messages=await self.messages,
|
|
|
|
action="continue",
|
|
|
|
conversation_id=self.fields.conversation_id,
|
|
|
|
parent_id=self.fields.message_id,
|
|
|
|
**kwargs
|
|
|
|
)
|
|
|
|
|
|
|
|
async def variant(self, **kwargs) -> Response:
|
|
|
|
if self.action != "next":
|
2023-11-20 13:00:40 +00:00
|
|
|
raise RuntimeError("Can't create variant from continue or variant request.")
|
2023-11-19 04:36:04 +00:00
|
|
|
return await OpenaiChat.create(
|
|
|
|
**self._options,
|
|
|
|
messages=self._messages,
|
|
|
|
action="variant",
|
|
|
|
conversation_id=self.fields.conversation_id,
|
|
|
|
parent_id=self.fields.message_id,
|
|
|
|
**kwargs
|
|
|
|
)
|
|
|
|
|
|
|
|
@async_cached_property
|
|
|
|
async def messages(self):
|
|
|
|
messages = self._messages
|
|
|
|
messages.append({
|
|
|
|
"role": "assistant", "content": await self.message
|
|
|
|
})
|
|
|
|
return messages
|
|
|
|
|
|
|
|
@property
|
|
|
|
def end_turn(self):
|
|
|
|
return self.fields._end_turn.is_end
|