OpenaiChat.py 34 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707
  1. from __future__ import annotations
  2. import os
  3. import re
  4. import asyncio
  5. import uuid
  6. import json
  7. import base64
  8. import time
  9. import random
  10. from typing import AsyncIterator, Iterator, Optional, Generator, Dict
  11. from copy import copy
  12. try:
  13. import nodriver
  14. has_nodriver = True
  15. except ImportError:
  16. has_nodriver = False
  17. from ..base_provider import AsyncAuthedProvider, ProviderModelMixin
  18. from ...typing import AsyncResult, Messages, Cookies, ImagesType
  19. from ...requests.raise_for_status import raise_for_status
  20. from ...requests import StreamSession
  21. from ...requests import get_nodriver
  22. from ...image import ImageRequest, to_image, to_bytes, is_accepted_format
  23. from ...errors import MissingAuthError, NoValidHarFileError
  24. from ...providers.response import JsonConversation, FinishReason, SynthesizeData, AuthResult, ImageResponse
  25. from ...providers.response import Sources, TitleGeneration, RequestLogin, Parameters, Reasoning
  26. from ..helper import format_cookies
  27. from ..openai.models import default_model, default_image_model, models, image_models, text_models
  28. from ..openai.har_file import get_request_config
  29. from ..openai.har_file import RequestConfig, arkReq, arkose_url, start_url, conversation_url, backend_url, backend_anon_url
  30. from ..openai.proofofwork import generate_proof_token
  31. from ..openai.new import get_requirements_token, get_config
  32. from ... import debug
  33. DEFAULT_HEADERS = {
  34. "accept": "*/*",
  35. "accept-encoding": "gzip, deflate, br, zstd",
  36. 'accept-language': 'en-US,en;q=0.8',
  37. "referer": "https://chatgpt.com/",
  38. "sec-ch-ua": "\"Google Chrome\";v=\"131\", \"Chromium\";v=\"131\", \"Not_A Brand\";v=\"24\"",
  39. "sec-ch-ua-mobile": "?0",
  40. "sec-ch-ua-platform": "\"Windows\"",
  41. "sec-fetch-dest": "empty",
  42. "sec-fetch-mode": "cors",
  43. "sec-fetch-site": "same-origin",
  44. "sec-gpc": "1",
  45. "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/131.0.0.0 Safari/537.36"
  46. }
  47. INIT_HEADERS = {
  48. 'accept': '*/*',
  49. 'accept-language': 'en-US,en;q=0.8',
  50. 'cache-control': 'no-cache',
  51. 'pragma': 'no-cache',
  52. 'priority': 'u=0, i',
  53. "sec-ch-ua": "\"Google Chrome\";v=\"131\", \"Chromium\";v=\"131\", \"Not_A Brand\";v=\"24\"",
  54. 'sec-ch-ua-arch': '"arm"',
  55. 'sec-ch-ua-bitness': '"64"',
  56. 'sec-ch-ua-mobile': '?0',
  57. 'sec-ch-ua-model': '""',
  58. "sec-ch-ua-platform": "\"Windows\"",
  59. 'sec-ch-ua-platform-version': '"14.4.0"',
  60. 'sec-fetch-dest': 'document',
  61. 'sec-fetch-mode': 'navigate',
  62. 'sec-fetch-site': 'none',
  63. 'sec-fetch-user': '?1',
  64. 'upgrade-insecure-requests': '1',
  65. "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/131.0.0.0 Safari/537.36"
  66. }
  67. UPLOAD_HEADERS = {
  68. "accept": "application/json, text/plain, */*",
  69. 'accept-language': 'en-US,en;q=0.8',
  70. "referer": "https://chatgpt.com/",
  71. "priority": "u=1, i",
  72. "sec-ch-ua": "\"Google Chrome\";v=\"131\", \"Chromium\";v=\"131\", \"Not_A Brand\";v=\"24\"",
  73. "sec-ch-ua-mobile": "?0",
  74. 'sec-ch-ua-platform': '"macOS"',
  75. "sec-fetch-dest": "empty",
  76. "sec-fetch-mode": "cors",
  77. "sec-fetch-site": "cross-site",
  78. "x-ms-blob-type": "BlockBlob",
  79. "x-ms-version": "2020-04-08",
  80. "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/131.0.0.0 Safari/537.36"
  81. }
  82. class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin):
  83. """A class for creating and managing conversations with OpenAI chat service"""
  84. label = "OpenAI ChatGPT"
  85. url = "https://chatgpt.com"
  86. working = True
  87. use_nodriver = True
  88. supports_gpt_4 = True
  89. supports_message_history = True
  90. supports_system_message = True
  91. default_model = default_model
  92. default_image_model = default_image_model
  93. image_models = image_models
  94. vision_models = text_models
  95. models = models
  96. synthesize_content_type = "audio/aac"
  97. request_config = RequestConfig()
  98. _expires: int = None
  99. @classmethod
  100. async def on_auth_async(cls, proxy: str = None, **kwargs) -> AsyncIterator:
  101. async for chunk in cls.login(proxy=proxy):
  102. yield chunk
  103. yield AuthResult(
  104. api_key=cls.request_config.access_token,
  105. cookies=cls.request_config.cookies or {},
  106. headers=cls.request_config.headers or cls.get_default_headers(),
  107. expires=cls._expires,
  108. proof_token=cls.request_config.proof_token,
  109. turnstile_token=cls.request_config.turnstile_token
  110. )
  111. @classmethod
  112. async def upload_images(
  113. cls,
  114. session: StreamSession,
  115. auth_result: AuthResult,
  116. images: ImagesType,
  117. ) -> ImageRequest:
  118. """
  119. Upload an image to the service and get the download URL
  120. Args:
  121. session: The StreamSession object to use for requests
  122. headers: The headers to include in the requests
  123. images: The images to upload, either a PIL Image object or a bytes object
  124. Returns:
  125. An ImageRequest object that contains the download URL, file name, and other data
  126. """
  127. async def upload_image(image, image_name):
  128. # Convert the image to a PIL Image object and get the extension
  129. data_bytes = to_bytes(image)
  130. image = to_image(data_bytes)
  131. extension = image.format.lower()
  132. data = {
  133. "file_name": "" if image_name is None else image_name,
  134. "file_size": len(data_bytes),
  135. "use_case": "multimodal"
  136. }
  137. # Post the image data to the service and get the image data
  138. headers = auth_result.headers if hasattr(auth_result, "headers") else None
  139. async with session.post(f"{cls.url}/backend-api/files", json=data, headers=headers) as response:
  140. cls._update_request_args(auth_result, session)
  141. await raise_for_status(response, "Create file failed")
  142. image_data = {
  143. **data,
  144. **await response.json(),
  145. "mime_type": is_accepted_format(data_bytes),
  146. "extension": extension,
  147. "height": image.height,
  148. "width": image.width
  149. }
  150. # Put the image bytes to the upload URL and check the status
  151. await asyncio.sleep(1)
  152. async with session.put(
  153. image_data["upload_url"],
  154. data=data_bytes,
  155. headers={
  156. **UPLOAD_HEADERS,
  157. "Content-Type": image_data["mime_type"],
  158. "x-ms-blob-type": "BlockBlob",
  159. "x-ms-version": "2020-04-08",
  160. "Origin": "https://chatgpt.com",
  161. }
  162. ) as response:
  163. await raise_for_status(response)
  164. # Post the file ID to the service and get the download URL
  165. async with session.post(
  166. f"{cls.url}/backend-api/files/{image_data['file_id']}/uploaded",
  167. json={},
  168. headers=auth_result.headers
  169. ) as response:
  170. cls._update_request_args(auth_result, session)
  171. await raise_for_status(response, "Get download url failed")
  172. image_data["download_url"] = (await response.json())["download_url"]
  173. return ImageRequest(image_data)
  174. if not images:
  175. return
  176. return [await upload_image(image, image_name) for image, image_name in images]
  177. @classmethod
  178. def create_messages(cls, messages: Messages, image_requests: ImageRequest = None, system_hints: list = None):
  179. """
  180. Create a list of messages for the user input
  181. Args:
  182. prompt: The user input as a string
  183. image_response: The image response object, if any
  184. Returns:
  185. A list of messages with the user input and the image, if any
  186. """
  187. # Create a message object with the user role and the content
  188. messages = [{
  189. "id": str(uuid.uuid4()),
  190. "author": {"role": message["role"]},
  191. "content": {"content_type": "text", "parts": [message["content"]]},
  192. "metadata": {"serialization_metadata": {"custom_symbol_offsets": []}, **({"system_hints": system_hints} if system_hints else {})},
  193. "create_time": time.time(),
  194. } for message in messages]
  195. # Check if there is an image response
  196. if image_requests:
  197. # Change content in last user message
  198. messages[-1]["content"] = {
  199. "content_type": "multimodal_text",
  200. "parts": [*[{
  201. "asset_pointer": f"file-service://{image_request.get('file_id')}",
  202. "height": image_request.get("height"),
  203. "size_bytes": image_request.get("file_size"),
  204. "width": image_request.get("width"),
  205. }
  206. for image_request in image_requests],
  207. messages[-1]["content"]["parts"][0]]
  208. }
  209. # Add the metadata object with the attachments
  210. messages[-1]["metadata"] = {
  211. "attachments": [{
  212. "height": image_request.get("height"),
  213. "id": image_request.get("file_id"),
  214. "mimeType": image_request.get("mime_type"),
  215. "name": image_request.get("file_name"),
  216. "size": image_request.get("file_size"),
  217. "width": image_request.get("width"),
  218. }
  219. for image_request in image_requests]
  220. }
  221. return messages
  222. @classmethod
  223. async def get_generated_image(cls, session: StreamSession, auth_result: AuthResult, element: dict, prompt: str = None) -> ImageResponse:
  224. try:
  225. prompt = element["metadata"]["dalle"]["prompt"]
  226. file_id = element["asset_pointer"].split("file-service://", 1)[1]
  227. except TypeError:
  228. return
  229. except Exception as e:
  230. raise RuntimeError(f"No Image: {e.__class__.__name__}: {e}")
  231. try:
  232. async with session.get(f"{cls.url}/backend-api/files/{file_id}/download", headers=auth_result.headers) as response:
  233. cls._update_request_args(auth_result, session)
  234. await raise_for_status(response)
  235. download_url = (await response.json())["download_url"]
  236. return ImageResponse(download_url, prompt)
  237. except Exception as e:
  238. raise RuntimeError(f"Error in downloading image: {e}")
  239. @classmethod
  240. async def create_authed(
  241. cls,
  242. model: str,
  243. messages: Messages,
  244. auth_result: AuthResult,
  245. proxy: str = None,
  246. timeout: int = 180,
  247. auto_continue: bool = False,
  248. history_disabled: bool = False,
  249. action: str = "next",
  250. conversation_id: str = None,
  251. conversation: Conversation = None,
  252. images: ImagesType = None,
  253. return_conversation: bool = False,
  254. max_retries: int = 0,
  255. web_search: bool = False,
  256. **kwargs
  257. ) -> AsyncResult:
  258. """
  259. Create an asynchronous generator for the conversation.
  260. Args:
  261. model (str): The model name.
  262. messages (Messages): The list of previous messages.
  263. proxy (str): Proxy to use for requests.
  264. timeout (int): Timeout for requests.
  265. api_key (str): Access token for authentication.
  266. auto_continue (bool): Flag to automatically continue the conversation.
  267. history_disabled (bool): Flag to disable history and training.
  268. action (str): Type of action ('next', 'continue', 'variant').
  269. conversation_id (str): ID of the conversation.
  270. images (ImagesType): Images to include in the conversation.
  271. return_conversation (bool): Flag to include response fields in the output.
  272. **kwargs: Additional keyword arguments.
  273. Yields:
  274. AsyncResult: Asynchronous results from the generator.
  275. Raises:
  276. RuntimeError: If an error occurs during processing.
  277. """
  278. async with StreamSession(
  279. proxy=proxy,
  280. impersonate="chrome",
  281. timeout=timeout
  282. ) as session:
  283. image_requests = None
  284. if not cls.needs_auth:
  285. if not cls.request_config.headers:
  286. cls._create_request_args(cls.request_config.cookies)
  287. async with session.get(cls.url, headers=INIT_HEADERS) as response:
  288. cls._update_request_args(auth_result, session)
  289. await raise_for_status(response)
  290. else:
  291. if not cls.request_config.headers and getattr(auth_result, "cookies", None):
  292. cls._create_request_args(auth_result.cookies, auth_result.headers)
  293. if not cls._set_api_key(getattr(auth_result, "api_key", None)):
  294. raise MissingAuthError("Access token is not valid")
  295. async with session.get(cls.url, headers=cls.request_config.headers) as response:
  296. cls._update_request_args(auth_result, session)
  297. await raise_for_status(response)
  298. try:
  299. image_requests = await cls.upload_images(session, auth_result, images) if images else None
  300. except Exception as e:
  301. debug.error("OpenaiChat: Upload image failed")
  302. debug.error(e)
  303. model = cls.get_model(model)
  304. if conversation is None:
  305. conversation = Conversation(conversation_id, str(uuid.uuid4()), getattr(auth_result, "cookies", {}).get("oai-did"))
  306. else:
  307. conversation = copy(conversation)
  308. if getattr(auth_result, "cookies", {}).get("oai-did") != getattr(conversation, "user_id", None):
  309. conversation = Conversation(None, str(uuid.uuid4()))
  310. if cls.request_config.access_token is None:
  311. auto_continue = False
  312. conversation.finish_reason = None
  313. sources = Sources([])
  314. while conversation.finish_reason is None:
  315. async with session.post(
  316. f"{cls.url}/backend-anon/sentinel/chat-requirements"
  317. if cls.request_config.access_token is None else
  318. f"{cls.url}/backend-api/sentinel/chat-requirements",
  319. json={"p": None if not getattr(auth_result, "proof_token", None) else get_requirements_token(getattr(auth_result, "proof_token", None))},
  320. headers=cls.request_config.headers
  321. ) as response:
  322. if response.status in (401, 403):
  323. auth_result.reset()
  324. else:
  325. cls._update_request_args(auth_result, session)
  326. await raise_for_status(response)
  327. chat_requirements = await response.json()
  328. need_turnstile = chat_requirements.get("turnstile", {}).get("required", False)
  329. need_arkose = chat_requirements.get("arkose", {}).get("required", False)
  330. chat_token = chat_requirements.get("token")
  331. # if need_arkose and cls.request_config.arkose_token is None:
  332. # await get_request_config(proxy)
  333. # cls._create_request_args(auth_result.cookies, auth_result.headers)
  334. # cls._set_api_key(auth_result.access_token)
  335. # if auth_result.arkose_token is None:
  336. # raise MissingAuthError("No arkose token found in .har file")
  337. if "proofofwork" in chat_requirements:
  338. user_agent = getattr(auth_result, "headers", {}).get("user-agent")
  339. proof_token = getattr(auth_result, "proof_token", None)
  340. if proof_token is None:
  341. auth_result.proof_token = get_config(user_agent)
  342. proofofwork = generate_proof_token(
  343. **chat_requirements["proofofwork"],
  344. user_agent=user_agent,
  345. proof_token=proof_token
  346. )
  347. [debug.log(text) for text in (
  348. #f"Arkose: {'False' if not need_arkose else auth_result.arkose_token[:12]+'...'}",
  349. #f"Proofofwork: {'False' if proofofwork is None else proofofwork[:12]+'...'}",
  350. #f"AccessToken: {'False' if cls._api_key is None else cls._api_key[:12]+'...'}",
  351. )]
  352. if action == "continue" and conversation.message_id is None:
  353. action = "next"
  354. data = {
  355. "action": action,
  356. "parent_message_id": conversation.message_id,
  357. "model": model,
  358. "timezone_offset_min":-60,
  359. "timezone":"Europe/Berlin",
  360. "suggestions":[],
  361. "history_and_training_disabled": history_disabled and not auto_continue and not return_conversation or not cls.needs_auth,
  362. "conversation_mode":{"kind":"primary_assistant","plugin_ids":None},
  363. "force_paragen":False,
  364. "force_paragen_model_slug":"",
  365. "force_rate_limit":False,
  366. "reset_rate_limits":False,
  367. "websocket_request_id": str(uuid.uuid4()),
  368. "system_hints": ["search"] if web_search else None,
  369. "supported_encodings":["v1"],
  370. "conversation_origin":None,
  371. "client_contextual_info":{"is_dark_mode":False,"time_since_loaded":random.randint(20, 500),"page_height":578,"page_width":1850,"pixel_ratio":1,"screen_height":1080,"screen_width":1920},
  372. "paragen_stream_type_override":None,
  373. "paragen_cot_summary_display_override":"allow",
  374. "supports_buffering":True
  375. }
  376. if conversation.conversation_id is not None:
  377. data["conversation_id"] = conversation.conversation_id
  378. debug.log(f"OpenaiChat: Use conversation: {conversation.conversation_id}")
  379. if action != "continue":
  380. data["parent_message_id"] = getattr(conversation, "parent_message_id", conversation.message_id)
  381. conversation.parent_message_id = None
  382. messages = messages if conversation_id is None else [messages[-1]]
  383. data["messages"] = cls.create_messages(messages, image_requests, ["search"] if web_search else None)
  384. headers = {
  385. **cls.request_config.headers,
  386. "accept": "text/event-stream",
  387. "content-type": "application/json",
  388. "openai-sentinel-chat-requirements-token": chat_token,
  389. }
  390. #if cls.request_config.arkose_token:
  391. # headers["openai-sentinel-arkose-token"] = cls.request_config.arkose_token
  392. if proofofwork is not None:
  393. headers["openai-sentinel-proof-token"] = proofofwork
  394. if need_turnstile and getattr(auth_result, "turnstile_token", None) is not None:
  395. headers['openai-sentinel-turnstile-token'] = auth_result.turnstile_token
  396. async with session.post(
  397. f"{cls.url}/backend-anon/conversation"
  398. if cls.request_config.access_token is None else
  399. f"{cls.url}/backend-api/conversation",
  400. json=data,
  401. headers=headers
  402. ) as response:
  403. cls._update_request_args(auth_result, session)
  404. if response.status == 403:
  405. cls.request_config.proof_token = None
  406. raise MissingAuthError("Access token is not valid")
  407. await raise_for_status(response)
  408. buffer = u""
  409. async for line in response.iter_lines():
  410. async for chunk in cls.iter_messages_line(session, auth_result, line, conversation, sources):
  411. if isinstance(chunk, str):
  412. chunk = chunk.replace("\ue203", "").replace("\ue204", "").replace("\ue206", "")
  413. buffer += chunk
  414. if buffer.find(u"\ue200") != -1:
  415. if buffer.find(u"\ue201") != -1:
  416. buffer = buffer.replace("\ue200", "").replace("\ue202", "\n").replace("\ue201", "")
  417. buffer = buffer.replace("navlist\n", "#### ")
  418. def replacer(match):
  419. link = None
  420. if len(sources.list) > int(match.group(1)):
  421. link = sources.list[int(match.group(1))]["url"]
  422. return f"[[{int(match.group(1))+1}]]({link})"
  423. return f" [{int(match.group(1))+1}]"
  424. buffer = re.sub(r'(?:cite\nturn0search|cite\nturn0news|turn0news)(\d+)', replacer, buffer)
  425. else:
  426. continue
  427. yield buffer
  428. buffer = ""
  429. else:
  430. yield chunk
  431. if conversation.finish_reason is not None:
  432. break
  433. if sources.list:
  434. yield sources
  435. if return_conversation:
  436. yield conversation
  437. if not history_disabled and auth_result.api_key is not None:
  438. yield SynthesizeData(cls.__name__, {
  439. "conversation_id": conversation.conversation_id,
  440. "message_id": conversation.message_id,
  441. "voice": "maple",
  442. })
  443. if auto_continue and conversation.finish_reason == "max_tokens":
  444. conversation.finish_reason = None
  445. action = "continue"
  446. await asyncio.sleep(5)
  447. else:
  448. break
  449. yield FinishReason(conversation.finish_reason)
  450. @classmethod
  451. async def iter_messages_line(cls, session: StreamSession, auth_result: AuthResult, line: bytes, fields: Conversation, sources: Sources) -> AsyncIterator:
  452. if not line.startswith(b"data: "):
  453. return
  454. elif line.startswith(b"data: [DONE]"):
  455. if fields.finish_reason is None:
  456. fields.finish_reason = "error"
  457. return
  458. try:
  459. line = json.loads(line[6:])
  460. except:
  461. return
  462. if not isinstance(line, dict):
  463. return
  464. if "type" in line:
  465. if line["type"] == "title_generation":
  466. yield TitleGeneration(line["title"])
  467. if "v" in line:
  468. v = line.get("v")
  469. if isinstance(v, str) and fields.is_recipient:
  470. if "p" not in line or line.get("p") == "/message/content/parts/0":
  471. yield Reasoning(token=v) if fields.is_thinking else v
  472. elif isinstance(v, list):
  473. for m in v:
  474. if m.get("p") == "/message/content/parts/0" and fields.is_recipient:
  475. yield m.get("v")
  476. elif m.get("p") == "/message/metadata/search_result_groups":
  477. for entry in [p.get("entries") for p in m.get("v")]:
  478. for link in entry:
  479. sources.add_source(link)
  480. elif re.match(r"^/message/metadata/content_references/\d+$", m.get("p")):
  481. sources.add_source(m.get("v"))
  482. elif m.get("p") == "/message/metadata/finished_text":
  483. fields.is_thinking = False
  484. yield Reasoning(status=m.get("v"))
  485. elif m.get("p") == "/message/metadata":
  486. fields.finish_reason = m.get("v", {}).get("finish_details", {}).get("type")
  487. break
  488. elif isinstance(v, dict):
  489. if fields.conversation_id is None:
  490. fields.conversation_id = v.get("conversation_id")
  491. debug.log(f"OpenaiChat: New conversation: {fields.conversation_id}")
  492. m = v.get("message", {})
  493. fields.is_recipient = m.get("recipient", "all") == "all"
  494. if fields.is_recipient:
  495. c = m.get("content", {})
  496. if c.get("content_type") == "text" and m.get("author", {}).get("role") == "tool" and "initial_text" in m.get("metadata", {}):
  497. fields.is_thinking = True
  498. yield Reasoning(status=m.get("metadata", {}).get("initial_text"))
  499. if c.get("content_type") == "multimodal_text":
  500. generated_images = []
  501. for element in c.get("parts"):
  502. if isinstance(element, dict) and element.get("content_type") == "image_asset_pointer":
  503. image = cls.get_generated_image(session, auth_result, element)
  504. generated_images.append(image)
  505. for image_response in await asyncio.gather(*generated_images):
  506. if image_response is not None:
  507. yield image_response
  508. if m.get("author", {}).get("role") == "assistant":
  509. if fields.parent_message_id is None:
  510. fields.parent_message_id = v.get("message", {}).get("id")
  511. fields.message_id = v.get("message", {}).get("id")
  512. return
  513. if "error" in line and line.get("error"):
  514. raise RuntimeError(line.get("error"))
  515. @classmethod
  516. async def synthesize(cls, params: dict) -> AsyncIterator[bytes]:
  517. async for _ in cls.login():
  518. pass
  519. async with StreamSession(
  520. impersonate="chrome",
  521. timeout=0
  522. ) as session:
  523. async with session.get(
  524. f"{cls.url}/backend-api/synthesize",
  525. params=params,
  526. headers=cls.request_config.headers
  527. ) as response:
  528. await raise_for_status(response)
  529. async for chunk in response.iter_content():
  530. yield chunk
  531. @classmethod
  532. async def login(
  533. cls,
  534. proxy: str = None,
  535. **kwargs
  536. ) -> AsyncIterator:
  537. if cls._expires is not None and (cls._expires - 60*10) < time.time():
  538. cls.request_config.headers = cls.request_config.access_token = None
  539. if cls.request_config.headers is None:
  540. cls.request_config.headers = {}
  541. if cls.request_config.access_token is not None:
  542. cls._create_request_args(cls.request_config.cookies, cls.request_config.headers)
  543. cls._set_api_key(cls.request_config.access_token)
  544. else:
  545. try:
  546. cls.request_config = await get_request_config(cls.request_config, proxy)
  547. cls._create_request_args(cls.request_config.cookies, cls.request_config.headers)
  548. if cls.request_config.access_token is not None or cls.needs_auth:
  549. if not cls._set_api_key(cls.request_config.access_token):
  550. raise NoValidHarFileError(f"Access token is not valid: {cls.request_config.access_token}")
  551. except NoValidHarFileError:
  552. if has_nodriver:
  553. if cls.request_config.access_token is None:
  554. yield RequestLogin(cls.label, os.environ.get("G4F_LOGIN_URL", ""))
  555. await cls.nodriver_auth(proxy)
  556. else:
  557. raise
  558. @classmethod
  559. async def nodriver_auth(cls, proxy: str = None):
  560. browser, stop_browser = await get_nodriver(proxy=proxy)
  561. try:
  562. page = browser.main_tab
  563. def on_request(event: nodriver.cdp.network.RequestWillBeSent):
  564. if event.request.url == start_url or event.request.url.startswith(conversation_url):
  565. for key, value in event.request.headers.items():
  566. cls.request_config.headers[key.lower()] = value
  567. elif event.request.url in (backend_url, backend_anon_url):
  568. if "OpenAI-Sentinel-Proof-Token" in event.request.headers:
  569. cls.request_config.proof_token = json.loads(base64.b64decode(
  570. event.request.headers["OpenAI-Sentinel-Proof-Token"].split("gAAAAAB", 1)[-1].encode()
  571. ).decode())
  572. if "OpenAI-Sentinel-Turnstile-Token" in event.request.headers:
  573. cls.request_config.turnstile_token = event.request.headers["OpenAI-Sentinel-Turnstile-Token"]
  574. if "Authorization" in event.request.headers:
  575. cls.request_config.access_token = event.request.headers["Authorization"].split()[-1]
  576. elif event.request.url == arkose_url:
  577. cls.request_config.arkose_request = arkReq(
  578. arkURL=event.request.url,
  579. arkBx=None,
  580. arkHeader=event.request.headers,
  581. arkBody=event.request.post_data,
  582. userAgent=event.request.headers.get("User-Agent")
  583. )
  584. await page.send(nodriver.cdp.network.enable())
  585. page.add_handler(nodriver.cdp.network.RequestWillBeSent, on_request)
  586. page = await browser.get(cls.url)
  587. user_agent = await page.evaluate("window.navigator.userAgent")
  588. await page.select("#prompt-textarea", 240)
  589. await page.evaluate("document.getElementById('prompt-textarea').innerText = 'Hello'")
  590. await page.evaluate("document.querySelector('[data-testid=\"send-button\"]').click()")
  591. while True:
  592. if cls.request_config.access_token is not None or not cls.needs_auth:
  593. break
  594. body = await page.evaluate("JSON.stringify(window.__remixContext)")
  595. if body:
  596. match = re.search(r'"accessToken":"(.*?)"', body)
  597. if match:
  598. cls.request_config.access_token = match.group(1)
  599. break
  600. await asyncio.sleep(1)
  601. while True:
  602. if cls.request_config.proof_token:
  603. break
  604. await asyncio.sleep(1)
  605. cls.request_config.data_build = await page.evaluate("document.documentElement.getAttribute('data-build')")
  606. cls.request_config.cookies = await page.send(get_cookies([cls.url]))
  607. await page.close()
  608. cls._create_request_args(cls.request_config.cookies, cls.request_config.headers, user_agent=user_agent)
  609. cls._set_api_key(cls.request_config.access_token)
  610. finally:
  611. stop_browser()
  612. @staticmethod
  613. def get_default_headers() -> Dict[str, str]:
  614. return {
  615. **DEFAULT_HEADERS,
  616. "content-type": "application/json",
  617. }
  618. @classmethod
  619. def _create_request_args(cls, cookies: Cookies = None, headers: dict = None, user_agent: str = None):
  620. cls.request_config.headers = cls.get_default_headers() if headers is None else headers
  621. if user_agent is not None:
  622. cls.request_config.headers["user-agent"] = user_agent
  623. cls.request_config.cookies = {} if cookies is None else cookies
  624. cls._update_cookie_header()
  625. @classmethod
  626. def _update_request_args(cls, auth_result: AuthResult, session: StreamSession):
  627. if hasattr(auth_result, "cookies"):
  628. for c in session.cookie_jar if hasattr(session, "cookie_jar") else session.cookies.jar:
  629. auth_result.cookies[getattr(c, "key", getattr(c, "name", ""))] = c.value
  630. cls.request_config.cookies = auth_result.cookies
  631. cls._update_cookie_header()
  632. @classmethod
  633. def _set_api_key(cls, api_key: str):
  634. if api_key:
  635. exp = api_key.split(".")[1]
  636. exp = (exp + "=" * (4 - len(exp) % 4)).encode()
  637. cls._expires = json.loads(base64.b64decode(exp)).get("exp")
  638. debug.log(f"OpenaiChat: API key expires at\n {cls._expires} we have:\n {time.time()}")
  639. if time.time() > cls._expires:
  640. debug.log(f"OpenaiChat: API key is expired")
  641. else:
  642. cls.request_config.access_token = api_key
  643. cls.request_config.headers["authorization"] = f"Bearer {api_key}"
  644. return True
  645. return False
  646. @classmethod
  647. def _update_cookie_header(cls):
  648. if cls.request_config.cookies:
  649. cls.request_config.headers["cookie"] = format_cookies(cls.request_config.cookies)
  650. class Conversation(JsonConversation):
  651. """
  652. Class to encapsulate response fields.
  653. """
  654. def __init__(self, conversation_id: str = None, message_id: str = None, user_id: str = None, finish_reason: str = None, parent_message_id: str = None, is_thinking: bool = False):
  655. self.conversation_id = conversation_id
  656. self.message_id = message_id
  657. self.finish_reason = finish_reason
  658. self.is_recipient = False
  659. self.parent_message_id = message_id if parent_message_id is None else parent_message_id
  660. self.user_id = user_id
  661. self.is_thinking = is_thinking
  662. def get_cookies(
  663. urls: Optional[Iterator[str]] = None
  664. ) -> Generator[Dict, Dict, Dict[str, str]]:
  665. params = {}
  666. if urls is not None:
  667. params['urls'] = [i for i in urls]
  668. cmd_dict = {
  669. 'method': 'Network.getCookies',
  670. 'params': params,
  671. }
  672. json = yield cmd_dict
  673. return {c["name"]: c["value"] for c in json['cookies']} if 'cookies' in json else {}