summaryrefslogtreecommitdiffstats
path: root/g4f/requests.py
diff options
context:
space:
mode:
Diffstat (limited to 'g4f/requests.py')
-rw-r--r--g4f/requests.py143
1 files changed, 22 insertions, 121 deletions
diff --git a/g4f/requests.py b/g4f/requests.py
index 46f3b457..275e108b 100644
--- a/g4f/requests.py
+++ b/g4f/requests.py
@@ -1,130 +1,21 @@
from __future__ import annotations
-import json
-from functools import partialmethod
-from typing import AsyncGenerator
from urllib.parse import urlparse
try:
- from curl_cffi.requests import AsyncSession, Session, Response
+ from curl_cffi.requests import Session
+ from .requests_curl_cffi import StreamResponse, StreamSession
has_curl_cffi = True
except ImportError:
Session = type
+ from .requests_aiohttp import StreamResponse, StreamSession
has_curl_cffi = False
from .webdriver import WebDriver, WebDriverSession, bypass_cloudflare, get_driver_cookies
from .errors import MissingRequirementsError
-if not has_curl_cffi:
- from aiohttp import ClientSession, ClientResponse, ClientTimeout
- from .Provider.helper import get_connector
-
- class StreamResponse(ClientResponse):
- async def iter_lines(self) -> iter[bytes, None]:
- async for line in self.content:
- yield line.rstrip(b"\r\n")
-
- async def json(self):
- return await super().json(content_type=None)
-
- class StreamSession(ClientSession):
- def __init__(self, headers: dict = {}, timeout: int = None, proxies: dict = {}, impersonate = None, **kwargs):
- if impersonate:
- headers = {
- 'Accept-Encoding': 'gzip, deflate, br',
- 'Accept-Language': 'en-US',
- 'Connection': 'keep-alive',
- 'Sec-Fetch-Dest': 'empty',
- 'Sec-Fetch-Mode': 'cors',
- 'Sec-Fetch-Site': 'same-site',
- "User-Agent": 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/107.0.0.0 Safari/537.36',
- 'Accept': '*/*',
- 'sec-ch-ua': '"Google Chrome";v="107", "Chromium";v="107", "Not?A_Brand";v="24"',
- 'sec-ch-ua-mobile': '?0',
- 'sec-ch-ua-platform': '"Windows"',
- **headers
- }
- super().__init__(
- **kwargs,
- timeout=ClientTimeout(timeout) if timeout else None,
- response_class=StreamResponse,
- connector=get_connector(kwargs.get("connector"), proxies.get("https")),
- headers=headers
- )
-else:
- class StreamResponse:
- """
- A wrapper class for handling asynchronous streaming responses.
-
- Attributes:
- inner (Response): The original Response object.
- """
-
- def __init__(self, inner: Response) -> None:
- """Initialize the StreamResponse with the provided Response object."""
- self.inner: Response = inner
-
- async def text(self) -> str:
- """Asynchronously get the response text."""
- return await self.inner.atext()
-
- def raise_for_status(self) -> None:
- """Raise an HTTPError if one occurred."""
- self.inner.raise_for_status()
-
- async def json(self, **kwargs) -> dict:
- """Asynchronously parse the JSON response content."""
- return json.loads(await self.inner.acontent(), **kwargs)
-
- async def iter_lines(self) -> AsyncGenerator[bytes, None]:
- """Asynchronously iterate over the lines of the response."""
- async for line in self.inner.aiter_lines():
- yield line
-
- async def iter_content(self) -> AsyncGenerator[bytes, None]:
- """Asynchronously iterate over the response content."""
- async for chunk in self.inner.aiter_content():
- yield chunk
-
- async def __aenter__(self):
- """Asynchronously enter the runtime context for the response object."""
- inner: Response = await self.inner
- self.inner = inner
- self.request = inner.request
- self.status_code: int = inner.status_code
- self.reason: str = inner.reason
- self.ok: bool = inner.ok
- self.headers = inner.headers
- self.cookies = inner.cookies
- return self
-
- async def __aexit__(self, *args):
- """Asynchronously exit the runtime context for the response object."""
- await self.inner.aclose()
-
- class StreamSession(AsyncSession):
- """
- An asynchronous session class for handling HTTP requests with streaming.
-
- Inherits from AsyncSession.
- """
- def request(
- self, method: str, url: str, **kwargs
- ) -> StreamResponse:
- """Create and return a StreamResponse object for the given HTTP request."""
- return StreamResponse(super().request(method, url, stream=True, **kwargs))
-
- # Defining HTTP methods as partial methods of the request method.
- head = partialmethod(request, "HEAD")
- get = partialmethod(request, "GET")
- post = partialmethod(request, "POST")
- put = partialmethod(request, "PUT")
- patch = partialmethod(request, "PATCH")
- delete = partialmethod(request, "DELETE")
-
-
-def get_session_from_browser(url: str, webdriver: WebDriver = None, proxy: str = None, timeout: int = 120) -> Session:
+def get_args_from_browser(url: str, webdriver: WebDriver = None, proxy: str = None, timeout: int = 120) -> dict:
"""
Create a Session object using a WebDriver to handle cookies and headers.
@@ -137,26 +28,36 @@ def get_session_from_browser(url: str, webdriver: WebDriver = None, proxy: str =
Returns:
Session: A Session object configured with cookies and headers from the WebDriver.
"""
- if not has_curl_cffi:
- raise MissingRequirementsError('Install "curl_cffi" package')
- with WebDriverSession(webdriver, "", proxy=proxy, virtual_display=True) as driver:
+ with WebDriverSession(webdriver, "", proxy=proxy, virtual_display=False) as driver:
bypass_cloudflare(driver, url, timeout)
cookies = get_driver_cookies(driver)
user_agent = driver.execute_script("return navigator.userAgent")
-
parse = urlparse(url)
- return Session(
- cookies=cookies,
- headers={
+ return {
+ 'cookies': cookies,
+ 'headers': {
'accept': '*/*',
+ "accept-language": "en-US",
+ "accept-encoding": "gzip, deflate, br",
'authority': parse.netloc,
'origin': f'{parse.scheme}://{parse.netloc}',
'referer': url,
+ "sec-ch-ua": "\"Google Chrome\";v=\"121\", \"Not;A=Brand\";v=\"8\", \"Chromium\";v=\"121\"",
+ "sec-ch-ua-mobile": "?0",
+ "sec-ch-ua-platform": "Windows",
'sec-fetch-dest': 'empty',
'sec-fetch-mode': 'cors',
'sec-fetch-site': 'same-origin',
- 'user-agent': user_agent
+ 'user-agent': user_agent,
},
+ }
+
+def get_session_from_browser(url: str, webdriver: WebDriver = None, proxy: str = None, timeout: int = 120) -> Session:
+ if not has_curl_cffi:
+ raise MissingRequirementsError('Install "curl_cffi" package')
+ args = get_args_from_browser(url, webdriver, proxy, timeout)
+ return Session(
+ **args,
proxies={"https": proxy, "http": proxy},
timeout=timeout,
impersonate="chrome110"