diff options
Diffstat (limited to 'g4f/Provider')
-rw-r--r-- | g4f/Provider/Ecosia.py | 47 | ||||
-rw-r--r-- | g4f/Provider/You.py | 1 | ||||
-rw-r--r-- | g4f/Provider/__init__.py | 3 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/OpenRouter.py | 2 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/Openai.py | 11 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/OpenaiAccount.py | 7 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/OpenaiChat.py | 28 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/__init__.py | 3 |
8 files changed, 84 insertions, 18 deletions
diff --git a/g4f/Provider/Ecosia.py b/g4f/Provider/Ecosia.py new file mode 100644 index 00000000..1cae3560 --- /dev/null +++ b/g4f/Provider/Ecosia.py @@ -0,0 +1,47 @@ + +from __future__ import annotations + +import base64 +import json +from aiohttp import ClientSession, BaseConnector + +from ..typing import AsyncResult, Messages +from ..requests.raise_for_status import raise_for_status +from .base_provider import AsyncGeneratorProvider, ProviderModelMixin +from .helper import get_connector + +class Ecosia(AsyncGeneratorProvider, ProviderModelMixin): + url = "https://www.ecosia.org" + working = True + supports_gpt_35_turbo = True + default_model = "gpt-3.5-turbo-0125" + model_aliases = {"gpt-3.5-turbo": "gpt-3.5-turbo-0125"} + + @classmethod + async def create_async_generator( + cls, + model: str, + messages: Messages, + connector: BaseConnector = None, + green: bool = False, + proxy: str = None, + **kwargs + ) -> AsyncResult: + cls.get_model(model) + headers = { + "authority": "api.ecosia.org", + "accept": "*/*", + "origin": cls.url, + "referer": f"{cls.url}/", + "user-agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.114 Safari/537.36", + } + async with ClientSession(headers=headers, connector=get_connector(connector, proxy)) as session: + data = { + "messages": base64.b64encode(json.dumps(messages).encode()).decode() + } + api_url = f"https://api.ecosia.org/v2/chat/?sp={'eco' if green else 'productivity'}" + async with session.post(api_url, json=data) as response: + await raise_for_status(response) + async for chunk in response.content.iter_any(): + if chunk: + yield chunk.decode(errors="ignore")
\ No newline at end of file diff --git a/g4f/Provider/You.py b/g4f/Provider/You.py index 3ebd40f2..3d3a3513 100644 --- a/g4f/Provider/You.py +++ b/g4f/Provider/You.py @@ -132,7 +132,6 @@ class You(AsyncGeneratorProvider, ProviderModelMixin): @classmethod async def get_cookies(cls, client: StreamSession) -> Cookies: - if not cls._cookies or cls._cookies_used >= 5: cls._cookies = await cls.create_cookies(client) cls._cookies_used = 0 diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index ea64f80a..d5913e3c 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -23,9 +23,11 @@ from .ChatgptFree import ChatgptFree from .ChatgptNext import ChatgptNext from .ChatgptX import ChatgptX from .Cnote import Cnote +from .Cohere import Cohere from .DeepInfra import DeepInfra from .DeepInfraImage import DeepInfraImage from .DuckDuckGo import DuckDuckGo +from .Ecosia import Ecosia from .Feedough import Feedough from .FlowGpt import FlowGpt from .FreeChatgpt import FreeChatgpt @@ -46,7 +48,6 @@ from .ReplicateImage import ReplicateImage from .Vercel import Vercel from .WhiteRabbitNeo import WhiteRabbitNeo from .You import You -from .Cohere import Cohere import sys diff --git a/g4f/Provider/needs_auth/OpenRouter.py b/g4f/Provider/needs_auth/OpenRouter.py index 773d9203..7945784a 100644 --- a/g4f/Provider/needs_auth/OpenRouter.py +++ b/g4f/Provider/needs_auth/OpenRouter.py @@ -9,7 +9,7 @@ class OpenRouter(Openai): label = "OpenRouter" url = "https://openrouter.ai" working = True - default_model = "openrouter/auto" + default_model = "mistralai/mistral-7b-instruct:free" @classmethod def get_models(cls): diff --git a/g4f/Provider/needs_auth/Openai.py b/g4f/Provider/needs_auth/Openai.py index 80318f6d..f73c1011 100644 --- a/g4f/Provider/needs_auth/Openai.py +++ b/g4f/Provider/needs_auth/Openai.py @@ -56,6 +56,7 @@ class Openai(AsyncGeneratorProvider, ProviderModelMixin): await raise_for_status(response) if not stream: data = await response.json() + cls.raise_error(data) choice = data["choices"][0] if "content" in choice["message"]: yield choice["message"]["content"].strip() @@ -70,8 +71,7 @@ class Openai(AsyncGeneratorProvider, ProviderModelMixin): if chunk == b"[DONE]": break data = json.loads(chunk) - if "error_message" in data: - raise ResponseError(data["error_message"]) + cls.raise_error(data) choice = data["choices"][0] if "content" in choice["delta"] and choice["delta"]["content"]: delta = choice["delta"]["content"] @@ -89,6 +89,13 @@ class Openai(AsyncGeneratorProvider, ProviderModelMixin): if "finish_reason" in choice and choice["finish_reason"] is not None: return FinishReason(choice["finish_reason"]) + @staticmethod + def raise_error(data: dict): + if "error_message" in data: + raise ResponseError(data["error_message"]) + elif "error" in data: + raise ResponseError(f'Error {data["error"]["code"]}: {data["error"]["message"]}') + @classmethod def get_headers(cls, stream: bool, api_key: str = None, headers: dict = None) -> dict: return { diff --git a/g4f/Provider/needs_auth/OpenaiAccount.py b/g4f/Provider/needs_auth/OpenaiAccount.py new file mode 100644 index 00000000..5c90b1de --- /dev/null +++ b/g4f/Provider/needs_auth/OpenaiAccount.py @@ -0,0 +1,7 @@ +from __future__ import annotations + +from .OpenaiChat import OpenaiChat + +class OpenaiAccount(OpenaiChat): + label = "OpenAI ChatGPT with Account" + needs_auth = True
\ No newline at end of file diff --git a/g4f/Provider/needs_auth/OpenaiChat.py b/g4f/Provider/needs_auth/OpenaiChat.py index b34daa3e..07620c80 100644 --- a/g4f/Provider/needs_auth/OpenaiChat.py +++ b/g4f/Provider/needs_auth/OpenaiChat.py @@ -23,7 +23,7 @@ except ImportError: from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin from ...webdriver import get_browser -from ...typing import AsyncResult, Messages, Cookies, ImageType, Union, AsyncIterator +from ...typing import AsyncResult, Messages, Cookies, ImageType, AsyncIterator from ...requests import get_args_from_browser, raise_for_status from ...requests.aiohttp import StreamSession from ...image import to_image, to_bytes, ImageResponse, ImageRequest @@ -35,7 +35,7 @@ from ... import debug class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin): """A class for creating and managing conversations with OpenAI chat service""" - lebel = "OpenAI ChatGPT" + label = "OpenAI ChatGPT" url = "https://chat.openai.com" working = True supports_gpt_35_turbo = True @@ -295,7 +295,7 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin): model: str, messages: Messages, proxy: str = None, - timeout: int = 120, + timeout: int = 180, api_key: str = None, cookies: Cookies = None, auto_continue: bool = False, @@ -348,7 +348,7 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin): if api_key is not None: cls._set_api_key(api_key) - if cls.default_model is None and cls._api_key is not None: + if cls.default_model is None and (not cls.needs_auth or cls._api_key is not None): try: if not model: cls.default_model = cls.get_model(await cls.get_default_model(session, cls._headers)) @@ -368,12 +368,12 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin): arkose_token, api_key, cookies = await getArkoseAndAccessToken(proxy) cls._create_request_args(cookies) cls._set_api_key(api_key) - except NoValidHarFileError: + except NoValidHarFileError as e: ... if cls._api_key is None: - if debug.logging: - print("Getting access token with nodriver.") await cls.nodriver_access_token() + if cls._api_key is None and cls.needs_auth: + raise e cls.default_model = cls.get_model(await cls.get_default_model(session, cls._headers)) async with session.post( @@ -589,10 +589,11 @@ this.fetch = async (url, options) => { user_data_dir = user_config_dir("g4f-nodriver") except: user_data_dir = None - + if debug.logging: + print(f"Open nodriver with user_dir: {user_data_dir}") browser = await uc.start(user_data_dir=user_data_dir) page = await browser.get("https://chat.openai.com/") - while await page.query_selector("#prompt-textarea") is None: + while await page.find("[id^=headlessui-menu-button-]") is None: await asyncio.sleep(1) api_key = await page.evaluate( "(async () => {" @@ -609,8 +610,9 @@ this.fetch = async (url, options) => { for c in await page.browser.cookies.get_all(): if c.domain.endswith("chat.openai.com"): cookies[c.name] = c.value + user_agent = await page.evaluate("window.navigator.userAgent") await page.close() - cls._create_request_args(cookies) + cls._create_request_args(cookies, user_agent) cls._set_api_key(api_key) @classmethod @@ -662,7 +664,7 @@ this.fetch = async (url, options) => { "content-type": "application/json", "oai-device-id": str(uuid.uuid4()), "oai-language": "en-US", - "sec-ch-ua": "\"Chromium\";v=\"122\", \"Not(A:Brand\";v=\"24\", \"Google Chrome\";v=\"122\"", + "sec-ch-ua": "\"Google Chrome\";v=\"123\", \"Not:A-Brand\";v=\"8\", \"Chromium\";v=\"123\"", "sec-ch-ua-mobile": "?0", "sec-ch-ua-platform": "\"Linux\"", "sec-fetch-dest": "empty", @@ -675,8 +677,10 @@ this.fetch = async (url, options) => { return "; ".join(f"{k}={v}" for k, v in cookies.items() if k != "access_token") @classmethod - def _create_request_args(cls, cookies: Cookies = None): + def _create_request_args(cls, cookies: Cookies = None, user_agent: str = None): cls._headers = cls.get_default_headers() + if user_agent is not None: + cls._headers["user-agent"] = user_agent cls._cookies = {} if cookies is None else cookies cls._update_cookie_header() diff --git a/g4f/Provider/needs_auth/__init__.py b/g4f/Provider/needs_auth/__init__.py index 7b793223..805d9fca 100644 --- a/g4f/Provider/needs_auth/__init__.py +++ b/g4f/Provider/needs_auth/__init__.py @@ -6,4 +6,5 @@ from .OpenaiChat import OpenaiChat from .Poe import Poe from .Openai import Openai from .Groq import Groq -from .OpenRouter import OpenRouter
\ No newline at end of file +from .OpenRouter import OpenRouter +from .OpenaiAccount import OpenaiAccount
\ No newline at end of file |