From ddb6fa30d9b0a5c0c7a611969c9439b39a597214 Mon Sep 17 00:00:00 2001 From: Heiner Lohaus Date: Sat, 9 Sep 2023 23:07:00 +0200 Subject: Add CodeLinkAva Provider, Improve test_async.py --- g4f/Provider/CodeLinkAva.py | 63 +++++++++++++++++++++++++++++++++++++++++++++ g4f/Provider/HuggingChat.py | 2 +- g4f/Provider/OpenaiChat.py | 2 +- g4f/Provider/__init__.py | 2 ++ 4 files changed, 67 insertions(+), 2 deletions(-) create mode 100644 g4f/Provider/CodeLinkAva.py (limited to 'g4f') diff --git a/g4f/Provider/CodeLinkAva.py b/g4f/Provider/CodeLinkAva.py new file mode 100644 index 00000000..3ab4e264 --- /dev/null +++ b/g4f/Provider/CodeLinkAva.py @@ -0,0 +1,63 @@ +from __future__ import annotations + +from aiohttp import ClientSession +import json + +from ..typing import AsyncGenerator +from .base_provider import AsyncGeneratorProvider + + +class CodeLinkAva(AsyncGeneratorProvider): + url = "https://ava-ai-ef611.web.app" + supports_gpt_35_turbo = True + working = True + + @classmethod + async def create_async_generator( + cls, + model: str, + messages: list[dict[str, str]], + **kwargs + ) -> AsyncGenerator: + headers = { + "User-Agent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36", + "Accept" : "*/*", + "Accept-language" : "en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3", + "Origin" : cls.url, + "Referer" : cls.url + "/", + "Sec-Fetch-Dest" : "empty", + "Sec-Fetch-Mode" : "cors", + "Sec-Fetch-Site" : "same-origin", + } + async with ClientSession( + headers=headers + ) as session: + data = { + "messages": messages, + "temperature": 0.6, + "stream": True, + **kwargs + } + async with session.post("https://ava-alpha-api.codelink.io/api/chat", json=data) as response: + response.raise_for_status() + start = "data: " + async for line in response.content: + line = line.decode() + if line.startswith("data: ") and not line.startswith("data: [DONE]"): + line = json.loads(line[len(start):-1]) + content = line["choices"][0]["delta"].get("content") + if content: + yield content + + + @classmethod + @property + def params(cls): + params = [ + ("model", "str"), + ("messages", "list[dict[str, str]]"), + ("stream", "bool"), + ("temperature", "float"), + ] + param = ", ".join([": ".join(p) for p in params]) + return f"g4f.provider.{cls.__name__} supports: ({param})" \ No newline at end of file diff --git a/g4f/Provider/HuggingChat.py b/g4f/Provider/HuggingChat.py index 11310a69..85f879f3 100644 --- a/g4f/Provider/HuggingChat.py +++ b/g4f/Provider/HuggingChat.py @@ -68,7 +68,7 @@ class HuggingChat(AsyncGeneratorProvider): if "error" in data: raise RuntimeError(data["error"]) elif isinstance(data, list): - yield data[0]["generated_text"] + yield data[0]["generated_text"].strip() else: raise RuntimeError(f"Response: {data}") else: diff --git a/g4f/Provider/OpenaiChat.py b/g4f/Provider/OpenaiChat.py index c93977ec..cbe886f0 100644 --- a/g4f/Provider/OpenaiChat.py +++ b/g4f/Provider/OpenaiChat.py @@ -34,7 +34,7 @@ class OpenaiChat(AsyncProvider): "https": proxy } if not access_token: - access_token = await cls.get_access_token(cookies) + access_token = await cls.get_access_token(cookies, proxies) headers = { "Accept": "text/event-stream", "Authorization": f"Bearer {access_token}", diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index fa1bdb87..3ddd5742 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -8,6 +8,7 @@ from .Bard import Bard from .Bing import Bing from .ChatgptAi import ChatgptAi from .ChatgptLogin import ChatgptLogin +from .CodeLinkAva import CodeLinkAva from .DeepAi import DeepAi from .DfeHub import DfeHub from .EasyChat import EasyChat @@ -44,6 +45,7 @@ __all__ = [ 'Bing', 'ChatgptAi', 'ChatgptLogin', + 'CodeLinkAva', 'DeepAi', 'DfeHub', 'EasyChat', -- cgit v1.2.3