From d6382a2145e7e1c90b3795e3116e3c574e787868 Mon Sep 17 00:00:00 2001 From: Heiner Lohaus Date: Tue, 12 Sep 2023 00:47:03 +0200 Subject: Add new providers: Aivvm, Vitalentum and Ylokh --- g4f/Provider/Aivvm.py | 78 ++++++++++++++++++++++++++++++++++++++++++++++ g4f/Provider/ChatBase.py | 8 ++--- g4f/Provider/Vercel.py | 2 +- g4f/Provider/Vitalentum.py | 66 +++++++++++++++++++++++++++++++++++++++ g4f/Provider/Ylokh.py | 76 ++++++++++++++++++++++++++++++++++++++++++++ g4f/Provider/You.py | 40 ++++++++++++------------ g4f/Provider/__init__.py | 6 ++++ 7 files changed, 251 insertions(+), 25 deletions(-) create mode 100644 g4f/Provider/Aivvm.py create mode 100644 g4f/Provider/Vitalentum.py create mode 100644 g4f/Provider/Ylokh.py (limited to 'g4f') diff --git a/g4f/Provider/Aivvm.py b/g4f/Provider/Aivvm.py new file mode 100644 index 00000000..dbfc588d --- /dev/null +++ b/g4f/Provider/Aivvm.py @@ -0,0 +1,78 @@ +from __future__ import annotations + +from aiohttp import ClientSession + +from .base_provider import AsyncGeneratorProvider +from ..typing import AsyncGenerator + +models = { + "gpt-4": { + "id": "gpt-4", + "name": "GPT-4", + }, + "gpt-3.5-turbo": { + "id": "gpt-3.5-turbo", + "name": "GPT-3.5", + }, + "gpt-3.5-turbo-16k": { + "id": "gpt-3.5-turbo-16k", + "name": "GPT-3.5-16k", + }, +} + +class Aivvm(AsyncGeneratorProvider): + url = "https://chat.aivvm.com" + working = True + supports_gpt_35_turbo = True + supports_gpt_4 = True + + + @classmethod + async def create_async_generator( + cls, + model: str, + messages: list[dict[str, str]], + proxy: str = None, + **kwargs + ) -> AsyncGenerator: + model = model if model else "gpt-3.5-turbo" + if model not in models: + raise ValueError(f"Model are not supported: {model}") + headers = { + "User-Agent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36", + "Accept" : "*/*", + "Accept-language" : "en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3", + "Origin" : cls.url, + "Referer" : cls.url + "/", + "Sec-Fetch-Dest" : "empty", + "Sec-Fetch-Mode" : "cors", + "Sec-Fetch-Site" : "same-origin", + } + async with ClientSession( + headers=headers + ) as session: + data = { + "temperature": 1, + "key": "", + "messages": messages, + "model": models[model], + "prompt": "", + **kwargs + } + async with session.post(cls.url + "/api/chat", json=data, proxy=proxy) as response: + response.raise_for_status() + async for stream in response.content.iter_any(): + yield stream.decode() + + + @classmethod + @property + def params(cls): + params = [ + ("model", "str"), + ("messages", "list[dict[str, str]]"), + ("stream", "bool"), + ("temperature", "float"), + ] + param = ", ".join([": ".join(p) for p in params]) + return f"g4f.provider.{cls.__name__} supports: ({param})" \ No newline at end of file diff --git a/g4f/Provider/ChatBase.py b/g4f/Provider/ChatBase.py index 7d73fd2f..b98fe565 100644 --- a/g4f/Provider/ChatBase.py +++ b/g4f/Provider/ChatBase.py @@ -21,8 +21,10 @@ class ChatBase(AsyncGeneratorProvider): ) -> AsyncGenerator: if model == "gpt-4": chat_id = "quran---tafseer-saadi-pdf-wbgknt7zn" - elif model == "gpt-3.5-turbo" or True: + elif model == "gpt-3.5-turbo" or not model: chat_id = "chatbase--1--pdf-p680fxvnm" + else: + raise ValueError(f"Model are not supported: {model}") headers = { "User-Agent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36", "Accept" : "*/*", @@ -45,9 +47,7 @@ class ChatBase(AsyncGeneratorProvider): async with session.post("https://www.chatbase.co/api/fe/chat", json=data) as response: response.raise_for_status() async for stream in response.content.iter_any(): - stream = stream.decode() - if stream: - yield stream + yield stream.decode() @classmethod diff --git a/g4f/Provider/Vercel.py b/g4f/Provider/Vercel.py index b0e1d8af..79bcf3f4 100644 --- a/g4f/Provider/Vercel.py +++ b/g4f/Provider/Vercel.py @@ -55,7 +55,7 @@ class Vercel(AsyncProvider): def _create_payload(model: str, messages: list[dict[str, str]]) -> dict[str, Any]: if model not in model_info: - raise RuntimeError(f'Model "{model}" are not supported') + raise ValueError(f'Model are not supported: {model}') default_params = model_info[model]["default_params"] return { "messages": messages, diff --git a/g4f/Provider/Vitalentum.py b/g4f/Provider/Vitalentum.py new file mode 100644 index 00000000..31ad8b80 --- /dev/null +++ b/g4f/Provider/Vitalentum.py @@ -0,0 +1,66 @@ +from __future__ import annotations + +import json +from aiohttp import ClientSession + +from .base_provider import AsyncGeneratorProvider +from ..typing import AsyncGenerator + +class Vitalentum(AsyncGeneratorProvider): + url = "https://app.vitalentum.io" + working = True + supports_gpt_35_turbo = True + + + @classmethod + async def create_async_generator( + cls, + model: str, + messages: list[dict[str, str]], + proxy: str = None, + **kwargs + ) -> AsyncGenerator: + headers = { + "User-Agent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36", + "Accept" : "text/event-stream", + "Accept-language" : "de,en-US;q=0.7,en;q=0.3", + "Origin" : cls.url, + "Referer" : cls.url + "/", + "Sec-Fetch-Dest" : "empty", + "Sec-Fetch-Mode" : "cors", + "Sec-Fetch-Site" : "same-origin", + } + conversation = json.dumps({"history": [{ + "speaker": "human" if message["role"] == "user" else "bot", + "text": message["content"], + } for message in messages]}) + data = { + "conversation": conversation, + "temperature": 0.7, + **kwargs + } + async with ClientSession( + headers=headers + ) as session: + async with session.post(cls.url + "/api/converse-edge", json=data, proxy=proxy) as response: + response.raise_for_status() + async for line in response.content: + line = line.decode() + if line.startswith("data: ") and not line.startswith("data: [DONE]"): + line = json.loads(line[6:-1]) + content = line["choices"][0]["delta"].get("content") + if content: + yield content + + + @classmethod + @property + def params(cls): + params = [ + ("model", "str"), + ("messages", "list[dict[str, str]]"), + ("stream", "bool"), + ("temperature", "float"), + ] + param = ", ".join([": ".join(p) for p in params]) + return f"g4f.provider.{cls.__name__} supports: ({param})" \ No newline at end of file diff --git a/g4f/Provider/Ylokh.py b/g4f/Provider/Ylokh.py new file mode 100644 index 00000000..1986b6d3 --- /dev/null +++ b/g4f/Provider/Ylokh.py @@ -0,0 +1,76 @@ +from __future__ import annotations + +import json +from aiohttp import ClientSession + +from .base_provider import AsyncGeneratorProvider +from ..typing import AsyncGenerator + +class Ylokh(AsyncGeneratorProvider): + url = "https://chat.ylokh.xyz" + working = True + supports_gpt_35_turbo = True + + + @classmethod + async def create_async_generator( + cls, + model: str, + messages: list[dict[str, str]], + stream: bool = True, + proxy: str = None, + **kwargs + ) -> AsyncGenerator: + model = model if model else "gpt-3.5-turbo" + headers = { + "User-Agent" : "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/116.0", + "Accept" : "*/*", + "Accept-language" : "de,en-US;q=0.7,en;q=0.3", + "Origin" : cls.url, + "Referer" : cls.url + "/", + "Sec-Fetch-Dest" : "empty", + "Sec-Fetch-Mode" : "cors", + "Sec-Fetch-Site" : "same-origin", + } + data = { + "messages": messages, + "model": model, + "temperature": 1, + "presence_penalty": 0, + "top_p": 1, + "frequency_penalty": 0, + "allow_fallback": True, + "stream": stream, + **kwargs + } + async with ClientSession( + headers=headers + ) as session: + async with session.post("https://chatapi.ylokh.xyz/v1/chat/completions", json=data, proxy=proxy) as response: + response.raise_for_status() + if stream: + async for line in response.content: + line = line.decode() + if line.startswith("data: ") and not line.startswith("data: [DONE]"): + line = json.loads(line[6:-1]) + content = line["choices"][0]["delta"].get("content") + if content: + yield content + else: + chat = await response.json() + yield chat["choices"][0]["message"].get("content") + + + + @classmethod + @property + def params(cls): + params = [ + ("model", "str"), + ("messages", "list[dict[str, str]]"), + ("stream", "bool"), + ("proxy", "str"), + ("temperature", "float"), + ] + param = ", ".join([": ".join(p) for p in params]) + return f"g4f.provider.{cls.__name__} supports: ({param})" \ No newline at end of file diff --git a/g4f/Provider/You.py b/g4f/Provider/You.py index 121d1dbd..4f49f15e 100644 --- a/g4f/Provider/You.py +++ b/g4f/Provider/You.py @@ -2,39 +2,39 @@ from __future__ import annotations import json -from aiohttp import ClientSession +from curl_cffi.requests import AsyncSession from ..typing import AsyncGenerator -from .base_provider import AsyncGeneratorProvider, format_prompt, get_cookies +from .base_provider import AsyncGeneratorProvider, format_prompt class You(AsyncGeneratorProvider): url = "https://you.com" working = True supports_gpt_35_turbo = True - supports_stream = True + supports_stream = False - @staticmethod + + @classmethod async def create_async_generator( + cls, model: str, messages: list[dict[str, str]], - cookies: dict = None, + proxy: str = None, **kwargs, ) -> AsyncGenerator: - if not cookies: - cookies = get_cookies("you.com") - headers = { - "Accept": "text/event-stream", - "Referer": "https://you.com/search?fromSearchBar=true&tbm=youchat", - "User-Agent": "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/116.0" - } - async with ClientSession(headers=headers, cookies=cookies) as session: - async with session.get( + async with AsyncSession(proxies={"https": proxy}, impersonate="chrome107") as session: + headers = { + "Accept": "text/event-stream", + "Referer": "https://you.com/search?fromSearchBar=true&tbm=youchat", + } + response = await session.get( "https://you.com/api/streamingSearch", params={"q": format_prompt(messages), "domain": "youchat", "chat": ""}, - ) as response: - start = 'data: {"youChatToken": ' - async for line in response.content: - line = line.decode('utf-8') - if line.startswith(start): - yield json.loads(line[len(start): -2]) \ No newline at end of file + headers=headers + ) + response.raise_for_status() + start = 'data: {"youChatToken": ' + for line in response.text.splitlines(): + if line.startswith(start): + yield json.loads(line[len(start): -1]) \ No newline at end of file diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index a1abe452..2a3d820e 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -4,6 +4,7 @@ from .Aichat import Aichat from .Ails import Ails from .AiService import AiService from .AItianhu import AItianhu +from .Aivvm import Aivvm from .Bard import Bard from .Bing import Bing from .ChatBase import ChatBase @@ -25,7 +26,9 @@ from .OpenAssistant import OpenAssistant from .Raycast import Raycast from .Theb import Theb from .Vercel import Vercel +from .Vitalentum import Vitalentum from .Wewordle import Wewordle +from .Ylokh import Ylokh from .You import You from .Yqcloud import Yqcloud from .Equing import Equing @@ -42,6 +45,7 @@ __all__ = [ 'Ails', 'AiService', 'AItianhu', + 'Aivvm', 'Bard', 'Bing', 'ChatBase', @@ -63,7 +67,9 @@ __all__ = [ 'OpenAssistant', 'Theb', 'Vercel', + 'Vitalentum', 'Wewordle', + 'Ylokh', 'You', 'Yqcloud', 'Equing', -- cgit v1.2.3 From 618addd001077d9e4ed4d2a7aa03769a0ee0901b Mon Sep 17 00:00:00 2001 From: Heiner Lohaus Date: Tue, 12 Sep 2023 04:40:10 +0200 Subject: Disable not working providers: EasyChat, Equing, GetGpt, Liaobots and Wuguokai No and random best provider for a model Update providers in Readme --- g4f/Provider/EasyChat.py | 2 +- g4f/Provider/Equing.py | 6 +++--- g4f/Provider/GetGpt.py | 2 +- g4f/Provider/Liaobots.py | 2 +- g4f/Provider/Wuguokai.py | 11 +++-------- g4f/__init__.py | 13 +++++++++++-- g4f/models.py | 21 +++++++++++++-------- 7 files changed, 33 insertions(+), 24 deletions(-) (limited to 'g4f') diff --git a/g4f/Provider/EasyChat.py b/g4f/Provider/EasyChat.py index 946d4a4d..dae5196d 100644 --- a/g4f/Provider/EasyChat.py +++ b/g4f/Provider/EasyChat.py @@ -13,7 +13,7 @@ class EasyChat(BaseProvider): url: str = "https://free.easychat.work" supports_stream = True supports_gpt_35_turbo = True - working = True + working = False @staticmethod def create_completion( diff --git a/g4f/Provider/Equing.py b/g4f/Provider/Equing.py index 0ebb93a5..261c53c0 100644 --- a/g4f/Provider/Equing.py +++ b/g4f/Provider/Equing.py @@ -6,12 +6,12 @@ from abc import ABC, abstractmethod import requests from ..typing import Any, CreateResult +from .base_provider import BaseProvider -class Equing(ABC): +class Equing(BaseProvider): url: str = 'https://next.eqing.tech/' - working = True - needs_auth = False + working = False supports_stream = True supports_gpt_35_turbo = True supports_gpt_4 = False diff --git a/g4f/Provider/GetGpt.py b/g4f/Provider/GetGpt.py index 74e772b0..b96efaac 100644 --- a/g4f/Provider/GetGpt.py +++ b/g4f/Provider/GetGpt.py @@ -14,7 +14,7 @@ from .base_provider import BaseProvider class GetGpt(BaseProvider): url = 'https://chat.getgpt.world/' supports_stream = True - working = True + working = False supports_gpt_35_turbo = True @staticmethod diff --git a/g4f/Provider/Liaobots.py b/g4f/Provider/Liaobots.py index 2cc5ff99..33224d2e 100644 --- a/g4f/Provider/Liaobots.py +++ b/g4f/Provider/Liaobots.py @@ -31,7 +31,7 @@ models = { class Liaobots(AsyncGeneratorProvider): url = "https://liaobots.com" - working = True + working = False supports_gpt_35_turbo = True supports_gpt_4 = True _auth_code = None diff --git a/g4f/Provider/Wuguokai.py b/g4f/Provider/Wuguokai.py index a9614626..0a46f6ee 100644 --- a/g4f/Provider/Wuguokai.py +++ b/g4f/Provider/Wuguokai.py @@ -5,13 +5,13 @@ import random import requests from ..typing import Any, CreateResult -from .base_provider import BaseProvider +from .base_provider import BaseProvider, format_prompt class Wuguokai(BaseProvider): url = 'https://chat.wuguokai.xyz' supports_gpt_35_turbo = True - working = True + working = False @staticmethod def create_completion( @@ -20,11 +20,6 @@ class Wuguokai(BaseProvider): stream: bool, **kwargs: Any, ) -> CreateResult: - base = '' - for message in messages: - base += '%s: %s\n' % (message['role'], message['content']) - base += 'assistant:' - headers = { 'authority': 'ai-api.wuguokai.xyz', 'accept': 'application/json, text/plain, */*', @@ -41,7 +36,7 @@ class Wuguokai(BaseProvider): 'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36' } data ={ - "prompt": base, + "prompt": format_prompt(messages), "options": {}, "userId": f"#/chat/{random.randint(1,99999999)}", "usingContext": True diff --git a/g4f/__init__.py b/g4f/__init__.py index 065acee6..90b05c85 100644 --- a/g4f/__init__.py +++ b/g4f/__init__.py @@ -2,6 +2,7 @@ from __future__ import annotations from . import models from .Provider import BaseProvider from .typing import Any, CreateResult, Union +import random logging = False @@ -13,14 +14,22 @@ class ChatCompletion: provider : Union[type[BaseProvider], None] = None, stream : bool = False, auth : Union[str, None] = None, **kwargs: Any) -> Union[CreateResult, str]: - + if isinstance(model, str): try: model = models.ModelUtils.convert[model] except KeyError: raise Exception(f'The model: {model} does not exist') - provider = model.best_provider if provider == None else provider + + if not provider: + if isinstance(model.best_provider, tuple): + provider = random.choice(model.best_provider) + else: + provider = model.best_provider + + if not provider: + raise Exception(f'No provider found') if not provider.working: raise Exception(f'{provider.__name__} is not working') diff --git a/g4f/models.py b/g4f/models.py index 0c5eb961..e095ce7e 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -1,31 +1,37 @@ from __future__ import annotations from dataclasses import dataclass -from .Provider import Bard, BaseProvider, GetGpt, H2o, Liaobots, Vercel, Equing +from .Provider import BaseProvider, Bard, H2o, Vercel +from .Provider import Aichat, Aivvm, ChatBase, ChatgptAi, ChatgptLogin, CodeLinkAva +from .Provider import DeepAi, Vercel, Vitalentum, Ylokh, You, Yqcloud +from .typing import Union @dataclass class Model: name: str base_provider: str - best_provider: type[BaseProvider] + best_provider: Union[type[BaseProvider], tuple[type[BaseProvider]]] = None # Config for HuggingChat, OpenAssistant # Works for Liaobots, H2o, OpenaiChat, Yqcloud, You default = Model( name="", - base_provider="huggingface", - best_provider=H2o, + base_provider="huggingface" ) # GPT-3.5 / GPT-4 gpt_35_turbo = Model( name = 'gpt-3.5-turbo', base_provider = 'openai', - best_provider = GetGpt) + best_provider = ( + Vercel, Aichat, Aivvm, ChatBase, ChatgptAi, ChatgptLogin, + CodeLinkAva, DeepAi, Vitalentum, Ylokh, You, Yqcloud + ) +) gpt_4 = Model( name = 'gpt-4', base_provider = 'openai', - best_provider = Liaobots) +) # Bard palm = Model( @@ -117,8 +123,7 @@ gpt_35_turbo_16k = Model( gpt_35_turbo_16k_0613 = Model( name = 'openai:gpt-3.5-turbo-16k-0613', - base_provider = 'openai', - best_provider = Equing) + base_provider = 'openai') gpt_4_0613 = Model( name = 'openai:gpt-4-0613', -- cgit v1.2.3