From f6ef3cb2237d8c336e915ef77ddbe6f37934c4fd Mon Sep 17 00:00:00 2001 From: MIDORIBIN Date: Fri, 28 Jul 2023 19:07:17 +0900 Subject: refactor: refactor provider --- g4f/Provider/EasyChat.py | 83 ++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 83 insertions(+) create mode 100644 g4f/Provider/EasyChat.py (limited to 'g4f/Provider/EasyChat.py') diff --git a/g4f/Provider/EasyChat.py b/g4f/Provider/EasyChat.py new file mode 100644 index 00000000..59c46ffa --- /dev/null +++ b/g4f/Provider/EasyChat.py @@ -0,0 +1,83 @@ +import json + +import requests + +from ..typing import Any, CreateResult +from .base_provider import BaseProvider + + +class EasyChat(BaseProvider): + url = "https://free.easychat.work" + supports_stream = True + supports_gpt_35_turbo = True + + @staticmethod + def create_completion( + model: str, + messages: list[dict[str, str]], + stream: bool, + **kwargs: Any, + ) -> CreateResult: + active_servers = [ + "https://chat10.fastgpt.me", + "https://chat9.fastgpt.me", + "https://chat1.fastgpt.me", + "https://chat2.fastgpt.me", + "https://chat3.fastgpt.me", + "https://chat4.fastgpt.me", + ] + server = active_servers[kwargs.get("active_server", 0)] + headers = { + "authority": f"{server}".replace("https://", ""), + "accept": "text/event-stream", + "accept-language": "en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3,fa=0.2", + "content-type": "application/json", + "origin": f"{server}", + "referer": f"{server}/", + "sec-ch-ua": '"Not.A/Brand";v="8", "Chromium";v="114", "Google Chrome";v="114"', + "user-agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36", + "x-requested-with": "XMLHttpRequest", + } + + json_data = { + "messages": messages, + "stream": stream, + "model": model, + "temperature": kwargs.get("temperature", 0.5), + "presence_penalty": kwargs.get("presence_penalty", 0), + "frequency_penalty": kwargs.get("frequency_penalty", 0), + "top_p": kwargs.get("top_p", 1), + } + + session = requests.Session() + # init cookies from server + session.get(f"{server}/") + + response = session.post( + f"{server}/api/openai/v1/chat/completions", + headers=headers, + json=json_data, + ) + + response.raise_for_status() + print(response.text) + for chunk in response.iter_lines(): + if b"content" in chunk: + data = json.loads(chunk.decode().split("data: ")[1]) + yield data["choices"][0]["delta"]["content"] + + @classmethod + @property + def params(cls): + params = [ + ("model", "str"), + ("messages", "list[dict[str, str]]"), + ("stream", "bool"), + ("temperature", "float"), + ("presence_penalty", "int"), + ("frequency_penalty", "int"), + ("top_p", "int"), + ("active_server", "int"), + ] + param = ", ".join([": ".join(p) for p in params]) + return f"g4f.provider.{cls.__name__} supports: ({param})" -- cgit v1.2.3