summaryrefslogtreecommitdiffstats
path: root/g4f/Provider/AiAsk.py
diff options
context:
space:
mode:
Diffstat (limited to 'g4f/Provider/AiAsk.py')
-rw-r--r--g4f/Provider/AiAsk.py43
1 files changed, 43 insertions, 0 deletions
diff --git a/g4f/Provider/AiAsk.py b/g4f/Provider/AiAsk.py
new file mode 100644
index 00000000..906b5575
--- /dev/null
+++ b/g4f/Provider/AiAsk.py
@@ -0,0 +1,43 @@
+from aiohttp import ClientSession
+from ..typing import AsyncGenerator
+from .base_provider import AsyncGeneratorProvider
+
+class AiAsk(AsyncGeneratorProvider):
+ url = "https://e.aiask.me"
+ supports_gpt_35_turbo = True
+ working = True
+
+ @classmethod
+ async def create_async_generator(
+ cls,
+ model: str,
+ messages: list[dict[str, str]],
+ timeout: int = 30,
+ **kwargs
+ ) -> AsyncGenerator:
+ headers = {
+ "accept": "application/json, text/plain, */*",
+ "origin": cls.url,
+ "referer": f"{cls.url}/chat",
+ }
+ async with ClientSession(headers=headers, timeout=timeout) as session:
+ data = {
+ "continuous": True,
+ "id": "fRMSQtuHl91A4De9cCvKD",
+ "list": messages,
+ "models": "0",
+ "prompt": "",
+ "temperature": kwargs.get("temperature", 0.5),
+ "title": "",
+ }
+ buffer = ""
+ rate_limit = "您的免费额度不够使用这个模型啦,请点击右上角登录继续使用!"
+ async with session.post(f"{cls.url}/v1/chat/gpt/", json=data) as response:
+ response.raise_for_status()
+ async for chunk in response.content.iter_any():
+ buffer += chunk.decode()
+ if not rate_limit.startswith(buffer):
+ yield buffer
+ buffer = ""
+ elif buffer == rate_limit:
+ raise RuntimeError("Rate limit reached") \ No newline at end of file