summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--g4f/Provider/PerplexityLabs.py97
-rw-r--r--g4f/Provider/__init__.py5
2 files changed, 100 insertions, 2 deletions
diff --git a/g4f/Provider/PerplexityLabs.py b/g4f/Provider/PerplexityLabs.py
new file mode 100644
index 00000000..7e76aeef
--- /dev/null
+++ b/g4f/Provider/PerplexityLabs.py
@@ -0,0 +1,97 @@
+from __future__ import annotations
+
+import random
+import json
+from aiohttp import ClientSession, WSMsgType
+
+from ..typing import AsyncResult, Messages
+from .base_provider import AsyncGeneratorProvider
+
+API_URL = "https://labs-api.perplexity.ai/socket.io/"
+WS_URL = "wss://labs-api.perplexity.ai/socket.io/"
+MODELS = ['pplx-7b-online', 'pplx-70b-online', 'pplx-7b-chat', 'pplx-70b-chat', 'mistral-7b-instruct',
+ 'codellama-34b-instruct', 'llama-2-70b-chat', 'llava-7b-chat', 'mixtral-8x7b-instruct',
+ 'mistral-medium', 'related']
+DEFAULT_MODEL = MODELS[1]
+MODEL_MAP = {
+ "mistralai/Mistral-7B-Instruct-v0.1": "mistral-7b-instruct",
+ "meta-llama/Llama-2-70b-chat-hf": "llama-2-70b-chat",
+ "mistralai/Mixtral-8x7B-Instruct-v0.1": "mixtral-8x7b-instruct",
+}
+
+class PerplexityLabs(AsyncGeneratorProvider):
+ url = "https://labs.perplexity.ai"
+ working = True
+ supports_gpt_35_turbo = True
+
+ @classmethod
+ async def create_async_generator(
+ cls,
+ model: str,
+ messages: Messages,
+ proxy: str = None,
+ **kwargs
+ ) -> AsyncResult:
+ if not model:
+ model = DEFAULT_MODEL
+ elif model in MODEL_MAP:
+ model = MODEL_MAP[model]
+ elif model not in MODELS:
+ raise ValueError(f"Model is not supported: {model}")
+ headers = {
+ "User-Agent": "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:121.0) Gecko/20100101 Firefox/121.0",
+ "Accept": "*/*",
+ "Accept-Language": "de,en-US;q=0.7,en;q=0.3",
+ "Accept-Encoding": "gzip, deflate, br",
+ "Origin": cls.url,
+ "Connection": "keep-alive",
+ "Referer": f"{cls.url}/",
+ "Sec-Fetch-Dest": "empty",
+ "Sec-Fetch-Mode": "cors",
+ "Sec-Fetch-Site": "same-site",
+ "TE": "trailers",
+ }
+ async with ClientSession(headers=headers) as session:
+ t = format(random.getrandbits(32), '08x')
+ async with session.get(
+ f"{API_URL}?EIO=4&transport=polling&t={t}",
+ proxy=proxy
+ ) as response:
+ text = await response.text()
+
+ sid = json.loads(text[1:])['sid']
+ post_data = '40{"jwt":"anonymous-ask-user"}'
+ async with session.post(
+ f'{API_URL}?EIO=4&transport=polling&t={t}&sid={sid}',
+ data=post_data,
+ proxy=proxy
+ ) as response:
+ assert await response.text() == 'OK'
+
+ async with session.ws_connect(f'{WS_URL}?EIO=4&transport=websocket&sid={sid}', autoping=False) as ws:
+ await ws.send_str('2probe')
+ assert(await ws.receive_str() == '3probe')
+ await ws.send_str('5')
+ assert(await ws.receive_str())
+ assert(await ws.receive_str() == '6')
+ message_data = {
+ 'version': '2.2',
+ 'source': 'default',
+ 'model': model,
+ 'messages': messages
+ }
+ await ws.send_str('42' + json.dumps(['perplexity_playground', message_data]))
+ last_message = 0
+ while True:
+ message = await ws.receive_str()
+ if message == '2':
+ await ws.send_str('3')
+ continue
+ try:
+ data = json.loads(message[2:])[1]
+ yield data["output"][last_message:]
+ last_message = len(data["output"])
+ if data["final"]:
+ break
+ except:
+ raise RuntimeError(f"Message: {message}") \ No newline at end of file
diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py
index ee8d2c1b..5ac5cfca 100644
--- a/g4f/Provider/__init__.py
+++ b/g4f/Provider/__init__.py
@@ -9,10 +9,11 @@ from .needs_auth import *
from .unfinished import *
from .selenium import *
-from .Aura import Aura
from .AiAsk import AiAsk
from .AiChatOnline import AiChatOnline
from .AItianhu import AItianhu
+from .Aura import Aura
+from .Bestim import Bestim
from .Bing import Bing
from .ChatAnywhere import ChatAnywhere
from .ChatBase import ChatBase
@@ -45,12 +46,12 @@ from .Koala import Koala
from .Liaobots import Liaobots
from .Llama2 import Llama2
from .OnlineGpt import OnlineGpt
+from .PerplexityLabs import PerplexityLabs
from .Phind import Phind
from .Pi import Pi
from .Vercel import Vercel
from .Ylokh import Ylokh
from .You import You
-from .Bestim import Bestim
import sys