summaryrefslogtreecommitdiffstats
path: root/g4f/Provider/OpenAssistant.py
diff options
context:
space:
mode:
authorHeiner Lohaus <heiner@lohaus.eu>2023-10-04 07:20:51 +0200
committerHeiner Lohaus <heiner@lohaus.eu>2023-10-04 07:20:51 +0200
commit6a61cf811655fa87dbcb196025cc0b6040502293 (patch)
tree0e299a24d90095013854d04f9bf13617eebb8f6c /g4f/Provider/OpenAssistant.py
parentUse custom user dir (diff)
downloadgpt4free-6a61cf811655fa87dbcb196025cc0b6040502293.tar
gpt4free-6a61cf811655fa87dbcb196025cc0b6040502293.tar.gz
gpt4free-6a61cf811655fa87dbcb196025cc0b6040502293.tar.bz2
gpt4free-6a61cf811655fa87dbcb196025cc0b6040502293.tar.lz
gpt4free-6a61cf811655fa87dbcb196025cc0b6040502293.tar.xz
gpt4free-6a61cf811655fa87dbcb196025cc0b6040502293.tar.zst
gpt4free-6a61cf811655fa87dbcb196025cc0b6040502293.zip
Diffstat (limited to 'g4f/Provider/OpenAssistant.py')
-rw-r--r--g4f/Provider/OpenAssistant.py100
1 files changed, 0 insertions, 100 deletions
diff --git a/g4f/Provider/OpenAssistant.py b/g4f/Provider/OpenAssistant.py
deleted file mode 100644
index 1e9a0661..00000000
--- a/g4f/Provider/OpenAssistant.py
+++ /dev/null
@@ -1,100 +0,0 @@
-from __future__ import annotations
-
-import json
-
-from aiohttp import ClientSession
-
-from ..typing import Any, AsyncGenerator
-from .base_provider import AsyncGeneratorProvider, format_prompt, get_cookies
-
-
-class OpenAssistant(AsyncGeneratorProvider):
- url = "https://open-assistant.io/chat"
- needs_auth = True
- working = True
- model = "OA_SFT_Llama_30B_6"
-
- @classmethod
- async def create_async_generator(
- cls,
- model: str,
- messages: list[dict[str, str]],
- proxy: str = None,
- cookies: dict = None,
- **kwargs: Any
- ) -> AsyncGenerator:
- if not cookies:
- cookies = get_cookies("open-assistant.io")
-
- headers = {
- 'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36',
- }
- async with ClientSession(
- cookies=cookies,
- headers=headers
- ) as session:
- async with session.post("https://open-assistant.io/api/chat", proxy=proxy) as response:
- chat_id = (await response.json())["id"]
-
- data = {
- "chat_id": chat_id,
- "content": f"<s>[INST]\n{format_prompt(messages)}\n[/INST]",
- "parent_id": None
- }
- async with session.post("https://open-assistant.io/api/chat/prompter_message", proxy=proxy, json=data) as response:
- parent_id = (await response.json())["id"]
-
- data = {
- "chat_id": chat_id,
- "parent_id": parent_id,
- "model_config_name": model if model else cls.model,
- "sampling_parameters":{
- "top_k": 50,
- "top_p": None,
- "typical_p": None,
- "temperature": 0.35,
- "repetition_penalty": 1.1111111111111112,
- "max_new_tokens": 1024,
- **kwargs
- },
- "plugins":[]
- }
- async with session.post("https://open-assistant.io/api/chat/assistant_message", proxy=proxy, json=data) as response:
- data = await response.json()
- if "id" in data:
- message_id = data["id"]
- elif "message" in data:
- raise RuntimeError(data["message"])
- else:
- response.raise_for_status()
-
- params = {
- 'chat_id': chat_id,
- 'message_id': message_id,
- }
- async with session.post("https://open-assistant.io/api/chat/events", proxy=proxy, params=params) as response:
- start = "data: "
- async for line in response.content:
- line = line.decode("utf-8")
- if line and line.startswith(start):
- line = json.loads(line[len(start):])
- if line["event_type"] == "token":
- yield line["text"]
-
- params = {
- 'chat_id': chat_id,
- }
- async with session.delete("https://open-assistant.io/api/chat", proxy=proxy, params=params) as response:
- response.raise_for_status()
-
- @classmethod
- @property
- def params(cls):
- params = [
- ("model", "str"),
- ("messages", "list[dict[str, str]]"),
- ("stream", "bool"),
- ("proxy", "str"),
- ]
- param = ", ".join([": ".join(p) for p in params])
- return f"g4f.provider.{cls.__name__} supports: ({param})"