From 98895e5b09ea5e3f19b2263ddca028c7b296abb2 Mon Sep 17 00:00:00 2001 From: Heiner Lohaus Date: Sun, 1 Oct 2023 06:38:11 +0200 Subject: Update HuggingChat to new api Impersonate Aivvm Provider Add ChatForAi and FreeGpt Provider Update AItianhuSpace Provider Improve StreamRequest Support Update get_event_loop Helper --- g4f/Provider/HuggingChat.py | 62 ++++++++++++--------------------------------- 1 file changed, 16 insertions(+), 46 deletions(-) (limited to 'g4f/Provider/HuggingChat.py') diff --git a/g4f/Provider/HuggingChat.py b/g4f/Provider/HuggingChat.py index b2cf9793..f1cec775 100644 --- a/g4f/Provider/HuggingChat.py +++ b/g4f/Provider/HuggingChat.py @@ -1,6 +1,6 @@ from __future__ import annotations -import json +import json, uuid from aiohttp import ClientSession @@ -12,7 +12,7 @@ class HuggingChat(AsyncGeneratorProvider): url = "https://huggingface.co/chat" needs_auth = True working = True - model = "OpenAssistant/oasst-sft-6-llama-30b-xor" + model = "meta-llama/Llama-2-70b-chat-hf" @classmethod async def create_async_generator( @@ -37,55 +37,25 @@ class HuggingChat(AsyncGeneratorProvider): cookies=cookies, headers=headers ) as session: - async with session.post(f"{cls.url}/conversation", proxy=proxy, json={"model": model}) as response: + async with session.post(f"{cls.url}/conversation", json={"model": model}, proxy=proxy) as response: conversation_id = (await response.json())["conversationId"] send = { + "id": str(uuid.uuid4()), "inputs": format_prompt(messages), - "parameters": { - "temperature": 0.2, - "truncate": 1000, - "max_new_tokens": 1024, - "stop": [""], - "top_p": 0.95, - "repetition_penalty": 1.2, - "top_k": 50, - "return_full_text": False, - **kwargs - }, - "stream": stream, - "options": { - "id": "9e9b8bc4-6604-40c6-994e-8eb78fa32e37", - "response_id": "04ce2602-3bea-45e8-8efc-cef00680376a", - "is_retry": False, - "use_cache": False, - "web_search_id": "" - } + "is_retry": False, + "response_id": str(uuid.uuid4()), + "web_search": False } - async with session.post(f"{cls.url}/conversation/{conversation_id}", proxy=proxy, json=send) as response: - if not stream: - data = await response.json() - if "error" in data: - raise RuntimeError(data["error"]) - elif isinstance(data, list): - yield data[0]["generated_text"].strip() - else: - raise RuntimeError(f"Response: {data}") - else: - start = "data:" - first = True - async for line in response.content: - line = line.decode("utf-8") - if line.startswith(start): - line = json.loads(line[len(start):-1]) - if "token" not in line: - raise RuntimeError(f"Response: {line}") - if not line["token"]["special"]: - if first: - yield line["token"]["text"].lstrip() - first = False - else: - yield line["token"]["text"] + async with session.post(f"{cls.url}/conversation/{conversation_id}", json=send, proxy=proxy) as response: + async for line in response.content: + line = json.loads(line[:-1]) + if "type" not in line: + raise RuntimeError(f"Response: {line}") + elif line["type"] == "stream": + yield line["token"] + elif line["type"] == "finalAnswer": + break async with session.delete(f"{cls.url}/conversation/{conversation_id}", proxy=proxy) as response: response.raise_for_status() -- cgit v1.2.3 From 6a61cf811655fa87dbcb196025cc0b6040502293 Mon Sep 17 00:00:00 2001 From: Heiner Lohaus Date: Wed, 4 Oct 2023 07:20:51 +0200 Subject: Sort providers in new dirs --- g4f/Provider/HuggingChat.py | 74 --------------------------------------------- 1 file changed, 74 deletions(-) delete mode 100644 g4f/Provider/HuggingChat.py (limited to 'g4f/Provider/HuggingChat.py') diff --git a/g4f/Provider/HuggingChat.py b/g4f/Provider/HuggingChat.py deleted file mode 100644 index f1cec775..00000000 --- a/g4f/Provider/HuggingChat.py +++ /dev/null @@ -1,74 +0,0 @@ -from __future__ import annotations - -import json, uuid - -from aiohttp import ClientSession - -from ..typing import AsyncGenerator -from .base_provider import AsyncGeneratorProvider, format_prompt, get_cookies - - -class HuggingChat(AsyncGeneratorProvider): - url = "https://huggingface.co/chat" - needs_auth = True - working = True - model = "meta-llama/Llama-2-70b-chat-hf" - - @classmethod - async def create_async_generator( - cls, - model: str, - messages: list[dict[str, str]], - stream: bool = True, - proxy: str = None, - cookies: dict = None, - **kwargs - ) -> AsyncGenerator: - model = model if model else cls.model - if proxy and "://" not in proxy: - proxy = f"http://{proxy}" - if not cookies: - cookies = get_cookies(".huggingface.co") - - headers = { - 'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36', - } - async with ClientSession( - cookies=cookies, - headers=headers - ) as session: - async with session.post(f"{cls.url}/conversation", json={"model": model}, proxy=proxy) as response: - conversation_id = (await response.json())["conversationId"] - - send = { - "id": str(uuid.uuid4()), - "inputs": format_prompt(messages), - "is_retry": False, - "response_id": str(uuid.uuid4()), - "web_search": False - } - async with session.post(f"{cls.url}/conversation/{conversation_id}", json=send, proxy=proxy) as response: - async for line in response.content: - line = json.loads(line[:-1]) - if "type" not in line: - raise RuntimeError(f"Response: {line}") - elif line["type"] == "stream": - yield line["token"] - elif line["type"] == "finalAnswer": - break - - async with session.delete(f"{cls.url}/conversation/{conversation_id}", proxy=proxy) as response: - response.raise_for_status() - - - @classmethod - @property - def params(cls): - params = [ - ("model", "str"), - ("messages", "list[dict[str, str]]"), - ("stream", "bool"), - ("proxy", "str"), - ] - param = ", ".join([": ".join(p) for p in params]) - return f"g4f.provider.{cls.__name__} supports: ({param})" -- cgit v1.2.3