summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorkqlio67 <kqlio67@users.noreply.github.com>2024-09-02 10:55:51 +0200
committerkqlio67 <kqlio67@users.noreply.github.com>2024-09-02 10:55:51 +0200
commit3bece242045b885c71788f735cf77ac6187d2472 (patch)
treed986ab47b8ca641557489dee47e16fe65a434c97
parentUpdate g4f/Provider/ReplicateHome.py (diff)
downloadgpt4free-3bece242045b885c71788f735cf77ac6187d2472.tar
gpt4free-3bece242045b885c71788f735cf77ac6187d2472.tar.gz
gpt4free-3bece242045b885c71788f735cf77ac6187d2472.tar.bz2
gpt4free-3bece242045b885c71788f735cf77ac6187d2472.tar.lz
gpt4free-3bece242045b885c71788f735cf77ac6187d2472.tar.xz
gpt4free-3bece242045b885c71788f735cf77ac6187d2472.tar.zst
gpt4free-3bece242045b885c71788f735cf77ac6187d2472.zip
Diffstat (limited to '')
-rw-r--r--g4f/Provider/TeachAnything.py76
-rw-r--r--g4f/Provider/__init__.py6
-rw-r--r--g4f/models.py8
3 files changed, 84 insertions, 6 deletions
diff --git a/g4f/Provider/TeachAnything.py b/g4f/Provider/TeachAnything.py
new file mode 100644
index 00000000..3d34293f
--- /dev/null
+++ b/g4f/Provider/TeachAnything.py
@@ -0,0 +1,76 @@
+from __future__ import annotations
+
+from typing import Any, Dict
+
+from aiohttp import ClientSession, ClientTimeout
+
+from ..typing import AsyncResult, Messages
+from .base_provider import AsyncGeneratorProvider, ProviderModelMixin
+from .helper import format_prompt
+
+
+class TeachAnything(AsyncGeneratorProvider, ProviderModelMixin):
+ url = "https://www.teach-anything.com"
+ api_endpoint = "/api/generate"
+ working = True
+ default_model = "llama-3.1-70b"
+
+ @classmethod
+ async def create_async_generator(
+ cls,
+ model: str,
+ messages: Messages,
+ proxy: str | None = None,
+ **kwargs: Any
+ ) -> AsyncResult:
+ headers = cls._get_headers()
+
+ async with ClientSession(headers=headers) as session:
+ prompt = format_prompt(messages)
+ data = {"prompt": prompt}
+
+ timeout = ClientTimeout(total=60)
+
+ async with session.post(
+ f"{cls.url}{cls.api_endpoint}",
+ json=data,
+ proxy=proxy,
+ timeout=timeout
+ ) as response:
+ response.raise_for_status()
+ buffer = b""
+ async for chunk in response.content.iter_any():
+ buffer += chunk
+ try:
+ decoded = buffer.decode('utf-8')
+ yield decoded
+ buffer = b""
+ except UnicodeDecodeError:
+ # If we can't decode, we'll wait for more data
+ continue
+
+ # Handle any remaining data in the buffer
+ if buffer:
+ try:
+ yield buffer.decode('utf-8', errors='replace')
+ except Exception as e:
+ print(f"Error decoding final buffer: {e}")
+
+ @staticmethod
+ def _get_headers() -> Dict[str, str]:
+ return {
+ "accept": "*/*",
+ "accept-language": "en-US,en;q=0.9",
+ "content-type": "application/json",
+ "dnt": "1",
+ "origin": "https://www.teach-anything.com",
+ "priority": "u=1, i",
+ "referer": "https://www.teach-anything.com/",
+ "sec-ch-ua": '"Not/A)Brand";v="8", "Chromium";v="126"',
+ "sec-ch-ua-mobile": "?0",
+ "sec-ch-ua-platform": '"Linux"',
+ "sec-fetch-dest": "empty",
+ "sec-fetch-mode": "cors",
+ "sec-fetch-site": "same-origin",
+ "user-agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36"
+ }
diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py
index dc8acfbe..e5234685 100644
--- a/g4f/Provider/__init__.py
+++ b/g4f/Provider/__init__.py
@@ -17,8 +17,8 @@ from .Bing import Bing
from .BingCreateImages import BingCreateImages
from .Blackbox import Blackbox
from .ChatGot import ChatGot
-from .Chatgpt4o import Chatgpt4o
from .Chatgpt4Online import Chatgpt4Online
+from .Chatgpt4o import Chatgpt4o
from .ChatgptFree import ChatgptFree
from .DDG import DDG
from .DeepInfra import DeepInfra
@@ -33,7 +33,6 @@ from .GigaChat import GigaChat
from .GptTalkRu import GptTalkRu
from .HuggingChat import HuggingChat
from .HuggingFace import HuggingFace
-from .HuggingFace import HuggingFace
from .Koala import Koala
from .Liaobots import Liaobots
from .LiteIcoding import LiteIcoding
@@ -47,11 +46,12 @@ from .Ollama import Ollama
from .PerplexityLabs import PerplexityLabs
from .Pi import Pi
from .Pizzagpt import Pizzagpt
-from .Upstage import Upstage
from .Reka import Reka
from .Replicate import Replicate
from .ReplicateHome import ReplicateHome
from .Rocks import Rocks
+from .TeachAnything import TeachAnything
+from .Upstage import Upstage
from .Vercel import Vercel
from .WhiteRabbitNeo import WhiteRabbitNeo
from .You import You
diff --git a/g4f/models.py b/g4f/models.py
index 6ea2d2fd..4176f0a2 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -4,12 +4,13 @@ from dataclasses import dataclass
from .Provider import IterListProvider, ProviderType
from .Provider import (
+ AiChatOnline,
Allyfy,
Bing,
Blackbox,
ChatGot,
- Chatgpt4o,
Chatgpt4Online,
+ Chatgpt4o,
ChatgptFree,
DDG,
DeepInfra,
@@ -35,6 +36,7 @@ from .Provider import (
Reka,
Replicate,
ReplicateHome,
+ TeachAnything,
Upstage,
You,
)
@@ -106,7 +108,7 @@ gpt_4o_mini = Model(
name = 'gpt-4o-mini',
base_provider = 'openai',
best_provider = IterListProvider([
- DDG, Liaobots, You, FreeNetfly, MagickPenAsk, MagickPenChat, Pizzagpt, ChatgptFree, OpenaiChat, Koala,
+ DDG, Liaobots, You, FreeNetfly, MagickPenAsk, MagickPenChat, Pizzagpt, ChatgptFree, AiChatOnline, OpenaiChat, Koala,
])
)
@@ -163,7 +165,7 @@ llama_3_1_8b = Model(
llama_3_1_70b = Model(
name = "llama-3.1-70b",
base_provider = "meta",
- best_provider = IterListProvider([DDG, HuggingChat, FreeGpt, Blackbox, HuggingFace])
+ best_provider = IterListProvider([DDG, HuggingChat, FreeGpt, Blackbox, TeachAnything, HuggingFace])
)
llama_3_1_405b = Model(