diff options
-rw-r--r-- | docs/providers-and-models.md | 1 | ||||
-rw-r--r-- | g4f/Provider/AiMathGPT.py | 74 | ||||
-rw-r--r-- | g4f/Provider/__init__.py | 1 | ||||
-rw-r--r-- | g4f/models.py | 4 |
4 files changed, 1 insertions, 79 deletions
diff --git a/docs/providers-and-models.md b/docs/providers-and-models.md index 2a53cb22..17f47378 100644 --- a/docs/providers-and-models.md +++ b/docs/providers-and-models.md @@ -19,7 +19,6 @@ This document provides an overview of various AI providers and models, including |----------|-------------|--------------|---------------|--------|--------|------| |[ai4chat.co](https://www.ai4chat.co)|`g4f.Provider.Ai4Chat`|`gpt-4`|❌|❌|✔|![Active](https://img.shields.io/badge/Active-brightgreen)|❌| |[aichatfree.info](https://aichatfree.info)|`g4f.Provider.AIChatFree`|`gemini-pro`|❌|❌|✔|![Active](https://img.shields.io/badge/Active-brightgreen)|❌| -|[aimathgpt.forit.ai](https://aimathgpt.forit.ai)|`g4f.Provider.AiMathGPT`|`llama-3.1-70b`|❌|❌|✔|![Active](https://img.shields.io/badge/Active-brightgreen)|❌| |[api.airforce](https://api.airforce)|`g4f.Provider.Airforce`|`gpt-4o, gpt-4o-mini, gpt-4-turbo, llama-2-7b, llama-3.1-8b, llama-3.1-70b, hermes-2-pro, hermes-2-dpo, phi-2, deepseek-coder, openchat-3.5, openhermes-2.5, lfm-40b, german-7b, zephyr-7b, neural-7b`|`flux, flux-realism', flux-anime, flux-3d, flux-disney, flux-pixel, flux-4o, any-dark, sdxl`|❌|✔|![Active](https://img.shields.io/badge/Active-brightgreen)|❌| |[aiuncensored.info](https://www.aiuncensored.info)|`g4f.Provider.AIUncensored`|✔|✔|❌|✔|![Active](https://img.shields.io/badge/Active-brightgreen)|❌| |[allyfy.chat](https://allyfy.chat/)|`g4f.Provider.Allyfy`|`gpt-3.5-turbo`|❌|❌|✔|![Active](https://img.shields.io/badge/Active-brightgreen)|❌| diff --git a/g4f/Provider/AiMathGPT.py b/g4f/Provider/AiMathGPT.py deleted file mode 100644 index 90931691..00000000 --- a/g4f/Provider/AiMathGPT.py +++ /dev/null @@ -1,74 +0,0 @@ -from __future__ import annotations - -from aiohttp import ClientSession - -from ..typing import AsyncResult, Messages -from .base_provider import AsyncGeneratorProvider, ProviderModelMixin -from .helper import format_prompt - -class AiMathGPT(AsyncGeneratorProvider, ProviderModelMixin): - url = "https://aimathgpt.forit.ai" - api_endpoint = "https://aimathgpt.forit.ai/api/ai" - working = True - supports_stream = False - supports_system_message = True - supports_message_history = True - - default_model = 'llama3' - models = ['llama3'] - - model_aliases = {"llama-3.1-70b": "llama3",} - - @classmethod - def get_model(cls, model: str) -> str: - if model in cls.models: - return model - elif model in cls.model_aliases: - return cls.model_aliases[model] - else: - return cls.default_model - - @classmethod - async def create_async_generator( - cls, - model: str, - messages: Messages, - proxy: str = None, - **kwargs - ) -> AsyncResult: - model = cls.get_model(model) - - headers = { - 'accept': '*/*', - 'accept-language': 'en-US,en;q=0.9', - 'cache-control': 'no-cache', - 'content-type': 'application/json', - 'origin': cls.url, - 'pragma': 'no-cache', - 'priority': 'u=1, i', - 'referer': f'{cls.url}/', - 'sec-ch-ua': '"Chromium";v="129", "Not=A?Brand";v="8"', - 'sec-ch-ua-mobile': '?0', - 'sec-ch-ua-platform': '"Linux"', - 'sec-fetch-dest': 'empty', - 'sec-fetch-mode': 'cors', - 'sec-fetch-site': 'same-origin', - 'user-agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/129.0.0.0 Safari/537.36' - } - - async with ClientSession(headers=headers) as session: - data = { - "messages": [ - { - "role": "user", - "content": format_prompt(messages) - } - ], - "model": model - } - - async with session.post(cls.api_endpoint, json=data, proxy=proxy) as response: - response.raise_for_status() - response_data = await response.json() - filtered_response = response_data['result']['response'] - yield filtered_response diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index 368c4a25..60942581 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -13,7 +13,6 @@ from .local import * from .AIUncensored import AIUncensored from .Allyfy import Allyfy -from .AiMathGPT import AiMathGPT from .Airforce import Airforce from .Bing import Bing from .Blackbox import Blackbox diff --git a/g4f/models.py b/g4f/models.py index 87a076a8..8ce3688e 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -6,7 +6,6 @@ from .Provider import IterListProvider, ProviderType from .Provider import ( Ai4Chat, AIChatFree, - AiMathGPT, Airforce, AIUncensored, Allyfy, @@ -81,7 +80,6 @@ default = Model( ChatGptEs, ChatifyAI, Cloudflare, - AiMathGPT, AIUncensored, DarkAI, ]) @@ -184,7 +182,7 @@ llama_3_1_8b = Model( llama_3_1_70b = Model( name = "llama-3.1-70b", base_provider = "Meta Llama", - best_provider = IterListProvider([DDG, DeepInfraChat, Blackbox, TeachAnything, DarkAI, AiMathGPT, Airforce, RubiksAI, HuggingChat, HuggingFace, PerplexityLabs]) + best_provider = IterListProvider([DDG, DeepInfraChat, Blackbox, TeachAnything, DarkAI, Airforce, RubiksAI, HuggingChat, HuggingFace, PerplexityLabs]) ) llama_3_1_405b = Model( |