diff options
author | Tekky <98614666+xtekky@users.noreply.github.com> | 2024-11-15 11:58:58 +0100 |
---|---|---|
committer | Tekky <98614666+xtekky@users.noreply.github.com> | 2024-11-15 11:58:58 +0100 |
commit | 6ac545445a6abd80b5e664165442f3eccef8dd24 (patch) | |
tree | 4742a90026cb797a5f862c099523935e3659e83b | |
parent | Ai4Chat not working (diff) | |
download | gpt4free-6ac545445a6abd80b5e664165442f3eccef8dd24.tar gpt4free-6ac545445a6abd80b5e664165442f3eccef8dd24.tar.gz gpt4free-6ac545445a6abd80b5e664165442f3eccef8dd24.tar.bz2 gpt4free-6ac545445a6abd80b5e664165442f3eccef8dd24.tar.lz gpt4free-6ac545445a6abd80b5e664165442f3eccef8dd24.tar.xz gpt4free-6ac545445a6abd80b5e664165442f3eccef8dd24.tar.zst gpt4free-6ac545445a6abd80b5e664165442f3eccef8dd24.zip |
Diffstat (limited to '')
-rw-r--r-- | g4f/Provider/Pi.py | 1 | ||||
-rw-r--r-- | g4f/Provider/__init__.py | 1 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/Gemini.py | 1 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/Poe.py | 1 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/Raycast.py | 8 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/Theb.py | 1 | ||||
-rw-r--r-- | g4f/Provider/not_working/Allyfy.py (renamed from g4f/Provider/Allyfy.py) | 8 | ||||
-rw-r--r-- | g4f/models.py | 346 |
8 files changed, 188 insertions, 179 deletions
diff --git a/g4f/Provider/Pi.py b/g4f/Provider/Pi.py index 266647ba..68a7357f 100644 --- a/g4f/Provider/Pi.py +++ b/g4f/Provider/Pi.py @@ -12,6 +12,7 @@ class Pi(AbstractProvider): supports_stream = True _session = None default_model = "pi" + models = [default_model] @classmethod def create_completion( diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index dcf9c352..da0eacfe 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -12,7 +12,6 @@ from .not_working import * from .local import * from .AIUncensored import AIUncensored -from .Allyfy import Allyfy from .Airforce import Airforce from .Bing import Bing from .Blackbox import Blackbox diff --git a/g4f/Provider/needs_auth/Gemini.py b/g4f/Provider/needs_auth/Gemini.py index 8d741476..dad54c84 100644 --- a/g4f/Provider/needs_auth/Gemini.py +++ b/g4f/Provider/needs_auth/Gemini.py @@ -57,6 +57,7 @@ class Gemini(AsyncGeneratorProvider): default_model = 'gemini' image_models = ["gemini"] default_vision_model = "gemini" + models = ["gemini", "gemini-1.5-flash", "gemini-1.5-pro"] _cookies: Cookies = None _snlm0e: str = None _sid: str = None diff --git a/g4f/Provider/needs_auth/Poe.py b/g4f/Provider/needs_auth/Poe.py index 0c969d27..65fdbef9 100644 --- a/g4f/Provider/needs_auth/Poe.py +++ b/g4f/Provider/needs_auth/Poe.py @@ -26,6 +26,7 @@ class Poe(AbstractProvider): needs_auth = True supports_gpt_35_turbo = True supports_stream = True + models = models.keys() @classmethod def create_completion( diff --git a/g4f/Provider/needs_auth/Raycast.py b/g4f/Provider/needs_auth/Raycast.py index 07abeda3..b8ec5a97 100644 --- a/g4f/Provider/needs_auth/Raycast.py +++ b/g4f/Provider/needs_auth/Raycast.py @@ -16,6 +16,11 @@ class Raycast(AbstractProvider): needs_auth = True working = True + models = [ + "gpt-3.5-turbo", + "gpt-4" + ] + @staticmethod def create_completion( model: str, @@ -25,6 +30,9 @@ class Raycast(AbstractProvider): **kwargs, ) -> CreateResult: auth = kwargs.get('auth') + if not auth: + raise ValueError("Raycast needs an auth token, pass it with the `auth` parameter") + headers = { 'Accept': 'application/json', 'Accept-Language': 'en-US,en;q=0.9', diff --git a/g4f/Provider/needs_auth/Theb.py b/g4f/Provider/needs_auth/Theb.py index af690063..c7d7d58e 100644 --- a/g4f/Provider/needs_auth/Theb.py +++ b/g4f/Provider/needs_auth/Theb.py @@ -38,6 +38,7 @@ class Theb(AbstractProvider): supports_gpt_35_turbo = True supports_gpt_4 = True supports_stream = True + models = models.keys() @classmethod def create_completion( diff --git a/g4f/Provider/Allyfy.py b/g4f/Provider/not_working/Allyfy.py index 51f44963..a1c73499 100644 --- a/g4f/Provider/Allyfy.py +++ b/g4f/Provider/not_working/Allyfy.py @@ -4,15 +4,15 @@ import asyncio import json import uuid from aiohttp import ClientSession -from ..typing import AsyncResult, Messages -from .base_provider import AsyncGeneratorProvider, ProviderModelMixin -from .helper import format_prompt +from ...typing import AsyncResult, Messages +from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin +from ..helper import format_prompt class Allyfy(AsyncGeneratorProvider, ProviderModelMixin): url = "https://allyfy.chat" api_endpoint = "https://chatbot.allyfy.chat/api/v1/message/stream/super/chat" - working = True + working = False supports_stream = True supports_system_message = True supports_message_history = True diff --git a/g4f/models.py b/g4f/models.py index 9c6178be..0a7eed35 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -9,7 +9,6 @@ from .Provider import ( AIChatFree, Airforce, AIUncensored, - Allyfy, Bing, Blackbox, ChatGpt, @@ -599,203 +598,202 @@ class ModelUtils: """ convert: dict[str, Model] = { -############ -### Text ### -############ - -### OpenAI ### -# gpt-3.5 -'gpt-3.5-turbo': gpt_35_turbo, + ############ + ### Text ### + ############ + + ### OpenAI ### + # gpt-3.5 + 'gpt-3.5-turbo': gpt_35_turbo, -# gpt-4 -'gpt-4o': gpt_4o, -'gpt-4o-mini': gpt_4o_mini, -'gpt-4': gpt_4, -'gpt-4-turbo': gpt_4_turbo, + # gpt-4 + 'gpt-4o': gpt_4o, + 'gpt-4o-mini': gpt_4o_mini, + 'gpt-4': gpt_4, + 'gpt-4-turbo': gpt_4_turbo, -# o1 -'o1': o1, -'o1-mini': o1_mini, - - -### Meta ### -"meta-ai": meta, - -# llama-2 -'llama-2-7b': llama_2_7b, - -# llama-3 -'llama-3-8b': llama_3_8b, - -# llama-3.1 -'llama-3.1-8b': llama_3_1_8b, -'llama-3.1-70b': llama_3_1_70b, -'llama-3.1-405b': llama_3_1_405b, - -# llama-3.2 -'llama-3.2-1b': llama_3_2_1b, -'llama-3.2-11b': llama_3_2_11b, - - -### Mistral ### -'mistral-7b': mistral_7b, -'mixtral-8x7b': mixtral_8x7b, -'mistral-nemo': mistral_nemo, - - -### NousResearch ### -'hermes-2-pro': hermes_2_pro, -'hermes-2-dpo': hermes_2_dpo, -'hermes-3': hermes_3, + # o1 + 'o1': o1, + 'o1-mini': o1_mini, + + + ### Meta ### + "meta-ai": meta, + # llama-2 + 'llama-2-7b': llama_2_7b, + + # llama-3 + 'llama-3-8b': llama_3_8b, -### Microsoft ### -'phi-2': phi_2, -'phi-3.5-mini': phi_3_5_mini, + # llama-3.1 + 'llama-3.1-8b': llama_3_1_8b, + 'llama-3.1-70b': llama_3_1_70b, + 'llama-3.1-405b': llama_3_1_405b, + + # llama-3.2 + 'llama-3.2-1b': llama_3_2_1b, + 'llama-3.2-11b': llama_3_2_11b, + + ### Mistral ### + 'mistral-7b': mistral_7b, + 'mixtral-8x7b': mixtral_8x7b, + 'mistral-nemo': mistral_nemo, + + + ### NousResearch ### + 'hermes-2-pro': hermes_2_pro, + 'hermes-2-dpo': hermes_2_dpo, + 'hermes-3': hermes_3, + + + ### Microsoft ### + 'phi-2': phi_2, + 'phi-3.5-mini': phi_3_5_mini, + + + ### Google ### + # gemini + 'gemini': gemini, + 'gemini-pro': gemini_pro, + 'gemini-flash': gemini_flash, + + # gemma + 'gemma-2b': gemma_2b, -### Google ### -# gemini -'gemini': gemini, -'gemini-pro': gemini_pro, -'gemini-flash': gemini_flash, - -# gemma -'gemma-2b': gemma_2b, + ### Anthropic ### + 'claude-2.1': claude_2_1, + # claude 3 + 'claude-3-opus': claude_3_opus, + 'claude-3-sonnet': claude_3_sonnet, + 'claude-3-haiku': claude_3_haiku, -### Anthropic ### -'claude-2.1': claude_2_1, + # claude 3.5 + 'claude-3.5-sonnet': claude_3_5_sonnet, + + + ### Reka AI ### + 'reka-core': reka_core, + + + ### Blackbox AI ### + 'blackboxai': blackboxai, + 'blackboxai-pro': blackboxai_pro, + + + ### CohereForAI ### + 'command-r+': command_r_plus, + -# claude 3 -'claude-3-opus': claude_3_opus, -'claude-3-sonnet': claude_3_sonnet, -'claude-3-haiku': claude_3_haiku, + ### GigaChat ### + 'gigachat': gigachat, + + + + ### Qwen ### + # qwen 1.5 + 'qwen-1.5-7b': qwen_1_5_7b, -# claude 3.5 -'claude-3.5-sonnet': claude_3_5_sonnet, - - -### Reka AI ### -'reka-core': reka_core, - - -### Blackbox AI ### -'blackboxai': blackboxai, -'blackboxai-pro': blackboxai_pro, - - -### CohereForAI ### -'command-r+': command_r_plus, - + # qwen 2 + 'qwen-2-72b': qwen_2_72b, -### GigaChat ### -'gigachat': gigachat, - - - -### Qwen ### -# qwen 1.5 -'qwen-1.5-7b': qwen_1_5_7b, + # qwen 2.5 + 'qwen-2.5-coder-32b': qwen_2_5_coder_32b, + + + ### Upstage ### + 'solar-mini': solar_mini, + 'solar-pro': solar_pro, -# qwen 2 -'qwen-2-72b': qwen_2_72b, -# qwen 2.5 -'qwen-2.5-coder-32b': qwen_2_5_coder_32b, - - -### Upstage ### -'solar-mini': solar_mini, -'solar-pro': solar_pro, + ### Inflection ### + 'pi': pi, -### Inflection ### -'pi': pi, + ### DeepSeek ### + 'deepseek-coder': deepseek_coder, + + + ### Yorickvp ### + 'llava-13b': llava_13b, -### DeepSeek ### -'deepseek-coder': deepseek_coder, - - -### Yorickvp ### -'llava-13b': llava_13b, + ### WizardLM ### + 'wizardlm-2-8x22b': wizardlm_2_8x22b, + + + ### OpenChat ### + 'openchat-3.5': openchat_3_5, + + + ### x.ai ### + 'grok-2': grok_2, + 'grok-2-mini': grok_2_mini, + 'grok-beta': grok_beta, + + + ### Perplexity AI ### + 'sonar-online': sonar_online, + 'sonar-chat': sonar_chat, + + + ### TheBloke ### + 'german-7b': german_7b, + + + ### Nvidia ### + 'nemotron-70b': nemotron_70b, -### WizardLM ### -'wizardlm-2-8x22b': wizardlm_2_8x22b, + ### Teknium ### + 'openhermes-2.5': openhermes_2_5, - -### OpenChat ### -'openchat-3.5': openchat_3_5, - - -### x.ai ### -'grok-2': grok_2, -'grok-2-mini': grok_2_mini, -'grok-beta': grok_beta, - - -### Perplexity AI ### -'sonar-online': sonar_online, -'sonar-chat': sonar_chat, - - -### TheBloke ### -'german-7b': german_7b, - - -### Nvidia ### -'nemotron-70b': nemotron_70b, - - -### Teknium ### -'openhermes-2.5': openhermes_2_5, - - -### Liquid ### -'lfm-40b': lfm_40b, - - -### DiscoResearch ### -'german-7b': german_7b, - - -### HuggingFaceH4 ### -'zephyr-7b': zephyr_7b, - - -### Inferless ### -'neural-7b': neural_7b, - - - -############# -### Image ### -############# - -### Stability AI ### -'sdxl': sdxl, -'sd-3': sd_3, - - -### Playground ### -'playground-v2.5': playground_v2_5, + ### Liquid ### + 'lfm-40b': lfm_40b, + + + ### DiscoResearch ### + 'german-7b': german_7b, -### Flux AI ### -'flux': flux, -'flux-pro': flux_pro, -'flux-realism': flux_realism, -'flux-anime': flux_anime, -'flux-3d': flux_3d, -'flux-disney': flux_disney, -'flux-pixel': flux_pixel, -'flux-4o': flux_4o, + ### HuggingFaceH4 ### + 'zephyr-7b': zephyr_7b, -### Other ### -'any-dark': any_dark, + + ### Inferless ### + 'neural-7b': neural_7b, + + + + ############# + ### Image ### + ############# + + ### Stability AI ### + 'sdxl': sdxl, + 'sd-3': sd_3, + + + ### Playground ### + 'playground-v2.5': playground_v2_5, + + + ### Flux AI ### + 'flux': flux, + 'flux-pro': flux_pro, + 'flux-realism': flux_realism, + 'flux-anime': flux_anime, + 'flux-3d': flux_3d, + 'flux-disney': flux_disney, + 'flux-pixel': flux_pixel, + 'flux-4o': flux_4o, + + + ### Other ### + 'any-dark': any_dark, } _all_models = list(ModelUtils.convert.keys()) |