diff options
author | kqlio67 <kqlio67@users.noreply.github.com> | 2024-09-26 23:24:44 +0200 |
---|---|---|
committer | kqlio67 <kqlio67@users.noreply.github.com> | 2024-09-26 23:24:44 +0200 |
commit | cfe5acc15221806e9bc1155ef5d5f9e760f195a7 (patch) | |
tree | 6f79507646dd5d9afb8ea8f620ebbce7fca34059 /g4f | |
parent | feat(g4f/Provider/AIChatFree.py): add AIChatFree provider (diff) | |
download | gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.gz gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.bz2 gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.lz gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.xz gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.zst gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.zip |
Diffstat (limited to 'g4f')
-rw-r--r-- | g4f/Provider/DeepInfraChat.py | 6 | ||||
-rw-r--r-- | g4f/Provider/Liaobots.py | 60 | ||||
-rw-r--r-- | g4f/Provider/__init__.py | 1 | ||||
-rw-r--r-- | g4f/models.py | 25 |
4 files changed, 27 insertions, 65 deletions
diff --git a/g4f/Provider/DeepInfraChat.py b/g4f/Provider/DeepInfraChat.py index fa6dd2fe..b8cc6ab8 100644 --- a/g4f/Provider/DeepInfraChat.py +++ b/g4f/Provider/DeepInfraChat.py @@ -39,7 +39,7 @@ class DeepInfraChat(AsyncGeneratorProvider, ProviderModelMixin): model_aliases = { "llama-3.1-405b": "meta-llama/Meta-Llama-3.1-405B-Instruct", "llama-3.1-70b": "meta-llama/Meta-Llama-3.1-70B-Instruct", - "Llama-3.1-8B": "meta-llama/Meta-Llama-3.1-8B-Instruct", + "llama-3.1-8B": "meta-llama/Meta-Llama-3.1-8B-Instruct", "mixtral-8x22b": "mistralai/Mixtral-8x22B-Instruct-v0.1", "mixtral-8x7b": "mistralai/Mixtral-8x7B-Instruct-v0.1", "wizardlm-2-8x22b": "microsoft/WizardLM-2-8x22B", @@ -47,9 +47,9 @@ class DeepInfraChat(AsyncGeneratorProvider, ProviderModelMixin): "qwen-2-72b": "Qwen/Qwen2-72B-Instruct", "phi-3-medium-4k": "microsoft/Phi-3-medium-4k-instruct", "gemma-2b-27b": "google/gemma-2-27b-it", - "minicpm-llama-3-v2.5": "openbmb/MiniCPM-Llama3-V-2_5", + "minicpm-llama-3-v2.5": "openbmb/MiniCPM-Llama3-V-2_5", # Image upload is available "mistral-7b": "mistralai/Mistral-7B-Instruct-v0.3", - "lzlv_70b": "lizpreciatior/lzlv_70b_fp16_hf", + "lzlv-70b": "lizpreciatior/lzlv_70b_fp16_hf", "openchat-3.6-8b": "openchat/openchat-3.6-8b", "phind-codellama-34b-v2": "Phind/Phind-CodeLlama-34B-v2", "dolphin-2.9.1-llama-3-70b": "cognitivecomputations/dolphin-2.9.1-llama-3-70b", diff --git a/g4f/Provider/Liaobots.py b/g4f/Provider/Liaobots.py index 598da5ea..b292020e 100644 --- a/g4f/Provider/Liaobots.py +++ b/g4f/Provider/Liaobots.py @@ -36,36 +36,18 @@ models = { "tokenLimit": 7800, "context": "8K", }, - "o1-preview": { - "id": "o1-preview", - "name": "o1-preview", - "model": "o1", - "provider": "OpenAI", - "maxLength": 400000, - "tokenLimit": 100000, - "context": "128K", - }, - "o1-mini": { - "id": "o1-mini", - "name": "o1-mini", - "model": "o1", - "provider": "OpenAI", - "maxLength": 400000, - "tokenLimit": 100000, - "context": "128K", - }, - "gpt-4-turbo-2024-04-09": { - "id": "gpt-4-turbo-2024-04-09", - "name": "GPT-4-Turbo", + "gpt-4o-2024-08-06": { + "id": "gpt-4o-2024-08-06", + "name": "GPT-4o", "model": "ChatGPT", "provider": "OpenAI", "maxLength": 260000, "tokenLimit": 126000, "context": "128K", }, - "gpt-4o-2024-08-06": { - "id": "gpt-4o-2024-08-06", - "name": "GPT-4o", + "gpt-4-turbo-2024-04-09": { + "id": "gpt-4-turbo-2024-04-09", + "name": "GPT-4-Turbo", "model": "ChatGPT", "provider": "OpenAI", "maxLength": 260000, @@ -117,18 +99,18 @@ models = { "tokenLimit": 200000, "context": "200K", }, - "claude-3-sonnet-20240229": { - "id": "claude-3-sonnet-20240229", - "name": "Claude-3-Sonnet", + "claude-3-5-sonnet-20240620": { + "id": "claude-3-5-sonnet-20240620", + "name": "Claude-3.5-Sonnet", "model": "Claude", "provider": "Anthropic", "maxLength": 800000, "tokenLimit": 200000, "context": "200K", }, - "claude-3-5-sonnet-20240620": { - "id": "claude-3-5-sonnet-20240620", - "name": "Claude-3.5-Sonnet", + "claude-3-sonnet-20240229": { + "id": "claude-3-sonnet-20240229", + "name": "Claude-3-Sonnet", "model": "Claude", "provider": "Anthropic", "maxLength": 800000, @@ -153,8 +135,8 @@ models = { "tokenLimit": 200000, "context": "200K", }, - "gemini-1.5-flash-exp-0827": { - "id": "gemini-1.5-flash-exp-0827", + "gemini-1.5-flash-002": { + "id": "gemini-1.5-flash-002", "name": "Gemini-1.5-Flash-1M", "model": "Gemini", "provider": "Google", @@ -162,8 +144,8 @@ models = { "tokenLimit": 1000000, "context": "1024K", }, - "gemini-1.5-pro-exp-0827": { - "id": "gemini-1.5-pro-exp-0827", + "gemini-1.5-pro-002": { + "id": "gemini-1.5-pro-002", "name": "Gemini-1.5-Pro-1M", "model": "Gemini", "provider": "Google", @@ -186,11 +168,9 @@ class Liaobots(AsyncGeneratorProvider, ProviderModelMixin): model_aliases = { "gpt-4o-mini": "gpt-4o-mini-free", "gpt-4o": "gpt-4o-free", - - "o1": "o1-preview", - - "gpt-4-turbo": "gpt-4-turbo-2024-04-09", "gpt-4o": "gpt-4o-2024-08-06", + + "gpt-4-turbo": "gpt-4-turbo-2024-04-09", "gpt-4": "gpt-4-0613", "claude-3-opus": "claude-3-opus-20240229", @@ -201,8 +181,8 @@ class Liaobots(AsyncGeneratorProvider, ProviderModelMixin): "claude-3-haiku": "claude-3-haiku-20240307", "claude-2.1": "claude-2.1", - "gemini-flash": "gemini-1.5-flash-exp-0827", - "gemini-pro": "gemini-1.5-pro-exp-0827", + "gemini-flash": "gemini-1.5-flash-002", + "gemini-pro": "gemini-1.5-pro-002", } _auth_code = "" diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index 515f5912..d088b5ee 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -10,6 +10,7 @@ from .selenium import * from .needs_auth import * from .AI365VIP import AI365VIP +from .AIChatFree import AIChatFree from .Allyfy import Allyfy from .AiChatOnline import AiChatOnline from .AiChats import AiChats diff --git a/g4f/models.py b/g4f/models.py index 4d461ada..bac83b5b 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -4,6 +4,7 @@ from dataclasses import dataclass from .Provider import IterListProvider, ProviderType from .Provider import ( + AIChatFree, Airforce, Allyfy, Bing, @@ -146,22 +147,6 @@ gpt_4 = Model( ]) ) -# o1 -o1 = Model( - name = 'o1', - base_provider = 'OpenAI', - best_provider = IterListProvider([ - Liaobots - ]) -) - -o1_mini = Model( - name = 'o1-mini', - base_provider = 'OpenAI', - best_provider = IterListProvider([ - Liaobots - ]) -) ### GigaChat ### gigachat = Model( @@ -288,7 +273,7 @@ phi_3_5_mini = Model( gemini_pro = Model( name = 'gemini-pro', base_provider = 'Google DeepMind', - best_provider = IterListProvider([GeminiPro, LiteIcoding, Blackbox, Liaobots, Airforce]) + best_provider = IterListProvider([GeminiPro, LiteIcoding, Blackbox, AIChatFree, Liaobots, Airforce]) ) gemini_flash = Model( @@ -784,11 +769,7 @@ class ModelUtils: 'gpt-4o-mini': gpt_4o_mini, 'gpt-4': gpt_4, 'gpt-4-turbo': gpt_4_turbo, - -# o1 -'o1': o1, -'o1-mini': o1_mini, - + ### Meta ### "meta-ai": meta, |