diff options
author | kqlio67 <kqlio67@users.noreply.github.com> | 2024-09-26 23:24:44 +0200 |
---|---|---|
committer | kqlio67 <kqlio67@users.noreply.github.com> | 2024-09-26 23:24:44 +0200 |
commit | cfe5acc15221806e9bc1155ef5d5f9e760f195a7 (patch) | |
tree | 6f79507646dd5d9afb8ea8f620ebbce7fca34059 /g4f/Provider | |
parent | feat(g4f/Provider/AIChatFree.py): add AIChatFree provider (diff) | |
download | gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.gz gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.bz2 gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.lz gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.xz gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.tar.zst gpt4free-cfe5acc15221806e9bc1155ef5d5f9e760f195a7.zip |
Diffstat (limited to '')
-rw-r--r-- | g4f/Provider/DeepInfraChat.py | 6 | ||||
-rw-r--r-- | g4f/Provider/Liaobots.py | 60 | ||||
-rw-r--r-- | g4f/Provider/__init__.py | 1 |
3 files changed, 24 insertions, 43 deletions
diff --git a/g4f/Provider/DeepInfraChat.py b/g4f/Provider/DeepInfraChat.py index fa6dd2fe..b8cc6ab8 100644 --- a/g4f/Provider/DeepInfraChat.py +++ b/g4f/Provider/DeepInfraChat.py @@ -39,7 +39,7 @@ class DeepInfraChat(AsyncGeneratorProvider, ProviderModelMixin): model_aliases = { "llama-3.1-405b": "meta-llama/Meta-Llama-3.1-405B-Instruct", "llama-3.1-70b": "meta-llama/Meta-Llama-3.1-70B-Instruct", - "Llama-3.1-8B": "meta-llama/Meta-Llama-3.1-8B-Instruct", + "llama-3.1-8B": "meta-llama/Meta-Llama-3.1-8B-Instruct", "mixtral-8x22b": "mistralai/Mixtral-8x22B-Instruct-v0.1", "mixtral-8x7b": "mistralai/Mixtral-8x7B-Instruct-v0.1", "wizardlm-2-8x22b": "microsoft/WizardLM-2-8x22B", @@ -47,9 +47,9 @@ class DeepInfraChat(AsyncGeneratorProvider, ProviderModelMixin): "qwen-2-72b": "Qwen/Qwen2-72B-Instruct", "phi-3-medium-4k": "microsoft/Phi-3-medium-4k-instruct", "gemma-2b-27b": "google/gemma-2-27b-it", - "minicpm-llama-3-v2.5": "openbmb/MiniCPM-Llama3-V-2_5", + "minicpm-llama-3-v2.5": "openbmb/MiniCPM-Llama3-V-2_5", # Image upload is available "mistral-7b": "mistralai/Mistral-7B-Instruct-v0.3", - "lzlv_70b": "lizpreciatior/lzlv_70b_fp16_hf", + "lzlv-70b": "lizpreciatior/lzlv_70b_fp16_hf", "openchat-3.6-8b": "openchat/openchat-3.6-8b", "phind-codellama-34b-v2": "Phind/Phind-CodeLlama-34B-v2", "dolphin-2.9.1-llama-3-70b": "cognitivecomputations/dolphin-2.9.1-llama-3-70b", diff --git a/g4f/Provider/Liaobots.py b/g4f/Provider/Liaobots.py index 598da5ea..b292020e 100644 --- a/g4f/Provider/Liaobots.py +++ b/g4f/Provider/Liaobots.py @@ -36,36 +36,18 @@ models = { "tokenLimit": 7800, "context": "8K", }, - "o1-preview": { - "id": "o1-preview", - "name": "o1-preview", - "model": "o1", - "provider": "OpenAI", - "maxLength": 400000, - "tokenLimit": 100000, - "context": "128K", - }, - "o1-mini": { - "id": "o1-mini", - "name": "o1-mini", - "model": "o1", - "provider": "OpenAI", - "maxLength": 400000, - "tokenLimit": 100000, - "context": "128K", - }, - "gpt-4-turbo-2024-04-09": { - "id": "gpt-4-turbo-2024-04-09", - "name": "GPT-4-Turbo", + "gpt-4o-2024-08-06": { + "id": "gpt-4o-2024-08-06", + "name": "GPT-4o", "model": "ChatGPT", "provider": "OpenAI", "maxLength": 260000, "tokenLimit": 126000, "context": "128K", }, - "gpt-4o-2024-08-06": { - "id": "gpt-4o-2024-08-06", - "name": "GPT-4o", + "gpt-4-turbo-2024-04-09": { + "id": "gpt-4-turbo-2024-04-09", + "name": "GPT-4-Turbo", "model": "ChatGPT", "provider": "OpenAI", "maxLength": 260000, @@ -117,18 +99,18 @@ models = { "tokenLimit": 200000, "context": "200K", }, - "claude-3-sonnet-20240229": { - "id": "claude-3-sonnet-20240229", - "name": "Claude-3-Sonnet", + "claude-3-5-sonnet-20240620": { + "id": "claude-3-5-sonnet-20240620", + "name": "Claude-3.5-Sonnet", "model": "Claude", "provider": "Anthropic", "maxLength": 800000, "tokenLimit": 200000, "context": "200K", }, - "claude-3-5-sonnet-20240620": { - "id": "claude-3-5-sonnet-20240620", - "name": "Claude-3.5-Sonnet", + "claude-3-sonnet-20240229": { + "id": "claude-3-sonnet-20240229", + "name": "Claude-3-Sonnet", "model": "Claude", "provider": "Anthropic", "maxLength": 800000, @@ -153,8 +135,8 @@ models = { "tokenLimit": 200000, "context": "200K", }, - "gemini-1.5-flash-exp-0827": { - "id": "gemini-1.5-flash-exp-0827", + "gemini-1.5-flash-002": { + "id": "gemini-1.5-flash-002", "name": "Gemini-1.5-Flash-1M", "model": "Gemini", "provider": "Google", @@ -162,8 +144,8 @@ models = { "tokenLimit": 1000000, "context": "1024K", }, - "gemini-1.5-pro-exp-0827": { - "id": "gemini-1.5-pro-exp-0827", + "gemini-1.5-pro-002": { + "id": "gemini-1.5-pro-002", "name": "Gemini-1.5-Pro-1M", "model": "Gemini", "provider": "Google", @@ -186,11 +168,9 @@ class Liaobots(AsyncGeneratorProvider, ProviderModelMixin): model_aliases = { "gpt-4o-mini": "gpt-4o-mini-free", "gpt-4o": "gpt-4o-free", - - "o1": "o1-preview", - - "gpt-4-turbo": "gpt-4-turbo-2024-04-09", "gpt-4o": "gpt-4o-2024-08-06", + + "gpt-4-turbo": "gpt-4-turbo-2024-04-09", "gpt-4": "gpt-4-0613", "claude-3-opus": "claude-3-opus-20240229", @@ -201,8 +181,8 @@ class Liaobots(AsyncGeneratorProvider, ProviderModelMixin): "claude-3-haiku": "claude-3-haiku-20240307", "claude-2.1": "claude-2.1", - "gemini-flash": "gemini-1.5-flash-exp-0827", - "gemini-pro": "gemini-1.5-pro-exp-0827", + "gemini-flash": "gemini-1.5-flash-002", + "gemini-pro": "gemini-1.5-pro-002", } _auth_code = "" diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index 515f5912..d088b5ee 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -10,6 +10,7 @@ from .selenium import * from .needs_auth import * from .AI365VIP import AI365VIP +from .AIChatFree import AIChatFree from .Allyfy import Allyfy from .AiChatOnline import AiChatOnline from .AiChats import AiChats |