From 81cf5d7c771c4ff1a9d49db754ccef780f8b1d99 Mon Sep 17 00:00:00 2001
From: kqlio67 <166700875+kqlio67@users.noreply.github.com>
Date: Fri, 19 Apr 2024 12:10:56 +0000
Subject: A few small fixes in GUI and Providers (#1861)
---
g4f/Provider/Llama.py | 8 +++----
g4f/Provider/PerplexityLabs.py | 7 ++----
g4f/gui/client/index.html | 2 +-
g4f/models.py | 53 +++++++++++++++++++++++++++---------------
4 files changed, 41 insertions(+), 29 deletions(-)
(limited to 'g4f')
diff --git a/g4f/Provider/Llama.py b/g4f/Provider/Llama.py
index 4d19866e..8f3e9ea2 100644
--- a/g4f/Provider/Llama.py
+++ b/g4f/Provider/Llama.py
@@ -16,12 +16,12 @@ class Llama(AsyncGeneratorProvider, ProviderModelMixin):
"meta/llama-2-7b-chat",
"meta/llama-2-13b-chat",
"meta/llama-2-70b-chat",
- "meta/llama-3-8b-chat",
- "meta/llama-3-70b-chat",
+ "meta/meta-llama-3-8b-instruct",
+ "meta/meta-llama-3-70b-instruct",
]
model_aliases = {
- "meta-llama/Meta-Llama-3-8b": "meta/llama-3-8b-chat",
- "meta-llama/Meta-Llama-3-70b": "meta/llama-3-70b-chat",
+ "meta-llama/Meta-Llama-3-8b-instruct": "meta/meta-llama-3-8b-instruct",
+ "meta-llama/Meta-Llama-3-70b-instruct": "meta/meta-llama-3-70b-instruct",
"meta-llama/Llama-2-7b-chat-hf": "meta/llama-2-7b-chat",
"meta-llama/Llama-2-13b-chat-hf": "meta/llama-2-13b-chat",
"meta-llama/Llama-2-70b-chat-hf": "meta/llama-2-70b-chat",
diff --git a/g4f/Provider/PerplexityLabs.py b/g4f/Provider/PerplexityLabs.py
index ab36d284..b6fec53c 100644
--- a/g4f/Provider/PerplexityLabs.py
+++ b/g4f/Provider/PerplexityLabs.py
@@ -15,10 +15,7 @@ class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin):
working = True
default_model = "mixtral-8x7b-instruct"
models = [
- "sonar-small-online", "sonar-medium-online", "sonar-small-chat", "sonar-medium-chat", "mistral-7b-instruct",
- "codellama-70b-instruct", "llava-v1.5-7b-wrapper", "llava-v1.6-34b", "mixtral-8x7b-instruct",
- "gemma-2b-it", "gemma-7b-it"
- "mistral-medium", "related", "dbrx-instruct"
+ "sonar-small-online", "sonar-medium-online", "sonar-small-chat", "sonar-medium-chat", "dbrx-instruct", "claude-3-haiku-20240307", "llama-3-8b-instruct", "llama-3-70b-instruct", "codellama-70b-instruct", "mistral-7b-instruct", "llava-v1.5-7b-wrapper", "llava-v1.6-34b", "mixtral-8x7b-instruct", "mixtral-8x22b-instruct", "mistral-medium", "gemma-2b-it", "gemma-7b-it", "related"
]
model_aliases = {
"mistralai/Mistral-7B-Instruct-v0.1": "mistral-7b-instruct",
@@ -93,4 +90,4 @@ class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin):
if data["final"]:
break
except:
- raise RuntimeError(f"Message: {message}")
\ No newline at end of file
+ raise RuntimeError(f"Message: {message}")
diff --git a/g4f/gui/client/index.html b/g4f/gui/client/index.html
index 8668c21e..d84bbbe9 100644
--- a/g4f/gui/client/index.html
+++ b/g4f/gui/client/index.html
@@ -220,7 +220,7 @@
-
+
diff --git a/g4f/models.py b/g4f/models.py
index 4af1c31e..2b7c69f3 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -4,28 +4,33 @@ from dataclasses import dataclass
from .Provider import RetryProvider, ProviderType
from .Provider import (
+ Aichatos,
+ Bing,
+ Blackbox,
Chatgpt4Online,
- PerplexityLabs,
- GeminiProChat,
- ChatgptNext,
- HuggingChat,
- HuggingFace,
- OpenaiChat,
ChatgptAi,
+ ChatgptNext,
+ Cohere,
+ Cnote,
DeepInfra,
+ Feedough,
+ FreeGpt,
+ Gemini,
+ GeminiProChat,
GigaChat,
+ HuggingChat,
+ HuggingFace,
+ Koala,
Liaobots,
- FreeGpt,
Llama,
+ OpenaiChat,
+ PerplexityLabs,
+ Pi,
Vercel,
- Gemini,
- Koala,
- Cohere,
- Bing,
You,
- Pi,
)
+
@dataclass(unsafe_hash=True)
class Model:
"""
@@ -79,6 +84,9 @@ gpt_35_turbo = Model(
ChatgptNext,
Koala,
OpenaiChat,
+ Aichatos,
+ Cnote,
+ Feedough,
])
)
@@ -132,14 +140,14 @@ llama2_70b = Model(
best_provider = RetryProvider([Llama, DeepInfra, HuggingChat])
)
-llama3_8b = Model(
- name = "meta-llama/Meta-Llama-3-8b",
+llama3_8b_instruct = Model(
+ name = "meta-llama/Meta-Llama-3-8b-instruct",
base_provider = "meta",
best_provider = RetryProvider([Llama])
)
-llama3_70b = Model(
- name = "meta-llama/Meta-Llama-3-70b",
+llama3_70b_instruct = Model(
+ name = "meta-llama/Meta-Llama-3-70b-instruct",
base_provider = "meta",
best_provider = RetryProvider([Llama, HuggingChat])
)
@@ -291,6 +299,12 @@ command_r_plus = Model(
best_provider = RetryProvider([HuggingChat, Cohere])
)
+blackbox = Model(
+ name = 'blackbox',
+ base_provider = 'blackbox',
+ best_provider = Blackbox
+)
+
class ModelUtils:
"""
Utility class for mapping string identifiers to Model instances.
@@ -314,12 +328,12 @@ class ModelUtils:
'gpt-4-32k-0613' : gpt_4_32k_0613,
'gpt-4-turbo' : gpt_4_turbo,
- # Llama 2
+ # Llama
'llama2-7b' : llama2_7b,
'llama2-13b': llama2_13b,
'llama2-70b': llama2_70b,
- 'llama3-8b' : llama3_8b,
- 'llama3-70b': llama3_70b,
+ 'llama3-8b-instruct' : llama3_8b_instruct,
+ 'llama3-70b-instruct': llama3_70b_instruct,
'codellama-34b-instruct': codellama_34b_instruct,
'codellama-70b-instruct': codellama_70b_instruct,
@@ -345,6 +359,7 @@ class ModelUtils:
'claude-3-sonnet': claude_3_sonnet,
# other
+ 'blackbox': blackbox,
'command-r+': command_r_plus,
'dbrx-instruct': dbrx_instruct,
'lzlv-70b': lzlv_70b,
--
cgit v1.2.3