From 64b1583a03185956f06a36bea0072e05973241dc Mon Sep 17 00:00:00 2001 From: abc <98614666+xtekky@users.noreply.github.com> Date: Sun, 17 Sep 2023 22:37:14 +0100 Subject: ~ | improved stability with gpt-3.5-turbo improved stability with gpt-3.5-turbo --- g4f/models.py | 35 +++++++++++++++++++++-------------- 1 file changed, 21 insertions(+), 14 deletions(-) (limited to 'g4f') diff --git a/g4f/models.py b/g4f/models.py index e095ce7e..0e2680b6 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -1,9 +1,19 @@ -from __future__ import annotations +from __future__ import annotations from dataclasses import dataclass -from .Provider import BaseProvider, Bard, H2o, Vercel -from .Provider import Aichat, Aivvm, ChatBase, ChatgptAi, ChatgptLogin, CodeLinkAva -from .Provider import DeepAi, Vercel, Vitalentum, Ylokh, You, Yqcloud -from .typing import Union +from .typing import Union +from .Provider import BaseProvider +from .Provider import ( + ChatgptLogin, + CodeLinkAva, + ChatgptAi, + ChatBase, + Yqcloud, + Vercel, + DeepAi, + Aivvm, + Bard, + H2o +) @dataclass class Model: @@ -14,24 +24,21 @@ class Model: # Config for HuggingChat, OpenAssistant # Works for Liaobots, H2o, OpenaiChat, Yqcloud, You default = Model( - name="", - base_provider="huggingface" -) + name = "", + base_provider = "huggingface") # GPT-3.5 / GPT-4 gpt_35_turbo = Model( name = 'gpt-3.5-turbo', base_provider = 'openai', - best_provider = ( - Vercel, Aichat, Aivvm, ChatBase, ChatgptAi, ChatgptLogin, - CodeLinkAva, DeepAi, Vitalentum, Ylokh, You, Yqcloud - ) + best_provider = [ + Yqcloud, DeepAi, CodeLinkAva, ChatgptLogin, ChatgptAi, ChatBase, Aivvm + ] ) gpt_4 = Model( name = 'gpt-4', - base_provider = 'openai', -) + base_provider = 'openai') # Bard palm = Model( -- cgit v1.2.3