summaryrefslogtreecommitdiffstats
path: root/g4f/gui/server
diff options
context:
space:
mode:
Diffstat (limited to 'g4f/gui/server')
-rw-r--r--g4f/gui/server/api.py27
-rw-r--r--g4f/gui/server/backend.py2
2 files changed, 21 insertions, 8 deletions
diff --git a/g4f/gui/server/api.py b/g4f/gui/server/api.py
index df7b487d..966319e4 100644
--- a/g4f/gui/server/api.py
+++ b/g4f/gui/server/api.py
@@ -13,8 +13,12 @@ from g4f.errors import VersionNotFoundError
from g4f.Provider import ProviderType, __providers__, __map__
from g4f.providers.base_provider import ProviderModelMixin
from g4f.Provider.bing.create_images import patch_provider
+from g4f.Provider.Bing import Conversation
+
+conversations: dict[str, Conversation] = {}
+
+class Api():
-class Api():
def get_models(self) -> list[str]:
"""
Return a list of all models.
@@ -73,7 +77,8 @@ class Api():
def get_conversation(self, options: dict, **kwargs) -> Iterator:
window = webview.active_window()
for message in self._create_response_stream(
- self._prepare_conversation_kwargs(options, kwargs)
+ self._prepare_conversation_kwargs(options, kwargs),
+ options.get("conversation_id")
):
window.evaluate_js(f"this.add_message_chunk({json.dumps(message)})")
@@ -101,6 +106,10 @@ class Api():
from .internet import get_search_message
messages[-1]["content"] = get_search_message(messages[-1]["content"])
+ conversation_id = json_data.get("conversation_id")
+ if conversation_id and conversation_id in conversations:
+ kwargs["conversation"] = conversations[conversation_id]
+
model = json_data.get('model')
model = model if model else models.default
patch = patch_provider if json_data.get('patch_provider') else None
@@ -112,10 +121,11 @@ class Api():
"stream": True,
"ignore_stream": True,
"patch_provider": patch,
+ "return_conversation": True,
**kwargs
}
- def _create_response_stream(self, kwargs) -> Iterator:
+ def _create_response_stream(self, kwargs, conversation_id: str) -> Iterator:
"""
Creates and returns a streaming response for the conversation.
@@ -133,12 +143,15 @@ class Api():
for chunk in ChatCompletion.create(**kwargs):
if first:
first = False
- yield self._format_json('provider', get_last_provider(True))
- if isinstance(chunk, Exception):
+ yield self._format_json("provider", get_last_provider(True))
+ if isinstance(chunk, Conversation):
+ conversations[conversation_id] = chunk
+ yield self._format_json("conversation", conversation_id)
+ elif isinstance(chunk, Exception):
logging.exception(chunk)
- yield self._format_json('message', get_error_message(chunk))
+ yield self._format_json("message", get_error_message(chunk))
else:
- yield self._format_json('content', chunk)
+ yield self._format_json("content", chunk)
except Exception as e:
logging.exception(e)
yield self._format_json('error', get_error_message(e))
diff --git a/g4f/gui/server/backend.py b/g4f/gui/server/backend.py
index 8d388708..fb8404d4 100644
--- a/g4f/gui/server/backend.py
+++ b/g4f/gui/server/backend.py
@@ -85,7 +85,7 @@ class Backend_Api(Api):
kwargs = self._prepare_conversation_kwargs(json_data, kwargs)
return self.app.response_class(
- self._create_response_stream(kwargs),
+ self._create_response_stream(kwargs, json_data.get("conversation_id")),
mimetype='text/event-stream'
)