summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorHeiner Lohaus <hlohaus@users.noreply.github.com>2024-03-09 19:05:05 +0100
committerHeiner Lohaus <hlohaus@users.noreply.github.com>2024-03-09 19:05:05 +0100
commit74a33f10dd4bcac9e9f9716ffa27d3876fc976a0 (patch)
tree80ba1d0a2d2cdbe44b6ad83c378057772d92725f
parentAdd system message input to gui (diff)
downloadgpt4free-74a33f10dd4bcac9e9f9716ffa27d3876fc976a0.tar
gpt4free-74a33f10dd4bcac9e9f9716ffa27d3876fc976a0.tar.gz
gpt4free-74a33f10dd4bcac9e9f9716ffa27d3876fc976a0.tar.bz2
gpt4free-74a33f10dd4bcac9e9f9716ffa27d3876fc976a0.tar.lz
gpt4free-74a33f10dd4bcac9e9f9716ffa27d3876fc976a0.tar.xz
gpt4free-74a33f10dd4bcac9e9f9716ffa27d3876fc976a0.tar.zst
gpt4free-74a33f10dd4bcac9e9f9716ffa27d3876fc976a0.zip
-rw-r--r--g4f/Provider/GeminiPro.py2
-rw-r--r--g4f/Provider/bing/conversation.py2
-rw-r--r--g4f/gui/client/css/style.css22
-rw-r--r--g4f/gui/client/html/index.html5
-rw-r--r--g4f/gui/client/js/chat.v1.js210
5 files changed, 122 insertions, 119 deletions
diff --git a/g4f/Provider/GeminiPro.py b/g4f/Provider/GeminiPro.py
index a22304d5..4e31915a 100644
--- a/g4f/Provider/GeminiPro.py
+++ b/g4f/Provider/GeminiPro.py
@@ -32,7 +32,7 @@ class GeminiPro(AsyncGeneratorProvider, ProviderModelMixin):
connector: BaseConnector = None,
**kwargs
) -> AsyncResult:
- model = "gemini-pro-vision" if model is None and image is not None else model
+ model = "gemini-pro-vision" if not model and image is not None else model
model = cls.get_model(model)
if not api_key:
diff --git a/g4f/Provider/bing/conversation.py b/g4f/Provider/bing/conversation.py
index 919d9c1d..4af9e5fe 100644
--- a/g4f/Provider/bing/conversation.py
+++ b/g4f/Provider/bing/conversation.py
@@ -46,7 +46,7 @@ async def create_conversation(session: ClientSession, proxy: str = None) -> Conv
}
for k, v in headers.items():
session.headers[k] = v
- url = 'https://www.bing.com/turing/conversation/create?bundleVersion=1.1579.2'
+ url = 'https://www.bing.com/turing/conversation/create?bundleVersion=1.1626.1'
async with session.get(url, headers=headers, proxy=proxy) as response:
try:
data = await response.json()
diff --git a/g4f/gui/client/css/style.css b/g4f/gui/client/css/style.css
index 9e42b1e3..eb71fa72 100644
--- a/g4f/gui/client/css/style.css
+++ b/g4f/gui/client/css/style.css
@@ -205,6 +205,12 @@ body {
gap: 10px;
}
+.conversations .convo .choise {
+ position: absolute;
+ right: 8px;
+ background-color: var(--blur-bg);
+}
+
.conversations i {
color: var(--conversations);
cursor: pointer;
@@ -227,6 +233,10 @@ body {
padding-bottom: 0;
}
+.message.regenerate {
+ opacity: 0.75;
+}
+
.message:last-child {
animation: 0.6s show_message;
}
@@ -397,7 +407,7 @@ body {
padding: 6px var(--inner-gap);
}
-.stop_generating, .regenerate {
+.stop_generating, .toolbar .regenerate {
position: absolute;
z-index: 1000000;
top: 0;
@@ -405,14 +415,14 @@ body {
}
@media only screen and (min-width: 40em) {
- .stop_generating, .regenerate {
+ .stop_generating, .toolbar .regenerate {
left: 50%;
transform: translateX(-50%);
right: auto;
}
}
-.stop_generating button, .regenerate button{
+.stop_generating button, .toolbar .regenerate button{
backdrop-filter: blur(20px);
-webkit-backdrop-filter: blur(20px);
background-color: var(--blur-bg);
@@ -602,7 +612,6 @@ select {
.input-box {
display: flex;
align-items: center;
- padding-right: var(--inner-gap);
cursor: pointer;
}
@@ -991,8 +1000,9 @@ a:-webkit-any-link {
padding-right: 5px;
padding-top: 2px;
padding-bottom: 2px;
- top: 20px;
- left: 8px;
+ position: absolute;
+ bottom: 8px;
+ right: 8px;
}
#send-button:hover {
diff --git a/g4f/gui/client/html/index.html b/g4f/gui/client/html/index.html
index 5a9d1178..85192d23 100644
--- a/g4f/gui/client/html/index.html
+++ b/g4f/gui/client/html/index.html
@@ -37,10 +37,6 @@
import llamaTokenizer from "llama-tokenizer-js"
</script>
<script src="https://unpkg.com/gpt-tokenizer/dist/cl100k_base.js" async></script>
- <script type="module" async>
- import { countWords } from 'https://esm.run/alfaaz';
- window.countWords = countWords;
- </script>
<script>
const user_image = '<img src="/assets/img/user.png" alt="your avatar">';
const gpt_image = '<img src="/assets/img/gpt.png" alt="your avatar">';
@@ -55,7 +51,6 @@
}
#message-input {
- margin-right: 30px;
height: 82px;
margin-left: 20px;
}
diff --git a/g4f/gui/client/js/chat.v1.js b/g4f/gui/client/js/chat.v1.js
index 3d9945d9..8dd17275 100644
--- a/g4f/gui/client/js/chat.v1.js
+++ b/g4f/gui/client/js/chat.v1.js
@@ -1,13 +1,13 @@
const colorThemes = document.querySelectorAll('[name="theme"]');
const markdown = window.markdownit();
const message_box = document.getElementById(`messages`);
-const message_input = document.getElementById(`message-input`);
+const messageInput = document.getElementById(`message-input`);
const box_conversations = document.querySelector(`.top`);
const stop_generating = document.querySelector(`.stop_generating`);
const regenerate = document.querySelector(`.regenerate`);
const sidebar = document.querySelector(".conversations");
const sidebar_button = document.querySelector(".mobile-sidebar");
-const send_button = document.getElementById("send-button");
+const sendButton = document.getElementById("send-button");
const imageInput = document.getElementById("image");
const cameraInput = document.getElementById("camera");
const fileInput = document.getElementById("file");
@@ -19,11 +19,11 @@ let prompt_lock = false;
hljs.addPlugin(new CopyButtonPlugin());
-message_input.addEventListener("blur", () => {
+messageInput.addEventListener("blur", () => {
window.scrollTo(0, 0);
});
-message_input.addEventListener("focus", () => {
+messageInput.addEventListener("focus", () => {
document.documentElement.scrollTop = document.documentElement.scrollHeight;
});
@@ -60,7 +60,7 @@ const register_remove_message = async () => {
}
const message_el = el.parentElement.parentElement;
await remove_message(window.conversation_id, message_el.dataset.index);
- await load_conversation(window.conversation_id);
+ await load_conversation(window.conversation_id, false);
})
}
});
@@ -78,13 +78,13 @@ const delete_conversations = async () => {
};
const handle_ask = async () => {
- message_input.style.height = `82px`;
- message_input.focus();
+ messageInput.style.height = "82px";
+ messageInput.focus();
window.scrollTo(0, 0);
- message = message_input.value
+ message = messageInput.value
if (message.length > 0) {
- message_input.value = '';
+ messageInput.value = "";
prompt_lock = true;
count_input()
await add_conversation(window.conversation_id, message);
@@ -158,17 +158,19 @@ const prepare_messages = (messages, filter_last_message = true) => {
}
let new_messages = [];
- for (i in messages) {
- new_message = messages[i];
- // Remove generated images from history
- new_message.content = new_message.content.replaceAll(
- /<!-- generated images start -->[\s\S]+<!-- generated images end -->/gm,
- ""
- )
- delete new_message["provider"];
- // Remove regenerated messages
- if (!new_message.regenerate) {
- new_messages.push(new_message)
+ if (messages) {
+ for (i in messages) {
+ new_message = messages[i];
+ // Remove generated images from history
+ new_message.content = new_message.content.replaceAll(
+ /<!-- generated images start -->[\s\S]+<!-- generated images end -->/gm,
+ ""
+ )
+ delete new_message["provider"];
+ // Remove regenerated messages
+ if (!new_message.regenerate) {
+ new_messages.push(new_message)
+ }
}
}
@@ -263,11 +265,13 @@ const ask_gpt = async () => {
const { value, done } = await reader.read();
if (done) break;
for (const line of value.split("\n")) {
- if (!line) continue;
+ if (!line) {
+ continue;
+ }
const message = JSON.parse(line);
if (message.type == "content") {
text += message.content;
- } else if (message["type"] == "provider") {
+ } else if (message.type == "provider") {
provider_result = message.provider
content.querySelector('.provider').innerHTML = `
<a href="${provider_result.url}" target="_blank">
@@ -275,21 +279,21 @@ const ask_gpt = async () => {
</a>
${provider_result.model ? ' with ' + provider_result.model : ''}
`
- } else if (message["type"] == "error") {
- error = message["error"];
- } else if (message["type"] == "message") {
- console.error(message["message"])
+ } else if (message.type == "error") {
+ error = message.error;
+ } else if (messag.type == "message") {
+ console.error(messag.message)
}
}
if (error) {
console.error(error);
- content_inner.innerHTML += "<p>An error occured, please try again, if the problem persists, please use a other model or provider.</p>";
+ content_inner.innerHTML += `<p><strong>An error occured:</strong> ${error}</p>`;
} else {
html = markdown_render(text);
let lastElement, lastIndex = null;
- for (element of ['</p>', '</code></pre>', '</li>\n</ol>', '</li>\n</ul>']) {
+ for (element of ['</p>', '</code></pre>', '</p>\n</li>\n</ol>', '</li>\n</ol>', '</li>\n</ul>']) {
const index = html.lastIndexOf(element)
- if (index > lastIndex) {
+ if (index - element.length > lastIndex) {
lastElement = element;
lastIndex = index;
}
@@ -308,7 +312,6 @@ const ask_gpt = async () => {
}
}
if (!error) {
- // Remove cursor
html = markdown_render(text);
content_inner.innerHTML = html;
highlight(content_inner);
@@ -319,17 +322,16 @@ const ask_gpt = async () => {
}
} catch (e) {
console.error(e);
-
if (e.name != "AbortError") {
error = true;
text = "oops ! something went wrong, please try again / reload. [stacktrace in console]";
content_inner.innerHTML = text;
} else {
- content_inner.innerHTML += ` [aborted]`;
- text += ` [aborted]`
+ content_inner.innerHTML += " [aborted]";
+ if (text) text += " [aborted]";
}
}
- if (!error) {
+ if (!error && text) {
await add_message(window.conversation_id, "assistant", text, provider_result);
await load_conversation(window.conversation_id);
} else {
@@ -368,30 +370,22 @@ const clear_conversation = async () => {
while (messages.length > 0) {
message_box.removeChild(messages[0]);
}
-
- if (systemPrompt) {
- systemPrompt.value = "";
- }
};
const show_option = async (conversation_id) => {
const conv = document.getElementById(`conv-${conversation_id}`);
- const yes = document.getElementById(`yes-${conversation_id}`);
- const not = document.getElementById(`not-${conversation_id}`);
+ const choi = document.getElementById(`cho-${conversation_id}`);
- conv.style.display = `none`;
- yes.style.display = `block`;
- not.style.display = `block`;
+ conv.style.display = "none";
+ choi.style.display = "block";
};
const hide_option = async (conversation_id) => {
const conv = document.getElementById(`conv-${conversation_id}`);
- const yes = document.getElementById(`yes-${conversation_id}`);
- const not = document.getElementById(`not-${conversation_id}`);
+ const choi = document.getElementById(`cho-${conversation_id}`);
- conv.style.display = `block`;
- yes.style.display = `none`;
- not.style.display = `none`;
+ conv.style.display = "block";
+ choi.style.display = "none";
};
const delete_conversation = async (conversation_id) => {
@@ -422,12 +416,15 @@ const new_conversation = async () => {
window.conversation_id = uuid();
await clear_conversation();
+ if (systemPrompt) {
+ systemPrompt.value = "";
+ }
load_conversations();
hide_sidebar();
say_hello();
};
-const load_conversation = async (conversation_id) => {
+const load_conversation = async (conversation_id, scroll = true) => {
let conversation = await get_conversation(conversation_id);
let messages = conversation?.items || [];
@@ -451,7 +448,7 @@ const load_conversation = async (conversation_id) => {
</div>
` : "";
elements += `
- <div class="message" data-index="${i}">
+ <div class="message${item.regenerate ? " regenerate": ""}" data-index="${i}">
<div class="${item.role}">
${item.role == "assistant" ? gpt_image : user_image}
<i class="fa-solid fa-xmark"></i>
@@ -483,30 +480,14 @@ const load_conversation = async (conversation_id) => {
register_remove_message();
highlight(message_box);
- message_box.scrollTo({ top: message_box.scrollHeight, behavior: "smooth" });
-
- setTimeout(() => {
- message_box.scrollTop = message_box.scrollHeight;
- }, 500);
-};
+ if (scroll) {
+ message_box.scrollTo({ top: message_box.scrollHeight, behavior: "smooth" });
-function count_tokens(model, text) {
- if (model.startsWith("gpt-3") || model.startsWith("gpt-4")) {
- return GPTTokenizer_cl100k_base?.encode(text).length;
+ setTimeout(() => {
+ message_box.scrollTop = message_box.scrollHeight;
+ }, 500);
}
- if (model.startsWith("llama2") || model.startsWith("codellama")) {
- return llamaTokenizer?.encode(text).length;
- }
- if (model.startsWith("mistral") || model.startsWith("mixtral")) {
- return mistralTokenizer?.encode(text).length;
- }
-}
-
-function count_words_and_tokens(text, model) {
- const tokens_count = model ? count_tokens(model, text) : null;
- const tokens_append = tokens_count ? `, ${tokens_count} tokens` : "";
- return countWords ? `(${countWords(text)} words${tokens_append})` : "";
-}
+};
async function get_conversation(conversation_id) {
let conversation = await JSON.parse(
@@ -556,10 +537,12 @@ async function save_system_message() {
const hide_last_message = async (conversation_id) => {
const conversation = await get_conversation(conversation_id)
const last_message = conversation.items.pop();
- if (last_message["role"] == "assistant") {
- last_message["regenerate"] = true;
+ if (last_message !== null) {
+ if (last_message["role"] == "assistant") {
+ last_message["regenerate"] = true;
+ }
+ conversation.items.push(last_message);
}
- conversation.items.push(last_message);
await save_conversation(conversation_id, conversation);
};
@@ -568,7 +551,9 @@ const remove_message = async (conversation_id, index) => {
let new_items = [];
for (i in conversation.items) {
if (i == index - 1) {
- delete conversation.items[i]["regenerate"];
+ if (!conversation.items[index]?.regenerate) {
+ delete conversation.items[i]["regenerate"];
+ }
}
if (i != index) {
new_items.push(conversation.items[i])
@@ -609,8 +594,10 @@ const load_conversations = async () => {
<span class="convo-title">${conversation.title}</span>
</div>
<i onclick="show_option('${conversation.id}')" class="fa-regular fa-trash" id="conv-${conversation.id}"></i>
- <i onclick="delete_conversation('${conversation.id}')" class="fa-regular fa-check" id="yes-${conversation.id}" style="display:none;"></i>
- <i onclick="hide_option('${conversation.id}')" class="fa-regular fa-x" id="not-${conversation.id}" style="display:none;"></i>
+ <div id="cho-${conversation.id}" class="choise" style="display:none;">
+ <i onclick="delete_conversation('${conversation.id}')" class="fa-regular fa-check"></i>
+ <i onclick="hide_option('${conversation.id}')" class="fa-regular fa-x"></i>
+ </div>
</div>
`;
}
@@ -748,15 +735,45 @@ colorThemes.forEach((themeOption) => {
});
});
+function count_tokens(model, text) {
+ if (model) {
+ if (model.startsWith("llama2") || model.startsWith("codellama")) {
+ return llamaTokenizer?.encode(text).length;
+ }
+ if (model.startsWith("mistral") || model.startsWith("mixtral")) {
+ return mistralTokenizer?.encode(text).length;
+ }
+ }
+ return GPTTokenizer_cl100k_base?.encode(text).length;
+}
+
+function count_words(text) {
+ return text.trim().match(/[\w\u4E00-\u9FA5]+/gu)?.length || 0;
+}
+
+function count_words_and_tokens(text, model) {
+ return `(${count_words(text)} words, ${count_tokens(model, text)} tokens)`;
+}
+
+let countFocus = messageInput;
const count_input = async () => {
- if (message_input.value) {
+ if (countFocus.value) {
model = modelSelect.options[modelSelect.selectedIndex].value;
- inputCount.innerText = count_words_and_tokens(message_input.value, model);
+ inputCount.innerText = count_words_and_tokens(countFocus.value, model);
} else {
inputCount.innerHTML = "&nbsp;"
}
};
-message_input.addEventListener("keyup", count_input);
+messageInput.addEventListener("keyup", count_input);
+systemPrompt.addEventListener("keyup", count_input);
+systemPrompt.addEventListener("focus", function() {
+ countFocus = systemPrompt;
+ count_input();
+});
+systemPrompt.addEventListener("blur", function() {
+ countFocus = messageInput;
+ count_input();
+});
window.onload = async () => {
setTheme();
@@ -771,7 +788,7 @@ window.onload = async () => {
load_conversations();
- message_input.addEventListener("keydown", async (evt) => {
+ messageInput.addEventListener("keydown", async (evt) => {
if (prompt_lock) return;
if (evt.keyCode === 13 && !evt.shiftKey) {
@@ -779,41 +796,22 @@ window.onload = async () => {
console.log("pressed enter");
await handle_ask();
} else {
- message_input.style.removeProperty("height");
- message_input.style.height = message_input.scrollHeight + "px";
+ messageInput.style.removeProperty("height");
+ messageInput.style.height = messageInput.scrollHeight + "px";
}
});
- send_button.addEventListener(`click`, async () => {
+ sendButton.addEventListener(`click`, async () => {
console.log("clicked send");
if (prompt_lock) return;
await handle_ask();
});
+ messageInput.focus();
+
register_settings_localstorage();
};
-const observer = new MutationObserver((mutationsList) => {
- for (const mutation of mutationsList) {
- if (mutation.type === 'attributes' && mutation.attributeName === 'style') {
- const height = message_input.offsetHeight;
-
- let heightValues = {
- 81: "20px",
- 82: "20px",
- 100: "30px",
- 119: "39px",
- 138: "49px",
- 150: "55px"
- }
-
- send_button.style.top = heightValues[height] || '';
- }
- }
-});
-
-observer.observe(message_input, { attributes: true });
-
(async () => {
response = await fetch('/backend-api/v2/models')
models = await response.json()