summaryrefslogtreecommitdiffstats
path: root/g4f
diff options
context:
space:
mode:
authorHeiner Lohaus <hlohaus@users.noreply.github.com>2024-01-23 01:47:55 +0100
committerHeiner Lohaus <hlohaus@users.noreply.github.com>2024-01-23 01:47:55 +0100
commit07c944ad0a90e10e9401d34a626bbc7878ffab2a (patch)
tree05207374083d6e751cd84107259f5bd1e4d1e971 /g4f
parentImprove CreateImagesBing (diff)
downloadgpt4free-07c944ad0a90e10e9401d34a626bbc7878ffab2a.tar
gpt4free-07c944ad0a90e10e9401d34a626bbc7878ffab2a.tar.gz
gpt4free-07c944ad0a90e10e9401d34a626bbc7878ffab2a.tar.bz2
gpt4free-07c944ad0a90e10e9401d34a626bbc7878ffab2a.tar.lz
gpt4free-07c944ad0a90e10e9401d34a626bbc7878ffab2a.tar.xz
gpt4free-07c944ad0a90e10e9401d34a626bbc7878ffab2a.tar.zst
gpt4free-07c944ad0a90e10e9401d34a626bbc7878ffab2a.zip
Diffstat (limited to 'g4f')
-rw-r--r--g4f/Provider/bing/upload_image.py36
-rw-r--r--g4f/gui/client/html/index.html4
-rw-r--r--g4f/gui/client/js/chat.v1.js8
-rw-r--r--g4f/gui/server/backend.py2
-rw-r--r--g4f/image.py16
-rw-r--r--g4f/models.py11
6 files changed, 51 insertions, 26 deletions
diff --git a/g4f/Provider/bing/upload_image.py b/g4f/Provider/bing/upload_image.py
index 4d70659f..bb5687a8 100644
--- a/g4f/Provider/bing/upload_image.py
+++ b/g4f/Provider/bing/upload_image.py
@@ -82,13 +82,16 @@ def build_image_upload_payload(image_bin: str, tone: str) -> Tuple[str, str]:
Tuple[str, str]: The data and boundary for the payload.
"""
boundary = "----WebKitFormBoundary" + ''.join(random.choices(string.ascii_letters + string.digits, k=16))
- data = f"--{boundary}\r\n" \
- f"Content-Disposition: form-data; name=\"knowledgeRequest\"\r\n\r\n" \
- f"{json.dumps(build_knowledge_request(tone), ensure_ascii=False)}\r\n" \
- f"--{boundary}\r\n" \
- f"Content-Disposition: form-data; name=\"imageBase64\"\r\n\r\n" \
- f"{image_bin}\r\n" \
- f"--{boundary}--\r\n"
+ data = f"""--{boundary}
+Content-Disposition: form-data; name="knowledgeRequest"
+
+{json.dumps(build_knowledge_request(tone), ensure_ascii=False)}
+--{boundary}
+Content-Disposition: form-data; name="imageBase64"
+
+{image_bin}
+--{boundary}--
+"""
return data, boundary
def build_knowledge_request(tone: str) -> dict:
@@ -102,14 +105,17 @@ def build_knowledge_request(tone: str) -> dict:
dict: The knowledge request payload.
"""
return {
- 'invokedSkills': ["ImageById"],
- 'subscriptionId': "Bing.Chat.Multimodal",
- 'invokedSkillsRequestData': {
- 'enableFaceBlur': True
- },
- 'convoData': {
- 'convoid': "",
- 'convotone': tone
+ "imageInfo": {},
+ "knowledgeRequest": {
+ 'invokedSkills': ["ImageById"],
+ 'subscriptionId': "Bing.Chat.Multimodal",
+ 'invokedSkillsRequestData': {
+ 'enableFaceBlur': True
+ },
+ 'convoData': {
+ 'convoid': "",
+ 'convotone': tone
+ }
}
}
diff --git a/g4f/gui/client/html/index.html b/g4f/gui/client/html/index.html
index 95489ba4..5edb55e8 100644
--- a/g4f/gui/client/html/index.html
+++ b/g4f/gui/client/html/index.html
@@ -115,11 +115,11 @@
<textarea id="message-input" placeholder="Ask a question" cols="30" rows="10"
style="white-space: pre-wrap;resize: none;"></textarea>
<label for="image" title="Works only with Bing and OpenaiChat">
- <input type="file" id="image" name="image" accept="image/png, image/gif, image/jpeg" required/>
+ <input type="file" id="image" name="image" accept="image/png, image/gif, image/jpeg, image/svg+xml" required/>
<i class="fa-regular fa-image"></i>
</label>
<label for="file">
- <input type="file" id="file" name="file" accept="text/plain, text/html, text/xml, application/json, text/javascript, .sh, .py, .php, .css, .yaml, .sql, .svg, .log, .csv, .twig, .md" required/>
+ <input type="file" id="file" name="file" accept="text/plain, text/html, text/xml, application/json, text/javascript, .sh, .py, .php, .css, .yaml, .sql, .log, .csv, .twig, .md" required/>
<i class="fa-solid fa-paperclip"></i>
</label>
<div id="send-button">
diff --git a/g4f/gui/client/js/chat.v1.js b/g4f/gui/client/js/chat.v1.js
index 8b9bc181..d62c040b 100644
--- a/g4f/gui/client/js/chat.v1.js
+++ b/g4f/gui/client/js/chat.v1.js
@@ -660,7 +660,13 @@ observer.observe(message_input, { attributes: true });
}
document.getElementById("version_text").innerHTML = text
})()
-
+imageInput.addEventListener('click', async (event) => {
+ imageInput.value = '';
+});
+fileInput.addEventListener('click', async (event) => {
+ fileInput.value = '';
+ delete fileInput.dataset.text;
+});
fileInput.addEventListener('change', async (event) => {
if (fileInput.files.length) {
type = fileInput.files[0].type;
diff --git a/g4f/gui/server/backend.py b/g4f/gui/server/backend.py
index d5c59ed1..9346f489 100644
--- a/g4f/gui/server/backend.py
+++ b/g4f/gui/server/backend.py
@@ -137,7 +137,7 @@ class Backend_Api:
if 'image' in request.files:
file = request.files['image']
if file.filename != '' and is_allowed_extension(file.filename):
- kwargs['image'] = to_image(file.stream)
+ kwargs['image'] = to_image(file.stream, file.filename.endswith('.svg'))
if 'json' in request.form:
json_data = json.loads(request.form['json'])
else:
diff --git a/g4f/image.py b/g4f/image.py
index 24ded915..61081ea1 100644
--- a/g4f/image.py
+++ b/g4f/image.py
@@ -4,9 +4,9 @@ import base64
from .typing import ImageType, Union
from PIL import Image
-ALLOWED_EXTENSIONS = {'png', 'jpg', 'jpeg', 'gif', 'webp'}
+ALLOWED_EXTENSIONS = {'png', 'jpg', 'jpeg', 'gif', 'webp', 'svg'}
-def to_image(image: ImageType) -> Image.Image:
+def to_image(image: ImageType, is_svg: bool = False) -> Image.Image:
"""
Converts the input image to a PIL Image object.
@@ -16,6 +16,16 @@ def to_image(image: ImageType) -> Image.Image:
Returns:
Image.Image: The converted PIL Image object.
"""
+ if is_svg:
+ try:
+ import cairosvg
+ except ImportError:
+ raise RuntimeError('Install "cairosvg" package for open svg images')
+ if not isinstance(image, bytes):
+ image = image.read()
+ buffer = BytesIO()
+ cairosvg.svg2png(image, write_to=buffer)
+ image = Image.open(buffer)
if isinstance(image, str):
is_data_uri_an_image(image)
image = extract_data_uri(image)
@@ -153,6 +163,8 @@ def to_base64(image: Image.Image, compression_rate: float) -> str:
str: The base64-encoded image.
"""
output_buffer = BytesIO()
+ if image.mode != "RGB":
+ image = image.convert('RGB')
image.save(output_buffer, format="JPEG", quality=int(compression_rate * 100))
return base64.b64encode(output_buffer.getvalue()).decode()
diff --git a/g4f/models.py b/g4f/models.py
index 40a72d77..636058d3 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -5,6 +5,7 @@ from .Provider import (
Chatgpt4Online,
ChatgptDemoAi,
GeminiProChat,
+ PerplexityAi,
ChatgptNext,
HuggingChat,
ChatgptDemo,
@@ -78,7 +79,7 @@ gpt_35_long = Model(
gpt_35_turbo = Model(
name = 'gpt-3.5-turbo',
base_provider = 'openai',
- best_provider=RetryProvider([
+ best_provider = RetryProvider([
GptGo, You,
GptForLove, ChatBase,
Chatgpt4Online,
@@ -114,20 +115,20 @@ llama2_13b = Model(
llama2_70b = Model(
name = "meta-llama/Llama-2-70b-chat-hf",
base_provider = "huggingface",
- best_provider = RetryProvider([Llama2, DeepInfra, HuggingChat])
+ best_provider = RetryProvider([Llama2, DeepInfra, HuggingChat, PerplexityAi])
)
# Mistal
mixtral_8x7b = Model(
name = "mistralai/Mixtral-8x7B-Instruct-v0.1",
base_provider = "huggingface",
- best_provider = RetryProvider([DeepInfra, HuggingChat])
+ best_provider = RetryProvider([DeepInfra, HuggingChat, PerplexityAi])
)
mistral_7b = Model(
name = "mistralai/Mistral-7B-Instruct-v0.1",
base_provider = "huggingface",
- best_provider = RetryProvider([DeepInfra, HuggingChat])
+ best_provider = RetryProvider([DeepInfra, HuggingChat, PerplexityAi])
)
# Dolphin
@@ -311,7 +312,7 @@ llama70b_v2_chat = Model(
pi = Model(
name = 'pi',
base_provider = 'inflection',
- best_provider=Pi
+ best_provider = Pi
)
class ModelUtils: