summaryrefslogtreecommitdiffstats
path: root/interference
diff options
context:
space:
mode:
Diffstat (limited to 'interference')
-rw-r--r--interference/app.py126
1 files changed, 66 insertions, 60 deletions
diff --git a/interference/app.py b/interference/app.py
index eee67c57..c6c5130d 100644
--- a/interference/app.py
+++ b/interference/app.py
@@ -1,86 +1,92 @@
-import os
-import time
import json
import random
+import string
+import time
+from typing import Any
-from g4f import Model, ChatCompletion, Provider
-from flask import Flask, request, Response
+from flask import Flask, request
from flask_cors import CORS
+from g4f import ChatCompletion
+
app = Flask(__name__)
CORS(app)
-@app.route("/chat/completions", methods=['POST'])
+
+@app.route("/chat/completions", methods=["POST"])
def chat_completions():
- streaming = request.json.get('stream', False)
- model = request.json.get('model', 'gpt-3.5-turbo')
- messages = request.json.get('messages')
-
- response = ChatCompletion.create(model=model, stream=streaming,
- messages=messages)
-
- if not streaming:
- while 'curl_cffi.requests.errors.RequestsError' in response:
- response = ChatCompletion.create(model=model, stream=streaming,
- messages=messages)
+ model = request.get_json().get("model", "gpt-3.5-turbo")
+ stream = request.get_json().get("stream", False)
+ messages = request.get_json().get("messages")
- completion_timestamp = int(time.time())
- completion_id = ''.join(random.choices(
- 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789', k=28))
+ response = ChatCompletion.create(model=model, stream=stream, messages=messages)
+ completion_id = "".join(random.choices(string.ascii_letters + string.digits, k=28))
+ completion_timestamp = int(time.time())
+
+ if not stream:
return {
- 'id': 'chatcmpl-%s' % completion_id,
- 'object': 'chat.completion',
- 'created': completion_timestamp,
- 'model': model,
- 'usage': {
- 'prompt_tokens': None,
- 'completion_tokens': None,
- 'total_tokens': None
+ "id": f"chatcmpl-{completion_id}",
+ "object": "chat.completion",
+ "created": completion_timestamp,
+ "model": model,
+ "choices": [
+ {
+ "index": 0,
+ "message": {
+ "role": "assistant",
+ "content": response,
+ },
+ "finish_reason": "stop",
+ }
+ ],
+ "usage": {
+ "prompt_tokens": None,
+ "completion_tokens": None,
+ "total_tokens": None,
},
- 'choices': [{
- 'message': {
- 'role': 'assistant',
- 'content': response
- },
- 'finish_reason': 'stop',
- 'index': 0
- }]
}
- def stream():
- for token in response:
- completion_timestamp = int(time.time())
- completion_id = ''.join(random.choices(
- 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789', k=28))
-
+ def streaming():
+ for chunk in response:
completion_data = {
- 'id': f'chatcmpl-{completion_id}',
- 'object': 'chat.completion.chunk',
- 'created': completion_timestamp,
- 'model': 'gpt-3.5-turbo-0301',
- 'choices': [
+ "id": f"chatcmpl-{completion_id}",
+ "object": "chat.completion.chunk",
+ "created": completion_timestamp,
+ "model": model,
+ "choices": [
{
- 'delta': {
- 'content': token
+ "index": 0,
+ "delta": {
+ "content": chunk,
},
- 'index': 0,
- 'finish_reason': None
+ "finish_reason": None,
}
- ]
+ ],
}
- yield 'data: %s\n\n' % json.dumps(completion_data, separators=(',' ':'))
+ content = json.dumps(completion_data, separators=(",", ":"))
+ yield f"data: {content}\n\n"
time.sleep(0.1)
- return app.response_class(stream(), mimetype='text/event-stream')
+ end_completion_data: dict[str, Any] = {
+ "id": f"chatcmpl-{completion_id}",
+ "object": "chat.completion.chunk",
+ "created": completion_timestamp,
+ "model": model,
+ "choices": [
+ {
+ "index": 0,
+ "delta": {},
+ "finish_reason": "stop",
+ }
+ ],
+ }
+ content = json.dumps(end_completion_data, separators=(",", ":"))
+ yield f"data: {content}\n\n"
+ return app.response_class(streaming(), mimetype="text/event-stream")
-if __name__ == '__main__':
- config = {
- 'host': '0.0.0.0',
- 'port': 1337,
- 'debug': True
- }
- app.run(**config) \ No newline at end of file
+if __name__ == "__main__":
+ app.run(host="0.0.0.0", port=1337, debug=True)