summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorHeiner Lohaus <hlohaus@users.noreply.github.com>2024-01-13 16:18:31 +0100
committerHeiner Lohaus <hlohaus@users.noreply.github.com>2024-01-13 16:18:31 +0100
commitc190d0eae4c1fcc023035778e357e247d3b05259 (patch)
treee44d3bb05611a6e3fd86bd187636e845b6e8a331
parentAdded support for more models: FreeChatgpt (#1456) (diff)
downloadgpt4free-c190d0eae4c1fcc023035778e357e247d3b05259.tar
gpt4free-c190d0eae4c1fcc023035778e357e247d3b05259.tar.gz
gpt4free-c190d0eae4c1fcc023035778e357e247d3b05259.tar.bz2
gpt4free-c190d0eae4c1fcc023035778e357e247d3b05259.tar.lz
gpt4free-c190d0eae4c1fcc023035778e357e247d3b05259.tar.xz
gpt4free-c190d0eae4c1fcc023035778e357e247d3b05259.tar.zst
gpt4free-c190d0eae4c1fcc023035778e357e247d3b05259.zip
-rw-r--r--g4f/Provider/FreeChatgpt.py76
-rw-r--r--g4f/models.py12
2 files changed, 49 insertions, 39 deletions
diff --git a/g4f/Provider/FreeChatgpt.py b/g4f/Provider/FreeChatgpt.py
index 82fc3ef4..75514118 100644
--- a/g4f/Provider/FreeChatgpt.py
+++ b/g4f/Provider/FreeChatgpt.py
@@ -7,13 +7,9 @@ from ..typing import AsyncResult, Messages
from .base_provider import AsyncGeneratorProvider
-
models = {
- "claude-v1":"claude-2.1",
- "claude-v2":"claude-2.0",
- "gpt_35_turbo":"gpt-3.5-turbo-1106",
- "gpt-4":"gpt-4",
- "gemini-pro":"google-gemini-pro"
+ "claude-v2": "claude-2.0",
+ "gemini-pro": "google-gemini-pro"
}
class FreeChatgpt(AsyncGeneratorProvider):
@@ -31,31 +27,47 @@ class FreeChatgpt(AsyncGeneratorProvider):
proxy: str = None,
**kwargs
) -> AsyncResult:
- model = models[model] if model in models else "gpt-3.5-turbo-1106"
+ if model in models:
+ model = models[model]
+ elif not model:
+ model = "gpt-3.5-turbo"
headers = {
- "Accept": "application/json, text/event-stream",
- "Content-Type":"application/json",
- "Accept-Encoding": "gzip, deflate, br",
- "Accept-Language": "en-US,en;q=0.5",
- "Host":"free.chatgpt.org.uk",
- "Referer":f"{cls.url}/",
- "Origin":f"{cls.url}",
- "Sec-Fetch-Dest": "empty",
- "Sec-Fetch-Mode": "cors",
- "Sec-Fetch-Site": "same-origin",
- "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36",
-}
+ "Accept": "application/json, text/event-stream",
+ "Content-Type":"application/json",
+ "Accept-Encoding": "gzip, deflate, br",
+ "Accept-Language": "en-US,en;q=0.5",
+ "Host":"free.chatgpt.org.uk",
+ "Referer":f"{cls.url}/",
+ "Origin":f"{cls.url}",
+ "Sec-Fetch-Dest": "empty",
+ "Sec-Fetch-Mode": "cors",
+ "Sec-Fetch-Site": "same-origin",
+ "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36",
+ }
async with ClientSession(headers=headers) as session:
- data = {"messages":messages,"stream":True,"model":model,"temperature":0.5,"presence_penalty":0,"frequency_penalty":0,"top_p":1}
- async with session.post(f'{cls.url}/api/openai/v1/chat/completions',json=data) as result:
- async for chunk in result.content:
-
- line = chunk.decode()
- if line.startswith("data: [DONE]"):
- break
- elif line.startswith("data: "):
- line = json.loads(line[6:])
- if(line["choices"]==[]):
- continue
- if(line["choices"][0]["delta"].get("content") and line["choices"][0]["delta"]["content"]!=None):
- yield line["choices"][0]["delta"]["content"] \ No newline at end of file
+ data = {
+ "messages":messages,
+ "stream":True,
+ "model":model,
+ "temperature":0.5,
+ "presence_penalty":0,
+ "frequency_penalty":0,
+ "top_p":1,
+ **kwargs
+ }
+ async with session.post(f'{cls.url}/api/openai/v1/chat/completions', json=data, proxy=proxy) as response:
+ response.raise_for_status()
+ started = False
+ async for line in response.content:
+ if line.startswith(b"data: [DONE]"):
+ break
+ elif line.startswith(b"data: "):
+ line = json.loads(line[6:])
+ if(line["choices"]==[]):
+ continue
+ chunk = line["choices"][0]["delta"].get("content")
+ if chunk:
+ started = True
+ yield chunk
+ if not started:
+ raise RuntimeError("Empty response") \ No newline at end of file
diff --git a/g4f/models.py b/g4f/models.py
index 90d9f444..03deebf8 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -4,6 +4,7 @@ from .Provider import RetryProvider, ProviderType
from .Provider import (
Chatgpt4Online,
ChatgptDemoAi,
+ GeminiProChat,
ChatgptNext,
HuggingChat,
ChatgptDemo,
@@ -61,7 +62,6 @@ gpt_35_long = Model(
ChatgptNext,
ChatgptDemo,
Gpt6,
- FreeChatgpt,
])
)
@@ -70,7 +70,6 @@ gpt_35_turbo = Model(
name = 'gpt-3.5-turbo',
base_provider = 'openai',
best_provider=RetryProvider([
- FreeChatgpt,
GptGo, You,
GptForLove, ChatBase,
Chatgpt4Online,
@@ -81,7 +80,7 @@ gpt_4 = Model(
name = 'gpt-4',
base_provider = 'openai',
best_provider = RetryProvider([
- Bing, FreeChatgpt, Phind, Liaobots,
+ Bing, Phind, Liaobots,
])
)
@@ -159,12 +158,12 @@ claude_instant_v1 = Model(
claude_v1 = Model(
name = 'claude-v1',
base_provider = 'anthropic',
- best_provider = RetryProvider([FreeChatgpt,Vercel]))
+ best_provider = Vercel)
claude_v2 = Model(
name = 'claude-v2',
base_provider = 'anthropic',
- best_provider = RetryProvider([FreeChatgpt,Vercel]))
+ best_provider = RetryProvider([FreeChatgpt, Vercel]))
command_light_nightly = Model(
name = 'command-light-nightly',
@@ -246,11 +245,10 @@ gpt_4_32k_0613 = Model(
best_provider = gpt_4.best_provider
)
-#Gemini
gemini_pro = Model(
name = 'gemini-pro',
base_provider = 'google',
- best_provider = FreeChatgpt
+ best_provider = RetryProvider([FreeChatgpt, GeminiProChat])
)
text_ada_001 = Model(