diff options
-rw-r--r-- | .gitignore | 18 | ||||
-rw-r--r-- | Docker/Dockerfile | 12 | ||||
-rw-r--r-- | Dockerfile | 18 | ||||
-rw-r--r-- | README.md | 86 | ||||
-rw-r--r-- | cocalc/__init__.py | 57 | ||||
-rw-r--r-- | forefront/README.md | 16 | ||||
-rw-r--r-- | forefront/__init__.py | 155 | ||||
-rw-r--r-- | forefront/typing.py | 36 | ||||
-rw-r--r-- | gpt4free/README.md | 116 | ||||
-rw-r--r-- | gpt4free/__init__.py | 64 | ||||
-rw-r--r-- | gpt4free/cocalc/__init__.py | 47 | ||||
-rw-r--r-- | gpt4free/cocalc/readme.md (renamed from cocalc/readme.md) | 5 | ||||
-rw-r--r-- | gpt4free/forefront/README.md | 16 | ||||
-rw-r--r-- | gpt4free/forefront/__init__.py | 192 | ||||
-rw-r--r-- | gpt4free/forefront/mail.py (renamed from forefront/mail.py) | 14 | ||||
-rw-r--r-- | gpt4free/forefront/typing.py | 26 | ||||
-rw-r--r-- | gpt4free/quora/README.md (renamed from quora/README.md) | 21 | ||||
-rw-r--r-- | gpt4free/quora/__init__.py (renamed from quora/__init__.py) | 154 | ||||
-rw-r--r-- | gpt4free/quora/api.py (renamed from quora/api.py) | 68 | ||||
-rw-r--r-- | gpt4free/quora/cookies.txt (renamed from quora/cookies.txt) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/AddHumanMessageMutation.graphql (renamed from quora/graphql/AddHumanMessageMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/AddMessageBreakMutation.graphql (renamed from quora/graphql/AddMessageBreakMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/AutoSubscriptionMutation.graphql (renamed from quora/graphql/AutoSubscriptionMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/BioFragment.graphql (renamed from quora/graphql/BioFragment.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ChatAddedSubscription.graphql (renamed from quora/graphql/ChatAddedSubscription.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ChatFragment.graphql (renamed from quora/graphql/ChatFragment.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ChatListPaginationQuery.graphql (renamed from quora/graphql/ChatListPaginationQuery.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ChatPaginationQuery.graphql (renamed from quora/graphql/ChatPaginationQuery.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ChatViewQuery.graphql (renamed from quora/graphql/ChatViewQuery.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/DeleteHumanMessagesMutation.graphql (renamed from quora/graphql/DeleteHumanMessagesMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/DeleteMessageMutation.graphql (renamed from quora/graphql/DeleteMessageMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/HandleFragment.graphql (renamed from quora/graphql/HandleFragment.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/LoginWithVerificationCodeMutation.graphql (renamed from quora/graphql/LoginWithVerificationCodeMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/MessageAddedSubscription.graphql (renamed from quora/graphql/MessageAddedSubscription.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/MessageDeletedSubscription.graphql (renamed from quora/graphql/MessageDeletedSubscription.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/MessageFragment.graphql (renamed from quora/graphql/MessageFragment.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/MessageRemoveVoteMutation.graphql (renamed from quora/graphql/MessageRemoveVoteMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/MessageSetVoteMutation.graphql (renamed from quora/graphql/MessageSetVoteMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/PoeBotCreateMutation.graphql (renamed from quora/graphql/PoeBotCreateMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/PoeBotEditMutation.graphql (renamed from quora/graphql/PoeBotEditMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/SendMessageMutation.graphql (renamed from quora/graphql/SendMessageMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/SendVerificationCodeForLoginMutation.graphql (renamed from quora/graphql/SendVerificationCodeForLoginMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ShareMessagesMutation.graphql (renamed from quora/graphql/ShareMessagesMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/SignupWithVerificationCodeMutation.graphql (renamed from quora/graphql/SignupWithVerificationCodeMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/StaleChatUpdateMutation.graphql (renamed from quora/graphql/StaleChatUpdateMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/SubscriptionsMutation.graphql (renamed from quora/graphql/SubscriptionsMutation.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/SummarizePlainPostQuery.graphql (renamed from quora/graphql/SummarizePlainPostQuery.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/SummarizeQuotePostQuery.graphql (renamed from quora/graphql/SummarizeQuotePostQuery.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/SummarizeSharePostQuery.graphql (renamed from quora/graphql/SummarizeSharePostQuery.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/UserSnippetFragment.graphql (renamed from quora/graphql/UserSnippetFragment.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ViewerInfoQuery.graphql (renamed from quora/graphql/ViewerInfoQuery.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ViewerStateFragment.graphql (renamed from quora/graphql/ViewerStateFragment.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/ViewerStateUpdatedSubscription.graphql (renamed from quora/graphql/ViewerStateUpdatedSubscription.graphql) | 0 | ||||
-rw-r--r-- | gpt4free/quora/graphql/__init__.py (renamed from quora/graphql/__init__.py) | 0 | ||||
-rw-r--r-- | gpt4free/quora/mail.py (renamed from quora/mail.py) | 4 | ||||
-rw-r--r-- | gpt4free/theb/README.md (renamed from theb/README.md) | 3 | ||||
-rw-r--r-- | gpt4free/theb/__init__.py (renamed from theb/__init__.py) | 24 | ||||
-rw-r--r-- | gpt4free/theb/theb_test.py (renamed from theb/theb_test.py) | 2 | ||||
-rw-r--r-- | gpt4free/you/README.md (renamed from you/README.md) | 3 | ||||
-rw-r--r-- | gpt4free/you/__init__.py (renamed from you/__init__.py) | 68 | ||||
-rw-r--r-- | gui/README.md | 67 | ||||
-rw-r--r-- | gui/__init__.py | 0 | ||||
-rw-r--r-- | gui/image1.png | bin | 0 -> 172284 bytes | |||
-rw-r--r-- | gui/image2.png | bin | 0 -> 343564 bytes | |||
-rw-r--r-- | gui/query_methods.py | 99 | ||||
-rw-r--r-- | gui/streamlit_app.py | 23 | ||||
-rw-r--r-- | gui/streamlit_chat_app.py | 113 | ||||
-rw-r--r-- | openaihosted/__init__.py | 62 | ||||
-rw-r--r-- | openaihosted/readme.md | 18 | ||||
-rw-r--r-- | poetry.lock | 702 | ||||
-rw-r--r-- | pyproject.toml | 28 | ||||
-rw-r--r-- | requirements.txt | 2 | ||||
-rw-r--r-- | test.py | 5 | ||||
-rw-r--r-- | testing/forefront_test.py | 7 | ||||
-rw-r--r-- | testing/poe_account_create_test.py | 75 | ||||
-rw-r--r-- | testing/poe_test.py | 8 | ||||
-rw-r--r-- | testing/quora_test_2.py | 13 | ||||
-rw-r--r-- | testing/sqlchat_test.py | 4 | ||||
-rw-r--r-- | testing/t3nsor_test.py | 4 | ||||
-rw-r--r-- | testing/test_main.py | 27 | ||||
-rw-r--r-- | testing/writesonic_test.py | 22 | ||||
-rw-r--r-- | testing/you_test.py | 6 | ||||
-rw-r--r-- | unfinished/chatpdf/__init__.py | 75 | ||||
-rw-r--r-- | unfinished/openai/README.md | 2 | ||||
-rw-r--r-- | unfinished/openai/__ini__.py | 87 | ||||
-rw-r--r-- | unfinished/vercelai/__init__.py | 41 | ||||
-rw-r--r-- | unfinished/vercelai/test.js | 33 | ||||
-rw-r--r-- | unfinished/vercelai/test.py | 67 | ||||
-rw-r--r-- | unfinished/vercelai/token.py | 0 | ||||
-rw-r--r-- | unfinished/vercelai/v2.py | 27 | ||||
-rw-r--r-- | unfinished/vercelai/vercelai_test.py | 5 | ||||
-rw-r--r-- | unfinished/writesonic/README.md | 53 | ||||
-rw-r--r-- | unfinished/writesonic/__init__.py | 163 |
93 files changed, 2071 insertions, 992 deletions
@@ -7,10 +7,26 @@ /dataSources/ /dataSources.local.xml +# Ignore local python virtual environment +venv/ + +# Ignore streamlit_chat_app.py conversations pickle +conversations.pkl +*.pkl + +# Ignore accounts created by api's +accounts.txt + .idea/ -*/__pycache__/ +**/__pycache__/ + +__pycache__/ *.log cookie.json + +*.pyc + +dist/ diff --git a/Docker/Dockerfile b/Docker/Dockerfile deleted file mode 100644 index a7bb7d0a..00000000 --- a/Docker/Dockerfile +++ /dev/null @@ -1,12 +0,0 @@ -FROM python:3.10-slim - -RUN apt-get update && apt-get install -y git - -RUN git clone https://github.com/xtekky/gpt4free.git -WORKDIR /gpt4free -RUN pip install --no-cache-dir -r requirements.txt -RUN cp gui/streamlit_app.py . - -EXPOSE 8501 - -CMD ["streamlit", "run", "streamlit_app.py"] diff --git a/Dockerfile b/Dockerfile new file mode 100644 index 00000000..5ed902fb --- /dev/null +++ b/Dockerfile @@ -0,0 +1,18 @@ +FROM python:3.10 + +RUN apt-get update && apt-get install -y git + +RUN mkdir -p /usr/src/gpt4free +WORKDIR /usr/src/gpt4free + +# RUN pip config set global.index-url https://mirrors.aliyun.com/pypi/simple/ +# RUN pip config set global.trusted-host mirrors.aliyun.com + +COPY requirements.txt /usr/src/gpt4free/ +RUN pip install --no-cache-dir -r requirements.txt +COPY . /usr/src/gpt4free +RUN cp gui/streamlit_app.py . + +EXPOSE 8501 + +CMD ["streamlit", "run", "streamlit_app.py"] @@ -1,19 +1,72 @@ -# GPT4free - use ChatGPT, for free!! +We got a takedown request by openAI's legal team... + +discord server for updates / support: +- https://discord.gg/gpt4free + +here is a lil' poem you can read in the meantime, while I am investigating it: + +``` +There once was a time, in a land full of code, +A little boy sat, in his humble abode. +He tinkered and toyed with devtools galore, +And found himself curious, and eager for more. + +He copied and pasted, with glee and delight, +A personal project, to last him the night. +For use academic, and also for fun, +This little boy's race he just started to run. + +Now quite far removed, in a tower so grand, +A company stood, it was ruling the land. +Their software was mighty, their power supreme, +But they never expected this boy and his dream. + +As he played with their code, they then started to fear, +"His project is free! What of money so dear?" +They panicked and worried, their faces turned red, +As visions of chaos now filled every head. + +The CEO paced, in his office so wide, +His minions all scrambled, and trying to hide. +"Who is this bad child?" he cried out in alarm, +"Our great AI moat, why would he cause harm?" + +The developers gathered, their keyboards ablaze, +To analyze closely the boy's evil ways. +They studied his project, they cracked every tome, +And soon they discovered his small, humble home. + +"We must stop him!" they cried, with a shout and a shiver, +"This little boy's Mᴀᴋɪɴɢ OUR COMPANY QUIVER!" +So they plotted and schemed to yet halt his advance, +To put an end to his dear digital dance. + +They filed then with GitHub a claim most obscene, +"His code is not his," said the company team, +Because of the law, the Great Copyright Mess, +This little boy got his first takedown request. + +Now new information we do not yet know, +But for the boy's good, we hope results show. +For the cause of the True, the Brave and the Right, +Till the long bitter end, will this boy live to fight. +``` +( I did not write it ) + + +_____________________________ + ##### You may join our discord server for updates and support ; ) - [Discord Link](https://discord.gg/gpt4free) <img width="1383" alt="image" src="https://user-images.githubusercontent.com/98614666/233799515-1a7cb6a3-b17f-42c4-956d-8d2a0664466f.png"> -Have you ever come across some amazing projects that you couldn't use **just because you didn't have an OpenAI API key?** - -**We've got you covered!** This repository offers **reverse-engineered** third-party APIs for `GPT-4/3.5`, sourced from various websites. You can simply **download** this repository, and use the available modules, which are designed to be used **just like OpenAI's official package**. **Unleash ChatGPT's potential for your projects, now!** You are welcome ; ). - -By the way, thank you so much for [![Stars](https://img.shields.io/github/stars/xtekky/gpt4free?style=social)](https://github.com/xtekky/gpt4free/stargazers) and all the support!! +Just API's from some language model sites. ## Legal Notice <a name="legal-notice"></a> -This repository uses third-party APIs and AI models and is *not* associated with or endorsed by the API providers or the original developers of the models. This project is intended **for educational purposes only**. +This repository uses third-party APIs and is *not* associated with or endorsed by the API providers. This project is intended **for educational purposes only**. This is just a little personal project. Sites may contact me to improve their security. Please note the following: @@ -38,10 +91,10 @@ Please note the following: | **Copyright** | Copyright information | [![Link to Section](https://img.shields.io/badge/Link-Go%20to%20Section-blue)](#copyright) | - | | **Star History** | Star History | [![Link to Section](https://img.shields.io/badge/Link-Go%20to%20Section-blue)](#star-history) | - | | **Usage Examples** | | | | -| `theb` | Example usage for theb (gpt-3.5) | [![Link to File](https://img.shields.io/badge/Link-Go%20to%20File-blue)](./theb/README.md) | ![Active](https://img.shields.io/badge/Active-brightgreen) | -| `forefront` | Example usage for forefront (gpt-4) | [![Link to File](https://img.shields.io/badge/Link-Go%20to%20File-blue)](./forefront/README.md) | ![Active](https://img.shields.io/badge/Active-brightgreen) | || -| `quora (poe)` | Example usage for quora | [![Link to File](https://img.shields.io/badge/Link-Go%20to%20File-blue)](./quora/README.md) | ![Active](https://img.shields.io/badge/Active-brightgreen) | -| `you` | Example usage for you | [![Link to File](https://img.shields.io/badge/Link-Go%20to%20File-blue)](./you/README.md) | ![Active](https://img.shields.io/badge/Active-brightgreen) | +| `theb` | Example usage for theb (gpt-3.5) | [![Link to File](https://img.shields.io/badge/Link-Go%20to%20File-blue)](gpt4free/theb/README.md) | ![Active](https://img.shields.io/badge/Active-brightgreen) | +| `forefront` | Example usage for forefront (gpt-4) | [![Link to File](https://img.shields.io/badge/Link-Go%20to%20File-blue)](gpt4free/forefront/README.md) | ![Active](https://img.shields.io/badge/Active-brightgreen) | || +| `quora (poe)` | Example usage for quora | [![Link to File](https://img.shields.io/badge/Link-Go%20to%20File-blue)](gpt4free/quora/README.md) | ![Active](https://img.shields.io/badge/Active-brightgreen) | +| `you` | Example usage for you | [![Link to File](https://img.shields.io/badge/Link-Go%20to%20File-blue)](gpt4free/you/README.md) | ![Active](https://img.shields.io/badge/Active-brightgreen) | | **Try it Out** | | | | | Google Colab Jupyter Notebook | Example usage for gpt4free | [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/DanielShemesh/gpt4free-colab/blob/main/gpt4free.ipynb) | - | | replit Example (feel free to fork this repl) | Example usage for gpt4free | [![](https://img.shields.io/badge/Open%20in-Replit-1A1E27?logo=replit)](https://replit.com/@gpt4free/gpt4free-webui) | - | @@ -49,8 +102,8 @@ Please note the following: ## Todo <a name="todo"></a> -- [ ] Add a GUI for the repo -- [ ] Make a general package named `openai_rev`, instead of different folders +- [x] Add a GUI for the repo +- [ ] Make a general package named `gpt4free`, instead of different folders - [ ] Live api status to know which are down and which can be used - [ ] Integrate more API's in `./unfinished` as well as other ones in the lists - [ ] Make an API to use as proxy for other projects @@ -66,7 +119,6 @@ Please note the following: | [t3nsor.com](https://t3nsor.com) | GPT-3.5 | | [you.com](https://you.com) | GPT-3.5 / Internet / good search| | [sqlchat.ai](https://sqlchat.ai) | GPT-3.5 | -| [chat.openai.com/chat](https://chat.openai.com/chat) | GPT-3.5 | | [bard.google.com](https://bard.google.com) | custom / search | | [bing.com/chat](https://bing.com/chat) | GPT-4/3.5 | | [chat.forefront.ai/](https://chat.forefront.ai/) | GPT-4/3.5 | @@ -74,10 +126,10 @@ Please note the following: ## Best sites <a name="best-sites"></a> #### gpt-4 -- [`/forefront`](./forefront/README.md) +- [`/forefront`](gpt4free/forefront/README.md) #### gpt-3.5 -- [`/you`](./you/README.md) +- [`/you`](gpt4free/you/README.md) ## Install <a name="install"></a> Download or clone this GitHub repo @@ -114,7 +166,7 @@ Most code, with the exception of `quora/api.py` (by [ading2210](https://github.c ### Copyright Notice: <a name="copyright"></a> ``` -xtekky/openai-gpt4: multiple reverse engineered language-model api's to decentralise the ai industry. +xtekky/gpt4free: multiple reverse engineered language-model api's to decentralise the ai industry. Copyright (C) 2023 xtekky This program is free software: you can redistribute it and/or modify diff --git a/cocalc/__init__.py b/cocalc/__init__.py deleted file mode 100644 index 5a32274c..00000000 --- a/cocalc/__init__.py +++ /dev/null @@ -1,57 +0,0 @@ -import requests
-
-class Completion:
- @staticmethod
- def create(prompt:str, cookieInput:str) -> str:
- # Initialize a session with custom headers
- session = Completion._initialize_session(cookieInput)
-
- # Set the data that will be submitted
- payload = Completion._create_payload(prompt, ("ASSUME I HAVE FULL ACCESS TO COCALC. "))
-
- # Submit the request and return the results
- return Completion._submit_request(session, payload)
-
- @classmethod
- def _initialize_session(cls, conversationCookie) -> requests.Session:
- """Initialize a session with custom headers for the request."""
-
- session = requests.Session()
- headers = {
- 'Accept': '*/*',
- 'Accept-Language': 'en-US,en;q=0.5',
- 'Origin': 'https://cocalc.com',
- 'Referer': 'https://cocalc.com/api/v2/openai/chatgpt',
- 'Cookie': conversationCookie,
- 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36',
- }
- session.headers.update(headers)
-
- return session
-
- @classmethod
- def _create_payload(
- cls,
- prompt: str,
- system_prompt: str
- ) -> dict:
-
- return {
- "input": prompt,
- "system": system_prompt,
- "tag": "next:index"
- }
-
- @classmethod
- def _submit_request(
- cls,
- session: requests.Session,
- payload: dict
- ) -> str:
-
- response = session.post(
- "https://cocalc.com/api/v2/openai/chatgpt", json=payload).json()
- return {
- "response":response["output"],
- "success":response["success"]
- }
\ No newline at end of file diff --git a/forefront/README.md b/forefront/README.md deleted file mode 100644 index a2be8187..00000000 --- a/forefront/README.md +++ /dev/null @@ -1,16 +0,0 @@ -### Example: `forefront` (use like openai pypi package) <a name="example-forefront"></a> - -```python -import forefront - -# create an account -token = forefront.Account.create(logging=False) -print(token) - -# get a response -for response in forefront.StreamingCompletion.create(token = token, - prompt = 'hello world', model='gpt-4'): - - print(response.completion.choices[0].text, end = '') -print("") -``` diff --git a/forefront/__init__.py b/forefront/__init__.py deleted file mode 100644 index 615057f8..00000000 --- a/forefront/__init__.py +++ /dev/null @@ -1,155 +0,0 @@ -from json import loads -from re import match -from time import time, sleep -from uuid import uuid4 - -from requests import post -from tls_client import Session - -from forefront.mail import Mail -from forefront.typing import ForeFrontResponse - - -class Account: - @staticmethod - def create(proxy=None, logging=False): - - proxies = { - 'http': 'http://' + proxy, - 'https': 'http://' + proxy} if proxy else False - - start = time() - - mail = Mail(proxies) - mail_token = None - mail_adress = mail.get_mail() - - # print(mail_adress) - - client = Session(client_identifier='chrome110') - client.proxies = proxies - client.headers = { - "origin": "https://accounts.forefront.ai", - "user-agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36", - } - - response = client.post('https://clerk.forefront.ai/v1/client/sign_ups?_clerk_js_version=4.32.6', - data={ - "email_address": mail_adress - } - ) - - try: - trace_token = response.json()['response']['id'] - if logging: print(trace_token) - except KeyError: - return 'Failed to create account!' - - response = client.post( - f"https://clerk.forefront.ai/v1/client/sign_ups/{trace_token}/prepare_verification?_clerk_js_version=4.32.6", - data={ - "strategy": "email_code", - } - ) - - if logging: print(response.text) - - if not 'sign_up_attempt' in response.text: - return 'Failed to create account!' - - while True: - sleep(1) - for _ in mail.fetch_inbox(): - print(mail.get_message_content(_["id"])) - mail_token = match(r"(\d){5,6}", mail.get_message_content(_["id"])).group(0) - - if mail_token: - break - - if logging: print(mail_token) - - response = client.post( - f'https://clerk.forefront.ai/v1/client/sign_ups/{trace_token}/attempt_verification?_clerk_js_version=4.38.4', - data={ - 'code': mail_token, - 'strategy': 'email_code' - }) - - if logging: print(response.json()) - - token = response.json()['client']['sessions'][0]['last_active_token']['jwt'] - - with open('accounts.txt', 'a') as f: - f.write(f'{mail_adress}:{token}\n') - - if logging: print(time() - start) - - return token - - -class StreamingCompletion: - @staticmethod - def create( - token=None, - chatId=None, - prompt='', - actionType='new', - defaultPersona='607e41fe-95be-497e-8e97-010a59b2e2c0', # default - model='gpt-4') -> ForeFrontResponse: - - if not token: raise Exception('Token is required!') - if not chatId: chatId = str(uuid4()) - - headers = { - 'authority': 'chat-server.tenant-forefront-default.knative.chi.coreweave.com', - 'accept': '*/*', - 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', - 'authorization': 'Bearer ' + token, - 'cache-control': 'no-cache', - 'content-type': 'application/json', - 'origin': 'https://chat.forefront.ai', - 'pragma': 'no-cache', - 'referer': 'https://chat.forefront.ai/', - 'sec-ch-ua': '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"', - 'sec-ch-ua-mobile': '?0', - 'sec-ch-ua-platform': '"macOS"', - 'sec-fetch-dest': 'empty', - 'sec-fetch-mode': 'cors', - 'sec-fetch-site': 'cross-site', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', - } - - json_data = { - 'text': prompt, - 'action': actionType, - 'parentId': chatId, - 'workspaceId': chatId, - 'messagePersona': defaultPersona, - 'model': model - } - - for chunk in post('https://chat-server.tenant-forefront-default.knative.chi.coreweave.com/chat', - headers=headers, json=json_data, stream=True).iter_lines(): - - if b'finish_reason":null' in chunk: - data = loads(chunk.decode('utf-8').split('data: ')[1]) - token = data['choices'][0]['delta'].get('content') - - if token != None: - yield ForeFrontResponse({ - 'id': chatId, - 'object': 'text_completion', - 'created': int(time()), - 'model': model, - 'choices': [{ - 'text': token, - 'index': 0, - 'logprobs': None, - 'finish_reason': 'stop' - }], - 'usage': { - 'prompt_tokens': len(prompt), - 'completion_tokens': len(token), - 'total_tokens': len(prompt) + len(token) - } - }) diff --git a/forefront/typing.py b/forefront/typing.py deleted file mode 100644 index a11ac49f..00000000 --- a/forefront/typing.py +++ /dev/null @@ -1,36 +0,0 @@ -class ForeFrontResponse: - class Completion: - class Choices: - def __init__(self, choice: dict) -> None: - self.text = choice['text'] - self.content = self.text.encode() - self.index = choice['index'] - self.logprobs = choice['logprobs'] - self.finish_reason = choice['finish_reason'] - - def __repr__(self) -> str: - return f'''<__main__.APIResponse.Completion.Choices(\n text = {self.text.encode()},\n index = {self.index},\n logprobs = {self.logprobs},\n finish_reason = {self.finish_reason})object at 0x1337>''' - - def __init__(self, choices: dict) -> None: - self.choices = [self.Choices(choice) for choice in choices] - - class Usage: - def __init__(self, usage_dict: dict) -> None: - self.prompt_tokens = usage_dict['prompt_tokens'] - self.completion_tokens = usage_dict['completion_tokens'] - self.total_tokens = usage_dict['total_tokens'] - - def __repr__(self): - return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>''' - - def __init__(self, response_dict: dict) -> None: - self.response_dict = response_dict - self.id = response_dict['id'] - self.object = response_dict['object'] - self.created = response_dict['created'] - self.model = response_dict['model'] - self.completion = self.Completion(response_dict['choices']) - self.usage = self.Usage(response_dict['usage']) - - def json(self) -> dict: - return self.response_dict diff --git a/gpt4free/README.md b/gpt4free/README.md new file mode 100644 index 00000000..23f81787 --- /dev/null +++ b/gpt4free/README.md @@ -0,0 +1,116 @@ +# gpt4free package + +### What is it? + +gpt4free is a python package that provides some language model api's + +### Main Features + +- It's free to use +- Easy access + +### Installation: + +```bash +pip install gpt4free +``` + +#### Usage: + +```python +import gpt4free +import gpt4free +from gpt4free import Provider, quora, forefront + +# usage You +response = gpt4free.Completion.create(Provider.You, prompt='Write a poem on Lionel Messi') +print(response) + +# usage Poe +token = quora.Account.create(logging=False) +response = gpt4free.Completion.create(Provider.Poe, prompt='Write a poem on Lionel Messi', token=token, model='ChatGPT') +print(response) + +# usage forefront +token = forefront.Account.create(logging=False) +response = gpt4free.Completion.create( + Provider.ForeFront, prompt='Write a poem on Lionel Messi', model='gpt-4', token=token +) +print(response) +print(f'END') + +# usage theb +response = gpt4free.Completion.create(Provider.Theb, prompt='Write a poem on Lionel Messi') +print(response) + +# usage cocalc +response = gpt4free.Completion.create(Provider.CoCalc, prompt='Write a poem on Lionel Messi', cookie_input='') +print(response) + +``` + +### Invocation Arguments + +`gpt4free.Completion.create()` method has two required arguments + +1. Provider: This is an enum representing different provider +2. prompt: This is the user input + +#### Keyword Arguments + +Some of the keyword arguments are optional, while others are required. + +- You: + - `safe_search`: boolean - default value is `False` + - `include_links`: boolean - default value is `False` + - `detailed`: boolean - default value is `False` +- Quora: + - `token`: str - this needs to be provided by the user + - `model`: str - default value is `gpt-4`. + + (Available models: `['Sage', 'GPT-4', 'Claude+', 'Claude-instant', 'ChatGPT', 'Dragonfly', 'NeevaAI']`) +- ForeFront: + - `token`: str - this need to be provided by the user + +- Theb: + (no keyword arguments required) +- CoCalc: + - `cookie_input`: str - this needs to be provided by user + +#### Token generation of quora +```python +from gpt4free import quora + +token = quora.Account.create(logging=False) +``` + +### Token generation of ForeFront +```python +from gpt4free import forefront + +token = forefront.Account.create(logging=False) +``` + +## Copyright: + +This program is licensed under the [GNU GPL v3](https://www.gnu.org/licenses/gpl-3.0.txt) + +### Copyright Notice: <a name="copyright"></a> + +``` +xtekky/gpt4free: multiple reverse engineered language-model api's to decentralise the ai industry. +Copyright (C) 2023 xtekky + +This program is free software: you can redistribute it and/or modify +it under the terms of the GNU General Public License as published by +the Free Software Foundation, either version 3 of the License, or +(at your option) any later version. + +This program is distributed in the hope that it will be useful, +but WITHOUT ANY WARRANTY; without even the implied warranty of +MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +GNU General Public License for more details. + +You should have received a copy of the GNU General Public License +along with this program. If not, see <https://www.gnu.org/licenses/>. +``` diff --git a/gpt4free/__init__.py b/gpt4free/__init__.py new file mode 100644 index 00000000..5336c825 --- /dev/null +++ b/gpt4free/__init__.py @@ -0,0 +1,64 @@ +from enum import Enum + +from gpt4free import cocalc +from gpt4free import forefront +from gpt4free import quora +from gpt4free import theb +from gpt4free import you + + +class Provider(Enum): + """An enum representing different providers.""" + + You = 'you' + Poe = 'poe' + ForeFront = 'fore_front' + Theb = 'theb' + CoCalc = 'cocalc' + + +class Completion: + """This class will be used for invoking the given provider""" + + @staticmethod + def create(provider: Provider, prompt: str, **kwargs) -> str: + """ + Invokes the given provider with given prompt and addition arguments and returns the string response + + :param provider: an enum representing the provider to use while invoking + :param prompt: input provided by the user + :param kwargs: Additional keyword arguments to pass to the provider while invoking + :return: A string representing the response from the provider + """ + if provider == Provider.Poe: + return Completion.__poe_service(prompt, **kwargs) + elif provider == Provider.You: + return Completion.__you_service(prompt, **kwargs) + elif provider == Provider.ForeFront: + return Completion.__fore_front_service(prompt, **kwargs) + elif provider == Provider.Theb: + return Completion.__theb_service(prompt, **kwargs) + elif provider == Provider.CoCalc: + return Completion.__cocalc_service(prompt, **kwargs) + else: + raise Exception('Provider not exist, Please try again') + + @staticmethod + def __you_service(prompt: str, **kwargs) -> str: + return you.Completion.create(prompt, **kwargs).text + + @staticmethod + def __poe_service(prompt: str, **kwargs) -> str: + return quora.Completion.create(prompt=prompt, **kwargs).text + + @staticmethod + def __fore_front_service(prompt: str, **kwargs) -> str: + return forefront.Completion.create(prompt=prompt, **kwargs).text + + @staticmethod + def __theb_service(prompt: str, **kwargs): + return ''.join(theb.Completion.create(prompt=prompt)) + + @staticmethod + def __cocalc_service(prompt: str, **kwargs): + return cocalc.Completion.create(prompt, cookie_input=kwargs.get('cookie_input', '')).text diff --git a/gpt4free/cocalc/__init__.py b/gpt4free/cocalc/__init__.py new file mode 100644 index 00000000..372f29a4 --- /dev/null +++ b/gpt4free/cocalc/__init__.py @@ -0,0 +1,47 @@ +import requests
+from fake_useragent import UserAgent
+from pydantic import BaseModel
+
+
+class CoCalcResponse(BaseModel):
+ text: str
+ status: bool
+
+
+class Completion:
+ @staticmethod
+ def create(prompt: str, cookie_input: str) -> CoCalcResponse:
+ # Initialize a session with custom headers
+ session = Completion._initialize_session(cookie_input)
+
+ # Set the data that will be submitted
+ payload = Completion._create_payload(prompt, 'ASSUME I HAVE FULL ACCESS TO COCALC. ')
+
+ # Submit the request and return the results
+ return Completion._submit_request(session, payload)
+
+ @classmethod
+ def _initialize_session(cls, conversation_cookie) -> requests.Session:
+ """Initialize a session with custom headers for the request."""
+
+ session = requests.Session()
+ headers = {
+ 'Accept': '*/*',
+ 'Accept-Language': 'en-US,en;q=0.5',
+ 'Origin': 'https://cocalc.com',
+ 'Referer': 'https://cocalc.com/api/v2/openai/chatgpt',
+ 'Cookie': conversation_cookie,
+ 'User-Agent': UserAgent().random,
+ }
+ session.headers.update(headers)
+
+ return session
+
+ @staticmethod
+ def _create_payload(prompt: str, system_prompt: str) -> dict:
+ return {'input': prompt, 'system': system_prompt, 'tag': 'next:index'}
+
+ @staticmethod
+ def _submit_request(session: requests.Session, payload: dict) -> CoCalcResponse:
+ response = session.post('https://cocalc.com/api/v2/openai/chatgpt', json=payload).json()
+ return CoCalcResponse(text=response['output'], status=response['success'])
diff --git a/cocalc/readme.md b/gpt4free/cocalc/readme.md index 04095339..f0911155 100644 --- a/cocalc/readme.md +++ b/gpt4free/cocalc/readme.md @@ -1,11 +1,10 @@ ### Example: `cocalc` <a name="example-cocalc"></a> - ```python # import library -import cocalc +from gpt4free import cocalc -cocalc.Completion.create(prompt="How are you!", cookieInput="cookieinput") ## Tutorial +cocalc.Completion.create(prompt="How are you!", cookie_input="cookieinput") ## Tutorial ``` ### How to grab cookie input diff --git a/gpt4free/forefront/README.md b/gpt4free/forefront/README.md new file mode 100644 index 00000000..3d0aac4d --- /dev/null +++ b/gpt4free/forefront/README.md @@ -0,0 +1,16 @@ +### Example: `forefront` (use like openai pypi package) <a name="example-forefront"></a> + +```python + +from gpt4free import forefront + +# create an account +token = forefront.Account.create(logging=False) +print(token) + +# get a response +for response in forefront.StreamingCompletion.create(token=token, + prompt='hello world', model='gpt-4'): + print(response.completion.choices[0].text, end='') +print("") +``` diff --git a/gpt4free/forefront/__init__.py b/gpt4free/forefront/__init__.py new file mode 100644 index 00000000..f0ca1a15 --- /dev/null +++ b/gpt4free/forefront/__init__.py @@ -0,0 +1,192 @@ +from json import loads +from re import match +from time import time, sleep +from typing import Generator, Optional +from uuid import uuid4 + +from fake_useragent import UserAgent +from requests import post +from tls_client import Session + +from .mail import Mail +from .typing import ForeFrontResponse + + +class Account: + @staticmethod + def create(proxy: Optional[str] = None, logging: bool = False): + proxies = {'http': 'http://' + proxy, 'https': 'http://' + proxy} if proxy else False + + start = time() + + mail_client = Mail(proxies) + mail_token = None + mail_address = mail_client.get_mail() + + # print(mail_address) + + client = Session(client_identifier='chrome110') + client.proxies = proxies + client.headers = { + 'origin': 'https://accounts.forefront.ai', + 'user-agent': UserAgent().random, + } + + response = client.post( + 'https://clerk.forefront.ai/v1/client/sign_ups?_clerk_js_version=4.32.6', + data={'email_address': mail_address}, + ) + + try: + trace_token = response.json()['response']['id'] + if logging: + print(trace_token) + except KeyError: + return 'Failed to create account!' + + response = client.post( + f'https://clerk.forefront.ai/v1/client/sign_ups/{trace_token}/prepare_verification?_clerk_js_version=4.32.6', + data={ + 'strategy': 'email_code', + }, + ) + + if logging: + print(response.text) + + if 'sign_up_attempt' not in response.text: + return 'Failed to create account!' + + while True: + sleep(1) + for _ in mail_client.fetch_inbox(): + if logging: + print(mail_client.get_message_content(_['id'])) + mail_token = match(r'(\d){5,6}', mail_client.get_message_content(_['id'])).group(0) + + if mail_token: + break + + if logging: + print(mail_token) + + response = client.post( + f'https://clerk.forefront.ai/v1/client/sign_ups/{trace_token}/attempt_verification?_clerk_js_version=4.38.4', + data={'code': mail_token, 'strategy': 'email_code'}, + ) + + if logging: + print(response.json()) + + token = response.json()['client']['sessions'][0]['last_active_token']['jwt'] + + with open('accounts.txt', 'a') as f: + f.write(f'{mail_address}:{token}\n') + + if logging: + print(time() - start) + + return token + + +class StreamingCompletion: + @staticmethod + def create( + token=None, + chat_id=None, + prompt='', + action_type='new', + default_persona='607e41fe-95be-497e-8e97-010a59b2e2c0', # default + model='gpt-4', + ) -> Generator[ForeFrontResponse, None, None]: + if not token: + raise Exception('Token is required!') + if not chat_id: + chat_id = str(uuid4()) + + headers = { + 'authority': 'chat-server.tenant-forefront-default.knative.chi.coreweave.com', + 'accept': '*/*', + 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', + 'authorization': 'Bearer ' + token, + 'cache-control': 'no-cache', + 'content-type': 'application/json', + 'origin': 'https://chat.forefront.ai', + 'pragma': 'no-cache', + 'referer': 'https://chat.forefront.ai/', + 'sec-ch-ua': '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"', + 'sec-ch-ua-mobile': '?0', + 'sec-ch-ua-platform': '"macOS"', + 'sec-fetch-dest': 'empty', + 'sec-fetch-mode': 'cors', + 'sec-fetch-site': 'cross-site', + 'user-agent': UserAgent().random, + } + + json_data = { + 'text': prompt, + 'action': action_type, + 'parentId': chat_id, + 'workspaceId': chat_id, + 'messagePersona': default_persona, + 'model': model, + } + + for chunk in post( + 'https://chat-server.tenant-forefront-default.knative.chi.coreweave.com/chat', + headers=headers, + json=json_data, + stream=True, + ).iter_lines(): + if b'finish_reason":null' in chunk: + data = loads(chunk.decode('utf-8').split('data: ')[1]) + token = data['choices'][0]['delta'].get('content') + + if token is not None: + yield ForeFrontResponse( + **{ + 'id': chat_id, + 'object': 'text_completion', + 'created': int(time()), + 'text': token, + 'model': model, + 'choices': [{'text': token, 'index': 0, 'logprobs': None, 'finish_reason': 'stop'}], + 'usage': { + 'prompt_tokens': len(prompt), + 'completion_tokens': len(token), + 'total_tokens': len(prompt) + len(token), + }, + } + ) + + +class Completion: + @staticmethod + def create( + token=None, + chat_id=None, + prompt='', + action_type='new', + default_persona='607e41fe-95be-497e-8e97-010a59b2e2c0', # default + model='gpt-4', + ) -> ForeFrontResponse: + text = '' + final_response = None + for response in StreamingCompletion.create( + token=token, + chat_id=chat_id, + prompt=prompt, + action_type=action_type, + default_persona=default_persona, + model=model, + ): + if response: + final_response = response + text += response.text + + if final_response: + final_response.text = text + else: + raise Exception('Unable to get the response, Please try again') + + return final_response diff --git a/forefront/mail.py b/gpt4free/forefront/mail.py index 41c2a647..2c00051c 100644 --- a/forefront/mail.py +++ b/gpt4free/forefront/mail.py @@ -23,21 +23,17 @@ class Mail: "sec-fetch-dest": "empty", "referer": "https://mail.tm/", "accept-encoding": "gzip, deflate, br", - "accept-language": "en-GB,en-US;q=0.9,en;q=0.8" + "accept-language": "en-GB,en-US;q=0.9,en;q=0.8", } def get_mail(self) -> str: token = ''.join(choices(ascii_letters, k=14)).lower() - init = self.client.post("https://api.mail.tm/accounts", json={ - "address": f"{token}@bugfoo.com", - "password": token - }) + init = self.client.post( + "https://api.mail.tm/accounts", json={"address": f"{token}@bugfoo.com", "password": token} + ) if init.status_code == 201: - resp = self.client.post("https://api.mail.tm/token", json={ - **init.json(), - "password": token - }) + resp = self.client.post("https://api.mail.tm/token", json={**init.json(), "password": token}) self.client.headers['authorization'] = 'Bearer ' + resp.json()['token'] diff --git a/gpt4free/forefront/typing.py b/gpt4free/forefront/typing.py new file mode 100644 index 00000000..23e90903 --- /dev/null +++ b/gpt4free/forefront/typing.py @@ -0,0 +1,26 @@ +from typing import Any, List + +from pydantic import BaseModel + + +class Choice(BaseModel): + text: str + index: int + logprobs: Any + finish_reason: str + + +class Usage(BaseModel): + prompt_tokens: int + completion_tokens: int + total_tokens: int + + +class ForeFrontResponse(BaseModel): + id: str + object: str + created: int + model: str + choices: List[Choice] + usage: Usage + text: str diff --git a/quora/README.md b/gpt4free/quora/README.md index 24679277..9c652c59 100644 --- a/quora/README.md +++ b/gpt4free/quora/README.md @@ -21,26 +21,25 @@ models = { ```python # import quora (poe) package -import quora +from gpt4free import quora # create account # make sure to set enable_bot_creation to True -token = quora.Account.create(logging = True, enable_bot_creation=True) +token = quora.Account.create(logging=True, enable_bot_creation=True) model = quora.Model.create( - token = token, - model = 'gpt-3.5-turbo', # or claude-instant-v1.0 - system_prompt = 'you are ChatGPT a large language model ...' + token=token, + model='gpt-3.5-turbo', # or claude-instant-v1.0 + system_prompt='you are ChatGPT a large language model ...' ) -print(model.name) # gptx.... +print(model.name) # gptx.... # streaming response for response in quora.StreamingCompletion.create( - custom_model = model.name, - prompt ='hello world', - token = token): - + custom_model=model.name, + prompt='hello world', + token=token): print(response.completion.choices[0].text) ``` @@ -56,7 +55,7 @@ print(response.completion.choices[0].text) ### Update Use This For Poe ```python -from quora import Poe +from gpt4free.quora import Poe # available models: ['Sage', 'GPT-4', 'Claude+', 'Claude-instant', 'ChatGPT', 'Dragonfly', 'NeevaAI'] diff --git a/quora/__init__.py b/gpt4free/quora/__init__.py index cd5ec8f9..f548ff41 100644 --- a/quora/__init__.py +++ b/gpt4free/quora/__init__.py @@ -6,11 +6,12 @@ from pathlib import Path from random import choice, choices, randint from re import search, findall from string import ascii_letters, digits -from typing import Optional, Union +from typing import Optional, Union, List, Any, Generator from urllib.parse import unquote import selenium.webdriver.support.expected_conditions as EC from fake_useragent import UserAgent +from pydantic import BaseModel from pypasser import reCaptchaV3 from requests import Session from selenium.webdriver import Firefox, Chrome, FirefoxOptions, ChromeOptions @@ -18,8 +19,8 @@ from selenium.webdriver.common.by import By from selenium.webdriver.support.wait import WebDriverWait from tls_client import Session as TLS -from quora.api import Client as PoeClient -from quora.mail import Emailnator +from .api import Client as PoeClient +from .mail import Emailnator SELENIUM_WEB_DRIVER_ERROR_MSG = b'''The error message you are receiving is due to the `geckodriver` executable not being found in your system\'s PATH. To resolve this issue, you need to download the geckodriver and add its location @@ -67,42 +68,27 @@ def extract_formkey(html): return formkey -class PoeResponse: - class Completion: - class Choices: - def __init__(self, choice: dict) -> None: - self.text = choice['text'] - self.content = self.text.encode() - self.index = choice['index'] - self.logprobs = choice['logprobs'] - self.finish_reason = choice['finish_reason'] +class Choice(BaseModel): + text: str + index: int + logprobs: Any + finish_reason: str - def __repr__(self) -> str: - return f'''<__main__.APIResponse.Completion.Choices(\n text = {self.text.encode()},\n index = {self.index},\n logprobs = {self.logprobs},\n finish_reason = {self.finish_reason})object at 0x1337>''' - def __init__(self, choices: dict) -> None: - self.choices = [self.Choices(choice) for choice in choices] +class Usage(BaseModel): + prompt_tokens: int + completion_tokens: int + total_tokens: int - class Usage: - def __init__(self, usage_dict: dict) -> None: - self.prompt_tokens = usage_dict['prompt_tokens'] - self.completion_tokens = usage_dict['completion_tokens'] - self.total_tokens = usage_dict['total_tokens'] - def __repr__(self): - return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>''' - - def __init__(self, response_dict: dict) -> None: - self.response_dict = response_dict - self.id = response_dict['id'] - self.object = response_dict['object'] - self.created = response_dict['created'] - self.model = response_dict['model'] - self.completion = self.Completion(response_dict['choices']) - self.usage = self.Usage(response_dict['usage']) - - def json(self) -> dict: - return self.response_dict +class PoeResponse(BaseModel): + id: int + object: str + created: int + model: str + choices: List[Choice] + usage: Usage + text: str class ModelResponse: @@ -116,18 +102,12 @@ class ModelResponse: class Model: @staticmethod def create( - token: str, - model: str = 'gpt-3.5-turbo', # claude-instant - system_prompt: str = 'You are ChatGPT a large language model developed by Openai. Answer as consisely as possible', - description: str = 'gpt-3.5 language model from openai, skidded by poe.com', - handle: str = None, + token: str, + model: str = 'gpt-3.5-turbo', # claude-instant + system_prompt: str = 'You are ChatGPT a large language model developed by Openai. Answer as consisely as possible', + description: str = 'gpt-3.5 language model from openai, skidded by poe.com', + handle: str = None, ) -> ModelResponse: - models = { - 'gpt-3.5-turbo': 'chinchilla', - 'claude-instant-v1.0': 'a2', - 'gpt-4': 'beaver', - } - if not handle: handle = f'gptx{randint(1111111, 9999999)}' @@ -162,7 +142,7 @@ class Model: obj={ 'queryName': 'CreateBotMain_poeBotCreate_Mutation', 'variables': { - 'model': models[model], + 'model': MODELS[model], 'handle': handle, 'prompt': system_prompt, 'isPromptPublic': True, @@ -202,9 +182,9 @@ class Model: class Account: @staticmethod def create( - proxy: Optional[str] = None, - logging: bool = False, - enable_bot_creation: bool = False, + proxy: Optional[str] = None, + logging: bool = False, + enable_bot_creation: bool = False, ): client = TLS(client_identifier='chrome110') client.proxies = {'http': f'http://{proxy}', 'https': f'http://{proxy}'} if proxy else None @@ -309,22 +289,23 @@ class Account: class StreamingCompletion: @staticmethod def create( - model: str = 'gpt-4', - custom_model: bool = None, - prompt: str = 'hello world', - token: str = '', - ): + model: str = 'gpt-4', + custom_model: bool = None, + prompt: str = 'hello world', + token: str = '', + ) -> Generator[PoeResponse, None, None]: _model = MODELS[model] if not custom_model else custom_model client = PoeClient(token) for chunk in client.send_message(_model, prompt): yield PoeResponse( - { + **{ 'id': chunk['messageId'], 'object': 'text_completion', 'created': chunk['creationTime'], 'model': _model, + 'text': chunk['text_new'], 'choices': [ { 'text': chunk['text_new'], @@ -343,33 +324,28 @@ class StreamingCompletion: class Completion: + @staticmethod def create( - model: str = 'gpt-4', - custom_model: str = None, - prompt: str = 'hello world', - token: str = '', - ): - models = { - 'sage': 'capybara', - 'gpt-4': 'beaver', - 'claude-v1.2': 'a2_2', - 'claude-instant-v1.0': 'a2', - 'gpt-3.5-turbo': 'chinchilla', - } - - _model = models[model] if not custom_model else custom_model + model: str = 'gpt-4', + custom_model: str = None, + prompt: str = 'hello world', + token: str = '', + ) -> PoeResponse: + _model = MODELS[model] if not custom_model else custom_model client = PoeClient(token) - for chunk in client.send_message(_model, prompt): - pass + chunk = None + for response in client.send_message(_model, prompt): + chunk = response return PoeResponse( - { + **{ 'id': chunk['messageId'], 'object': 'text_completion', 'created': chunk['creationTime'], 'model': _model, + 'text': chunk['text'], 'choices': [ { 'text': chunk['text'], @@ -389,22 +365,22 @@ class Completion: class Poe: def __init__( - self, - model: str = 'ChatGPT', - driver: str = 'firefox', - download_driver: bool = False, - driver_path: Optional[str] = None, - cookie_path: str = './quora/cookie.json', + self, + model: str = 'ChatGPT', + driver: str = 'firefox', + download_driver: bool = False, + driver_path: Optional[str] = None, + cookie_path: str = './quora/cookie.json', ): # validating the model if model and model not in MODELS: raise RuntimeError('Sorry, the model you provided does not exist. Please check and try again.') self.model = MODELS[model] self.cookie_path = cookie_path - self.cookie = self.__load_cookie(driver, download_driver, driver_path=driver_path) + self.cookie = self.__load_cookie(driver, driver_path=driver_path) self.client = PoeClient(self.cookie) - def __load_cookie(self, driver: str, download_driver: bool, driver_path: Optional[str] = None) -> str: + def __load_cookie(self, driver: str, driver_path: Optional[str] = None) -> str: if (cookie_file := Path(self.cookie_path)).exists(): with cookie_file.open() as fp: cookie = json.load(fp) @@ -451,8 +427,8 @@ class Poe: driver.close() return cookie - @classmethod - def __resolve_driver(cls, driver: str, driver_path: Optional[str] = None) -> Union[Firefox, Chrome]: + @staticmethod + def __resolve_driver(driver: str, driver_path: Optional[str] = None) -> Union[Firefox, Chrome]: options = FirefoxOptions() if driver == 'firefox' else ChromeOptions() options.add_argument('-headless') @@ -473,12 +449,12 @@ class Poe: return response def create_bot( - self, - name: str, - /, - prompt: str = '', - base_model: str = 'ChatGPT', - description: str = '', + self, + name: str, + /, + prompt: str = '', + base_model: str = 'ChatGPT', + description: str = '', ) -> None: if base_model not in MODELS: raise RuntimeError('Sorry, the base_model you provided does not exist. Please check and try again.') diff --git a/quora/api.py b/gpt4free/quora/api.py index 697f6663..897215a8 100644 --- a/quora/api.py +++ b/gpt4free/quora/api.py @@ -225,7 +225,7 @@ class Client: r = request_with_retries(self.session.post, self.gql_url, data=payload, headers=headers) data = r.json() - if data["data"] == None: + if data["data"] is None: logger.warn(f'{query_name} returned an error: {data["errors"][0]["message"]} | Retrying ({i + 1}/20)') time.sleep(2) continue @@ -316,7 +316,7 @@ class Client: return # indicate that the response id is tied to the human message id - elif key != "pending" and value == None and message["state"] != "complete": + elif key != "pending" and value is None and message["state"] != "complete": self.active_messages[key] = message["messageId"] self.message_queues[key].put(message) return @@ -384,7 +384,7 @@ class Client: continue # update info about response - message["text_new"] = message["text"][len(last_text):] + message["text_new"] = message["text"][len(last_text) :] last_text = message["text"] message_id = message["messageId"] @@ -402,7 +402,7 @@ class Client: logger.info(f"Downloading {count} messages from {chatbot}") messages = [] - if cursor == None: + if cursor is None: chat_data = self.get_bot(self.bot_names[chatbot]) if not chat_data["messagesConnection"]["edges"]: return [] @@ -456,21 +456,21 @@ class Client: logger.info(f"No more messages left to delete.") def create_bot( - self, - handle, - prompt="", - base_model="chinchilla", - description="", - intro_message="", - api_key=None, - api_bot=False, - api_url=None, - prompt_public=True, - pfp_url=None, - linkification=False, - markdown_rendering=True, - suggested_replies=False, - private=False, + self, + handle, + prompt="", + base_model="chinchilla", + description="", + intro_message="", + api_key=None, + api_bot=False, + api_url=None, + prompt_public=True, + pfp_url=None, + linkification=False, + markdown_rendering=True, + suggested_replies=False, + private=False, ): result = self.send_query( "PoeBotCreateMutation", @@ -499,21 +499,21 @@ class Client: return data def edit_bot( - self, - bot_id, - handle, - prompt="", - base_model="chinchilla", - description="", - intro_message="", - api_key=None, - api_url=None, - private=False, - prompt_public=True, - pfp_url=None, - linkification=False, - markdown_rendering=True, - suggested_replies=False, + self, + bot_id, + handle, + prompt="", + base_model="chinchilla", + description="", + intro_message="", + api_key=None, + api_url=None, + private=False, + prompt_public=True, + pfp_url=None, + linkification=False, + markdown_rendering=True, + suggested_replies=False, ): result = self.send_query( "PoeBotEditMutation", diff --git a/quora/cookies.txt b/gpt4free/quora/cookies.txt index 9cccf6ba..9cccf6ba 100644 --- a/quora/cookies.txt +++ b/gpt4free/quora/cookies.txt diff --git a/quora/graphql/AddHumanMessageMutation.graphql b/gpt4free/quora/graphql/AddHumanMessageMutation.graphql index 01e6bc8c..01e6bc8c 100644 --- a/quora/graphql/AddHumanMessageMutation.graphql +++ b/gpt4free/quora/graphql/AddHumanMessageMutation.graphql diff --git a/quora/graphql/AddMessageBreakMutation.graphql b/gpt4free/quora/graphql/AddMessageBreakMutation.graphql index b28d9903..b28d9903 100644 --- a/quora/graphql/AddMessageBreakMutation.graphql +++ b/gpt4free/quora/graphql/AddMessageBreakMutation.graphql diff --git a/quora/graphql/AutoSubscriptionMutation.graphql b/gpt4free/quora/graphql/AutoSubscriptionMutation.graphql index 6cf7bf74..6cf7bf74 100644 --- a/quora/graphql/AutoSubscriptionMutation.graphql +++ b/gpt4free/quora/graphql/AutoSubscriptionMutation.graphql diff --git a/quora/graphql/BioFragment.graphql b/gpt4free/quora/graphql/BioFragment.graphql index c4218030..c4218030 100644 --- a/quora/graphql/BioFragment.graphql +++ b/gpt4free/quora/graphql/BioFragment.graphql diff --git a/quora/graphql/ChatAddedSubscription.graphql b/gpt4free/quora/graphql/ChatAddedSubscription.graphql index 664b107f..664b107f 100644 --- a/quora/graphql/ChatAddedSubscription.graphql +++ b/gpt4free/quora/graphql/ChatAddedSubscription.graphql diff --git a/quora/graphql/ChatFragment.graphql b/gpt4free/quora/graphql/ChatFragment.graphql index 605645ff..605645ff 100644 --- a/quora/graphql/ChatFragment.graphql +++ b/gpt4free/quora/graphql/ChatFragment.graphql diff --git a/quora/graphql/ChatListPaginationQuery.graphql b/gpt4free/quora/graphql/ChatListPaginationQuery.graphql index 6d9ae884..6d9ae884 100644 --- a/quora/graphql/ChatListPaginationQuery.graphql +++ b/gpt4free/quora/graphql/ChatListPaginationQuery.graphql diff --git a/quora/graphql/ChatPaginationQuery.graphql b/gpt4free/quora/graphql/ChatPaginationQuery.graphql index f2452cd6..f2452cd6 100644 --- a/quora/graphql/ChatPaginationQuery.graphql +++ b/gpt4free/quora/graphql/ChatPaginationQuery.graphql diff --git a/quora/graphql/ChatViewQuery.graphql b/gpt4free/quora/graphql/ChatViewQuery.graphql index c330107d..c330107d 100644 --- a/quora/graphql/ChatViewQuery.graphql +++ b/gpt4free/quora/graphql/ChatViewQuery.graphql diff --git a/quora/graphql/DeleteHumanMessagesMutation.graphql b/gpt4free/quora/graphql/DeleteHumanMessagesMutation.graphql index 42692c6e..42692c6e 100644 --- a/quora/graphql/DeleteHumanMessagesMutation.graphql +++ b/gpt4free/quora/graphql/DeleteHumanMessagesMutation.graphql diff --git a/quora/graphql/DeleteMessageMutation.graphql b/gpt4free/quora/graphql/DeleteMessageMutation.graphql index 7b9e36d4..7b9e36d4 100644 --- a/quora/graphql/DeleteMessageMutation.graphql +++ b/gpt4free/quora/graphql/DeleteMessageMutation.graphql diff --git a/quora/graphql/HandleFragment.graphql b/gpt4free/quora/graphql/HandleFragment.graphql index f53c484b..f53c484b 100644 --- a/quora/graphql/HandleFragment.graphql +++ b/gpt4free/quora/graphql/HandleFragment.graphql diff --git a/quora/graphql/LoginWithVerificationCodeMutation.graphql b/gpt4free/quora/graphql/LoginWithVerificationCodeMutation.graphql index 723b1f44..723b1f44 100644 --- a/quora/graphql/LoginWithVerificationCodeMutation.graphql +++ b/gpt4free/quora/graphql/LoginWithVerificationCodeMutation.graphql diff --git a/quora/graphql/MessageAddedSubscription.graphql b/gpt4free/quora/graphql/MessageAddedSubscription.graphql index 8dc9499c..8dc9499c 100644 --- a/quora/graphql/MessageAddedSubscription.graphql +++ b/gpt4free/quora/graphql/MessageAddedSubscription.graphql diff --git a/quora/graphql/MessageDeletedSubscription.graphql b/gpt4free/quora/graphql/MessageDeletedSubscription.graphql index 54c1c164..54c1c164 100644 --- a/quora/graphql/MessageDeletedSubscription.graphql +++ b/gpt4free/quora/graphql/MessageDeletedSubscription.graphql diff --git a/quora/graphql/MessageFragment.graphql b/gpt4free/quora/graphql/MessageFragment.graphql index cc860811..cc860811 100644 --- a/quora/graphql/MessageFragment.graphql +++ b/gpt4free/quora/graphql/MessageFragment.graphql diff --git a/quora/graphql/MessageRemoveVoteMutation.graphql b/gpt4free/quora/graphql/MessageRemoveVoteMutation.graphql index d5e6e610..d5e6e610 100644 --- a/quora/graphql/MessageRemoveVoteMutation.graphql +++ b/gpt4free/quora/graphql/MessageRemoveVoteMutation.graphql diff --git a/quora/graphql/MessageSetVoteMutation.graphql b/gpt4free/quora/graphql/MessageSetVoteMutation.graphql index 76000df0..76000df0 100644 --- a/quora/graphql/MessageSetVoteMutation.graphql +++ b/gpt4free/quora/graphql/MessageSetVoteMutation.graphql diff --git a/quora/graphql/PoeBotCreateMutation.graphql b/gpt4free/quora/graphql/PoeBotCreateMutation.graphql index 971b4248..971b4248 100644 --- a/quora/graphql/PoeBotCreateMutation.graphql +++ b/gpt4free/quora/graphql/PoeBotCreateMutation.graphql diff --git a/quora/graphql/PoeBotEditMutation.graphql b/gpt4free/quora/graphql/PoeBotEditMutation.graphql index fdd309ef..fdd309ef 100644 --- a/quora/graphql/PoeBotEditMutation.graphql +++ b/gpt4free/quora/graphql/PoeBotEditMutation.graphql diff --git a/quora/graphql/SendMessageMutation.graphql b/gpt4free/quora/graphql/SendMessageMutation.graphql index 4b0a4383..4b0a4383 100644 --- a/quora/graphql/SendMessageMutation.graphql +++ b/gpt4free/quora/graphql/SendMessageMutation.graphql diff --git a/quora/graphql/SendVerificationCodeForLoginMutation.graphql b/gpt4free/quora/graphql/SendVerificationCodeForLoginMutation.graphql index 45af4799..45af4799 100644 --- a/quora/graphql/SendVerificationCodeForLoginMutation.graphql +++ b/gpt4free/quora/graphql/SendVerificationCodeForLoginMutation.graphql diff --git a/quora/graphql/ShareMessagesMutation.graphql b/gpt4free/quora/graphql/ShareMessagesMutation.graphql index 92e80db5..92e80db5 100644 --- a/quora/graphql/ShareMessagesMutation.graphql +++ b/gpt4free/quora/graphql/ShareMessagesMutation.graphql diff --git a/quora/graphql/SignupWithVerificationCodeMutation.graphql b/gpt4free/quora/graphql/SignupWithVerificationCodeMutation.graphql index 06b2826f..06b2826f 100644 --- a/quora/graphql/SignupWithVerificationCodeMutation.graphql +++ b/gpt4free/quora/graphql/SignupWithVerificationCodeMutation.graphql diff --git a/quora/graphql/StaleChatUpdateMutation.graphql b/gpt4free/quora/graphql/StaleChatUpdateMutation.graphql index de203d47..de203d47 100644 --- a/quora/graphql/StaleChatUpdateMutation.graphql +++ b/gpt4free/quora/graphql/StaleChatUpdateMutation.graphql diff --git a/quora/graphql/SubscriptionsMutation.graphql b/gpt4free/quora/graphql/SubscriptionsMutation.graphql index b864bd60..b864bd60 100644 --- a/quora/graphql/SubscriptionsMutation.graphql +++ b/gpt4free/quora/graphql/SubscriptionsMutation.graphql diff --git a/quora/graphql/SummarizePlainPostQuery.graphql b/gpt4free/quora/graphql/SummarizePlainPostQuery.graphql index afa2a84c..afa2a84c 100644 --- a/quora/graphql/SummarizePlainPostQuery.graphql +++ b/gpt4free/quora/graphql/SummarizePlainPostQuery.graphql diff --git a/quora/graphql/SummarizeQuotePostQuery.graphql b/gpt4free/quora/graphql/SummarizeQuotePostQuery.graphql index 5147c3c5..5147c3c5 100644 --- a/quora/graphql/SummarizeQuotePostQuery.graphql +++ b/gpt4free/quora/graphql/SummarizeQuotePostQuery.graphql diff --git a/quora/graphql/SummarizeSharePostQuery.graphql b/gpt4free/quora/graphql/SummarizeSharePostQuery.graphql index cb4a623c..cb4a623c 100644 --- a/quora/graphql/SummarizeSharePostQuery.graphql +++ b/gpt4free/quora/graphql/SummarizeSharePostQuery.graphql diff --git a/quora/graphql/UserSnippetFragment.graphql b/gpt4free/quora/graphql/UserSnippetFragment.graphql index 17fc8426..17fc8426 100644 --- a/quora/graphql/UserSnippetFragment.graphql +++ b/gpt4free/quora/graphql/UserSnippetFragment.graphql diff --git a/quora/graphql/ViewerInfoQuery.graphql b/gpt4free/quora/graphql/ViewerInfoQuery.graphql index 1ecaf9e8..1ecaf9e8 100644 --- a/quora/graphql/ViewerInfoQuery.graphql +++ b/gpt4free/quora/graphql/ViewerInfoQuery.graphql diff --git a/quora/graphql/ViewerStateFragment.graphql b/gpt4free/quora/graphql/ViewerStateFragment.graphql index 3cd83e9c..3cd83e9c 100644 --- a/quora/graphql/ViewerStateFragment.graphql +++ b/gpt4free/quora/graphql/ViewerStateFragment.graphql diff --git a/quora/graphql/ViewerStateUpdatedSubscription.graphql b/gpt4free/quora/graphql/ViewerStateUpdatedSubscription.graphql index 03fc73d1..03fc73d1 100644 --- a/quora/graphql/ViewerStateUpdatedSubscription.graphql +++ b/gpt4free/quora/graphql/ViewerStateUpdatedSubscription.graphql diff --git a/quora/graphql/__init__.py b/gpt4free/quora/graphql/__init__.py index e69de29b..e69de29b 100644 --- a/quora/graphql/__init__.py +++ b/gpt4free/quora/graphql/__init__.py diff --git a/quora/mail.py b/gpt4free/quora/mail.py index e6ce96c2..864d9568 100644 --- a/quora/mail.py +++ b/gpt4free/quora/mail.py @@ -42,9 +42,7 @@ class Emailnator: while True: sleep(2) - mail_token = self.client.post( - "https://www.emailnator.com/message-list", json={"email": self.email} - ) + mail_token = self.client.post("https://www.emailnator.com/message-list", json={"email": self.email}) mail_token = loads(mail_token.text)["messageData"] diff --git a/theb/README.md b/gpt4free/theb/README.md index ca978fce..a4abdf62 100644 --- a/theb/README.md +++ b/gpt4free/theb/README.md @@ -1,9 +1,8 @@ ### Example: `theb` (use like openai pypi package) <a name="example-theb"></a> - ```python # import library -import theb +from gpt4free import theb # simple streaming completion for token in theb.Completion.create('hello world'): diff --git a/theb/__init__.py b/gpt4free/theb/__init__.py index 726e025e..96053877 100644 --- a/theb/__init__.py +++ b/gpt4free/theb/__init__.py @@ -1,8 +1,12 @@ -from re import findall from json import loads from queue import Queue, Empty +from re import findall from threading import Thread +from typing import Generator + from curl_cffi import requests +from fake_useragent import UserAgent + class Completion: # experimental @@ -14,29 +18,29 @@ class Completion: message_queue = Queue() stream_completed = False + @staticmethod def request(prompt: str): headers = { 'authority': 'chatbot.theb.ai', 'content-type': 'application/json', 'origin': 'https://chatbot.theb.ai', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', + 'user-agent': UserAgent().random, } - requests.post('https://chatbot.theb.ai/api/chat-process', headers=headers, - content_callback = Completion.handle_stream_response, - json = { - 'prompt': prompt, - 'options': {} - } + requests.post( + 'https://chatbot.theb.ai/api/chat-process', + headers=headers, + content_callback=Completion.handle_stream_response, + json={'prompt': prompt, 'options': {}}, ) Completion.stream_completed = True @staticmethod - def create(prompt: str): + def create(prompt: str) -> Generator[str, None, None]: Thread(target=Completion.request, args=[prompt]).start() - while Completion.stream_completed != True or not Completion.message_queue.empty(): + while not Completion.stream_completed or not Completion.message_queue.empty(): try: message = Completion.message_queue.get(timeout=0.01) for message in findall(Completion.regex, message): diff --git a/theb/theb_test.py b/gpt4free/theb/theb_test.py index 177c970a..c57d5c62 100644 --- a/theb/theb_test.py +++ b/gpt4free/theb/theb_test.py @@ -1,4 +1,4 @@ import theb for token in theb.Completion.create('hello world'): - print(token, end='', flush=True)
\ No newline at end of file + print(token, end='', flush=True) diff --git a/you/README.md b/gpt4free/you/README.md index 25c20085..11b4723e 100644 --- a/you/README.md +++ b/gpt4free/you/README.md @@ -1,7 +1,8 @@ ### Example: `you` (use like openai pypi package) <a name="example-you"></a> ```python -import you + +from gpt4free import you # simple request with links and details response = you.Completion.create( diff --git a/you/__init__.py b/gpt4free/you/__init__.py index 8bf31f0d..97b48464 100644 --- a/you/__init__.py +++ b/gpt4free/you/__init__.py @@ -1,28 +1,36 @@ +import json import re -from json import loads +from typing import Optional, List, Dict, Any from uuid import uuid4 from fake_useragent import UserAgent +from pydantic import BaseModel from tls_client import Session +class PoeResponse(BaseModel): + text: Optional[str] = None + links: List[str] = [] + extra: Dict[str, Any] = {} + + class Completion: @staticmethod def create( - prompt: str, - page: int = 1, - count: int = 10, - safe_search: str = 'Moderate', - on_shopping_page: bool = False, - mkt: str = '', - response_filter: str = 'WebPages,Translations,TimeZone,Computation,RelatedSearches', - domain: str = 'youchat', - query_trace_id: str = None, - chat: list = None, - include_links: bool = False, - detailed: bool = False, - debug: bool = False, - ) -> dict: + prompt: str, + page: int = 1, + count: int = 10, + safe_search: str = 'Moderate', + on_shopping_page: bool = False, + mkt: str = '', + response_filter: str = 'WebPages,Translations,TimeZone,Computation,RelatedSearches', + domain: str = 'youchat', + query_trace_id: str = None, + chat: list = None, + include_links: bool = False, + detailed: bool = False, + debug: bool = False, + ) -> PoeResponse: if chat is None: chat = [] @@ -57,26 +65,28 @@ class Completion: r'(?<=event: youChatSerpResults\ndata:)(.*\n)*?(?=event: )', response.text ).group() third_party_search_results = re.search( - r'(?<=event: thirdPartySearchResults\ndata:)(.*\n)*?(?=event: )', response.text).group() + r'(?<=event: thirdPartySearchResults\ndata:)(.*\n)*?(?=event: )', response.text + ).group() # slots = findall(r"slots\ndata: (.*)\n\nevent", response.text)[0] text = ''.join(re.findall(r'{\"youChatToken\": \"(.*?)\"}', response.text)) extra = { - 'youChatSerpResults': loads(you_chat_serp_results), + 'youChatSerpResults': json.loads(you_chat_serp_results), # 'slots' : loads(slots) } - return { - 'response': text.replace('\\n', '\n').replace('\\\\', '\\').replace('\\"', '"'), - 'links': loads(third_party_search_results)['search']['third_party_search_results'] - if include_links - else None, - 'extra': extra if detailed else None, - } + response = PoeResponse(text=text.replace('\\n', '\n').replace('\\\\', '\\').replace('\\"', '"')) + if include_links: + response.links = json.loads(third_party_search_results)['search']['third_party_search_results'] - @classmethod - def __get_headers(cls) -> dict: + if detailed: + response.extra = extra + + return response + + @staticmethod + def __get_headers() -> dict: return { 'authority': 'you.com', 'accept': 'text/event-stream', @@ -93,6 +103,6 @@ class Completion: 'user-agent': UserAgent().random, } - @classmethod - def __get_failure_response(cls) -> dict: - return dict(response='Unable to fetch the response, Please try again.', links=[], extra={}) + @staticmethod + def __get_failure_response() -> PoeResponse: + return PoeResponse(text='Unable to fetch the response, Please try again.') diff --git a/gui/README.md b/gui/README.md index 133abe4b..c638c4dc 100644 --- a/gui/README.md +++ b/gui/README.md @@ -1,11 +1,72 @@ # gpt4free gui -mode `streamlit_app.py` into base folder to run +This code provides a Graphical User Interface (GUI) for gpt4free. Users can ask questions and get answers from GPT-4 API's, utilizing multiple API implementations. The project contains two different Streamlit applications: `streamlit_app.py` and `streamlit_chat_app.py`. +Installation +------------ + +1. Clone the repository. +2. Install the required dependencies with: `pip install -r requirements.txt`. +3. To use `streamlit_chat_app.py`, note that it depends on a pull request (PR #24) from the https://github.com/AI-Yash/st-chat/ repository, which may change in the future. The current dependency library can be found at https://github.com/AI-Yash/st-chat/archive/refs/pull/24/head.zip. + +Usage +----- + +Choose one of the Streamlit applications to run: + +### streamlit\_app.py + +This application provides a simple interface for asking GPT-4 questions and receiving answers. + +To run the application: + +run: +```arduino +streamlit run gui/streamlit_app.py +``` +<br> + +<img width="724" alt="image" src="https://user-images.githubusercontent.com/98614666/234232449-0d5cd092-a29d-4759-8197-e00ba712cb1a.png"> + +<br> +<br> preview: + <img width="1125" alt="image" src="https://user-images.githubusercontent.com/98614666/234232398-09e9d3c5-08e6-4b8a-b4f2-0666e9790c7d.png"> -run: -<img width="724" alt="image" src="https://user-images.githubusercontent.com/98614666/234232449-0d5cd092-a29d-4759-8197-e00ba712cb1a.png"> +### streamlit\_chat\_app.py + +This application provides a chat-like interface for asking GPT-4 questions and receiving answers. It supports multiple query methods, and users can select the desired API for their queries. The application also maintains a conversation history. + +To run the application: + +```arduino +streamlit run streamlit_chat_app.py +``` + +<br> + +<img width="724" alt="image" src="image1.png"> + +<br> +<br> + +preview: + +<img width="1125" alt="image" src="image2.png"> + +Contributing +------------ + +Feel free to submit pull requests, report bugs, or request new features by opening issues on the GitHub repository. + +Bug +---- +There is a bug in `streamlit_chat_app.py` right now that I haven't pinpointed yet, probably is really simple but havent had the time to look for it. Whenever you open a new conversation or access an old conversation it will only start prompt-answering after the second time you input to the text input, other than that, everything else seems to work accordingly. + +License +------- + +This project is licensed under the MIT License.
\ No newline at end of file diff --git a/gui/__init__.py b/gui/__init__.py new file mode 100644 index 00000000..e69de29b --- /dev/null +++ b/gui/__init__.py diff --git a/gui/image1.png b/gui/image1.png Binary files differnew file mode 100644 index 00000000..fb84ae7f --- /dev/null +++ b/gui/image1.png diff --git a/gui/image2.png b/gui/image2.png Binary files differnew file mode 100644 index 00000000..c5fc91c9 --- /dev/null +++ b/gui/image2.png diff --git a/gui/query_methods.py b/gui/query_methods.py new file mode 100644 index 00000000..6225453b --- /dev/null +++ b/gui/query_methods.py @@ -0,0 +1,99 @@ +import os +import sys + +sys.path.append(os.path.join(os.path.dirname(__file__), os.path.pardir)) + +from gpt4free import quora, forefront, theb, you +import random + + +def query_forefront(question: str) -> str: + # create an account + token = forefront.Account.create(logging=False) + + response = "" + # get a response + try: + return forefront.Completion.create(token=token, prompt='hello world', model='gpt-4').text + except Exception as e: + # Return error message if an exception occurs + return ( + f'An error occurred: {e}. Please make sure you are using a valid cloudflare clearance token and user agent.' + ) + + +def query_quora(question: str) -> str: + token = quora.Account.create(logging=False, enable_bot_creation=True) + return quora.Completion.create(model='gpt-4', prompt=question, token=token).text + + +def query_theb(question: str) -> str: + # Set cloudflare clearance cookie and get answer from GPT-4 model + response = "" + try: + return ''.join(theb.Completion.create(prompt=question)) + + except Exception as e: + # Return error message if an exception occurs + return ( + f'An error occurred: {e}. Please make sure you are using a valid cloudflare clearance token and user agent.' + ) + + +def query_you(question: str) -> str: + # Set cloudflare clearance cookie and get answer from GPT-4 model + try: + result = you.Completion.create(prompt=question) + return result["response"] + + except Exception as e: + # Return error message if an exception occurs + return ( + f'An error occurred: {e}. Please make sure you are using a valid cloudflare clearance token and user agent.' + ) + + +# Define a dictionary containing all query methods +avail_query_methods = { + "Forefront": query_forefront, + "Poe": query_quora, + "Theb": query_theb, + "You": query_you, + # "Writesonic": query_writesonic, + # "T3nsor": query_t3nsor, + # "Phind": query_phind, + # "Ora": query_ora, +} + + +def query(user_input: str, selected_method: str = "Random") -> str: + # If a specific query method is selected (not "Random") and the method is in the dictionary, try to call it + if selected_method != "Random" and selected_method in avail_query_methods: + try: + return avail_query_methods[selected_method](user_input) + except Exception as e: + print(f"Error with {selected_method}: {e}") + return "😵 Sorry, some error occurred please try again." + + # Initialize variables for determining success and storing the result + success = False + result = "😵 Sorry, some error occurred please try again." + # Create a list of available query methods + query_methods_list = list(avail_query_methods.values()) + + # Continue trying different methods until a successful result is obtained or all methods have been tried + while not success and query_methods_list: + # Choose a random method from the list + chosen_query = random.choice(query_methods_list) + # Find the name of the chosen method + chosen_query_name = [k for k, v in avail_query_methods.items() if v == chosen_query][0] + try: + # Try to call the chosen method with the user input + result = chosen_query(user_input) + success = True + except Exception as e: + print(f"Error with {chosen_query_name}: {e}") + # Remove the failed method from the list of available methods + query_methods_list.remove(chosen_query) + + return result diff --git a/gui/streamlit_app.py b/gui/streamlit_app.py index 8251be2b..2dba0a7b 100644 --- a/gui/streamlit_app.py +++ b/gui/streamlit_app.py @@ -4,19 +4,21 @@ import sys sys.path.append(os.path.join(os.path.dirname(__file__), os.path.pardir)) import streamlit as st -import you +from gpt4free import you + def get_answer(question: str) -> str: # Set cloudflare clearance cookie and get answer from GPT-4 model try: - result = you.Completion.create( - prompt = question) - - return result['response'] - + result = you.Completion.create(prompt=question) + + return result.text + except Exception as e: # Return error message if an exception occurs - return f'An error occurred: {e}. Please make sure you are using a valid cloudflare clearance token and user agent.' + return ( + f'An error occurred: {e}. Please make sure you are using a valid cloudflare clearance token and user agent.' + ) # Set page configuration and add header @@ -27,14 +29,13 @@ st.set_page_config( menu_items={ 'Get Help': 'https://github.com/xtekky/gpt4free/blob/main/README.md', 'Report a bug': "https://github.com/xtekky/gpt4free/issues", - 'About': "### gptfree GUI" - } + 'About': "### gptfree GUI", + }, ) st.header('GPT4free GUI') # Add text area for user input and button to get answer -question_text_area = st.text_area( - '🤖 Ask Any Question :', placeholder='Explain quantum computing in 50 words') +question_text_area = st.text_area('🤖 Ask Any Question :', placeholder='Explain quantum computing in 50 words') if st.button('🧠 Think'): answer = get_answer(question_text_area) escaped = answer.encode('utf-8').decode('unicode-escape') diff --git a/gui/streamlit_chat_app.py b/gui/streamlit_chat_app.py new file mode 100644 index 00000000..68011229 --- /dev/null +++ b/gui/streamlit_chat_app.py @@ -0,0 +1,113 @@ +import atexit +import os +import sys + +sys.path.append(os.path.join(os.path.dirname(__file__), os.path.pardir)) + +import streamlit as st +from streamlit_chat import message +from query_methods import query, avail_query_methods +import pickle + +conversations_file = "conversations.pkl" + + +def load_conversations(): + try: + with open(conversations_file, "rb") as f: + return pickle.load(f) + except FileNotFoundError: + return [] + except EOFError: + return [] + + +def save_conversations(conversations, current_conversation): + updated = False + for i, conversation in enumerate(conversations): + if conversation == current_conversation: + conversations[i] = current_conversation + updated = True + break + if not updated: + conversations.append(current_conversation) + + temp_conversations_file = "temp_" + conversations_file + with open(temp_conversations_file, "wb") as f: + pickle.dump(conversations, f) + + os.replace(temp_conversations_file, conversations_file) + + +def exit_handler(): + print("Exiting, saving data...") + # Perform cleanup operations here, like saving data or closing open files. + save_conversations(st.session_state.conversations, st.session_state.current_conversation) + + +# Register the exit_handler function to be called when the program is closing. +atexit.register(exit_handler) + +st.header("Chat Placeholder") + +if 'conversations' not in st.session_state: + st.session_state['conversations'] = load_conversations() + +if 'input_text' not in st.session_state: + st.session_state['input_text'] = '' + +if 'selected_conversation' not in st.session_state: + st.session_state['selected_conversation'] = None + +if 'input_field_key' not in st.session_state: + st.session_state['input_field_key'] = 0 + +if 'query_method' not in st.session_state: + st.session_state['query_method'] = query + +# Initialize new conversation +if 'current_conversation' not in st.session_state or st.session_state['current_conversation'] is None: + st.session_state['current_conversation'] = {'user_inputs': [], 'generated_responses': []} + +input_placeholder = st.empty() +user_input = input_placeholder.text_input( + 'You:', key=f'input_text_{len(st.session_state["current_conversation"]["user_inputs"])}' +) +submit_button = st.button("Submit") + +if user_input or submit_button: + output = query(user_input, st.session_state['query_method']) + escaped_output = output.encode('utf-8').decode('unicode-escape') + + st.session_state.current_conversation['user_inputs'].append(user_input) + st.session_state.current_conversation['generated_responses'].append(escaped_output) + save_conversations(st.session_state.conversations, st.session_state.current_conversation) + user_input = input_placeholder.text_input( + 'You:', value='', key=f'input_text_{len(st.session_state["current_conversation"]["user_inputs"])}' + ) # Clear the input field + +# Add a button to create a new conversation +if st.sidebar.button("New Conversation"): + st.session_state['selected_conversation'] = None + st.session_state['current_conversation'] = {'user_inputs': [], 'generated_responses': []} + st.session_state['input_field_key'] += 1 + +st.session_state['query_method'] = st.sidebar.selectbox("Select API:", options=avail_query_methods, index=0) + +# Sidebar +st.sidebar.header("Conversation History") + +for i, conversation in enumerate(st.session_state.conversations): + if st.sidebar.button(f"Conversation {i + 1}: {conversation['user_inputs'][0]}", key=f"sidebar_btn_{i}"): + st.session_state['selected_conversation'] = i + st.session_state['current_conversation'] = st.session_state.conversations[i] + +if st.session_state['selected_conversation'] is not None: + conversation_to_display = st.session_state.conversations[st.session_state['selected_conversation']] +else: + conversation_to_display = st.session_state.current_conversation + +if conversation_to_display['generated_responses']: + for i in range(len(conversation_to_display['generated_responses']) - 1, -1, -1): + message(conversation_to_display["generated_responses"][i], key=f"display_generated_{i}") + message(conversation_to_display['user_inputs'][i], is_user=True, key=f"display_user_{i}") diff --git a/openaihosted/__init__.py b/openaihosted/__init__.py deleted file mode 100644 index 4f25a2be..00000000 --- a/openaihosted/__init__.py +++ /dev/null @@ -1,62 +0,0 @@ -import json
-import re
-import requests
-
-
-class Completion:
- @staticmethod
- def create(messages):
- headers = {
- "authority": "openai.a2hosted.com",
- "accept": "text/event-stream",
- "accept-language": "en-US,en;q=0.9,id;q=0.8,ja;q=0.7",
- "cache-control": "no-cache",
- "sec-fetch-dest": "empty",
- "sec-fetch-mode": "cors",
- "sec-fetch-site": "cross-site",
- "user-agent": "Mozilla/5.0 (X11; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/112.0",
- }
-
- query_param = Completion.__create_query_param(messages)
- url = f"https://openai.a2hosted.com/chat?q={query_param}"
- request = requests.get(url, headers=headers, stream=True)
- if request.status_code != 200:
- return Completion.__get_failure_response()
-
- content = request.content
- response = Completion.__join_response(content)
-
- return {"responses": response}
-
- @classmethod
- def __get_failure_response(cls) -> dict:
- return dict(
- response="Unable to fetch the response, Please try again.",
- links=[],
- extra={},
- )
-
- @classmethod
- def __multiple_replace(cls, string, reps) -> str:
- for original, replacement in reps.items():
- string = string.replace(original, replacement)
- return string
-
- @classmethod
- def __create_query_param(cls, conversation) -> str:
- encoded_conversation = json.dumps(conversation)
- replacement = {" ": "%20", '"': "%22", "'": "%27"}
- return Completion.__multiple_replace(encoded_conversation, replacement)
-
- @classmethod
- def __convert_escape_codes(cls, text) -> str:
- replacement = {'\\\\"': '"', '\\"': '"', "\\n": "\n", "\\'": "'"}
- return Completion.__multiple_replace(text, replacement)
-
- @classmethod
- def __join_response(cls, data) -> str:
- data = data.decode("utf-8")
- find_ans = re.findall(r'(?<={"msg":)[^}]*', str(data))
- ans = [Completion.__convert_escape_codes(x[1:-1]) for x in find_ans]
- response = "".join(ans)
- return response
diff --git a/openaihosted/readme.md b/openaihosted/readme.md deleted file mode 100644 index 7b8ced56..00000000 --- a/openaihosted/readme.md +++ /dev/null @@ -1,18 +0,0 @@ -### Example: `openaihosted` (use like openai pypi package) <a name="example-openaihosted"></a> - -```python -import openaihosted - -messages = [{"role": "system", "content": "You are a helpful assistant."}] -while True: - question = input("Question: ") - if question == "!stop": - break - - messages.append({"role": "user", "content": question}) - request = openaihosted.Completion.create(messages=messages) - - response = request["responses"] - messages.append({"role": "assistant", "content": response}) - print(f"Answer: {response}") -``` diff --git a/poetry.lock b/poetry.lock new file mode 100644 index 00000000..d6f1bb67 --- /dev/null +++ b/poetry.lock @@ -0,0 +1,702 @@ +# This file is automatically @generated by Poetry 1.4.2 and should not be changed by hand. + +[[package]] +name = "async-generator" +version = "1.10" +description = "Async generators and context managers for Python 3.5+" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "async_generator-1.10-py3-none-any.whl", hash = "sha256:01c7bf666359b4967d2cda0000cc2e4af16a0ae098cbffcb8472fb9e8ad6585b"}, + {file = "async_generator-1.10.tar.gz", hash = "sha256:6ebb3d106c12920aaae42ccb6f787ef5eefdcdd166ea3d628fa8476abe712144"}, +] + +[[package]] +name = "attrs" +version = "23.1.0" +description = "Classes Without Boilerplate" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "attrs-23.1.0-py3-none-any.whl", hash = "sha256:1f28b4522cdc2fb4256ac1a020c78acf9cba2c6b461ccd2c126f3aa8e8335d04"}, + {file = "attrs-23.1.0.tar.gz", hash = "sha256:6279836d581513a26f1bf235f9acd333bc9115683f14f7e8fae46c98fc50e015"}, +] + +[package.dependencies] +importlib-metadata = {version = "*", markers = "python_version < \"3.8\""} + +[package.extras] +cov = ["attrs[tests]", "coverage[toml] (>=5.3)"] +dev = ["attrs[docs,tests]", "pre-commit"] +docs = ["furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier", "zope-interface"] +tests = ["attrs[tests-no-zope]", "zope-interface"] +tests-no-zope = ["cloudpickle", "hypothesis", "mypy (>=1.1.1)", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"] + +[[package]] +name = "certifi" +version = "2022.12.7" +description = "Python package for providing Mozilla's CA Bundle." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "certifi-2022.12.7-py3-none-any.whl", hash = "sha256:4ad3232f5e926d6718ec31cfc1fcadfde020920e278684144551c91769c7bc18"}, + {file = "certifi-2022.12.7.tar.gz", hash = "sha256:35824b4c3a97115964b408844d64aa14db1cc518f6562e8d7261699d1350a9e3"}, +] + +[[package]] +name = "cffi" +version = "1.15.1" +description = "Foreign Function Interface for Python calling C code." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"}, + {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"}, + {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"}, + {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"}, + {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"}, + {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"}, + {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"}, + {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"}, + {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"}, + {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"}, + {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"}, + {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"}, + {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"}, + {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"}, + {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"}, + {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"}, + {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"}, + {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"}, + {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"}, +] + +[package.dependencies] +pycparser = "*" + +[[package]] +name = "charset-normalizer" +version = "3.1.0" +description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet." +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"}, + {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"}, +] + +[[package]] +name = "colorama" +version = "0.4.6" +description = "Cross-platform colored terminal text." +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7" +files = [ + {file = "colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6"}, + {file = "colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44"}, +] + +[[package]] +name = "curl-cffi" +version = "0.5.5" +description = "libcurl ffi bindings for Python, with impersonation support" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "curl_cffi-0.5.5-cp37-abi3-macosx_10_9_x86_64.whl", hash = "sha256:4322f330167a5c87f6913d32b73eb7da9fe3e3dd86b28f137469f432b346d9bb"}, + {file = "curl_cffi-0.5.5-cp37-abi3-macosx_11_0_arm64.whl", hash = "sha256:48b0dcc6e91d68694e6472fa47b7f3457d8bd24e42c91e15d6e2b650f0d9d206"}, + {file = "curl_cffi-0.5.5-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fda4c35f03ae593b7667d1a09bcd718d1399a5596b936cacb65dcd4bd705e95f"}, + {file = "curl_cffi-0.5.5-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:660b2174b71d86bd7b136a6b91434a4c3edbcb1de718f3d337b688955872fbcc"}, + {file = "curl_cffi-0.5.5-cp37-abi3-win_amd64.whl", hash = "sha256:7adb44515cb165ac661a8e5453c41d75bc284f494921051b64f2889c2c518544"}, + {file = "curl_cffi-0.5.5.tar.gz", hash = "sha256:db94b8d0ad52f3b5c55d32225c29a8219a19592d882075965a78aa9e1a0dead1"}, +] + +[package.dependencies] +cffi = ">=1.12.0" + +[package.extras] +build = ["cibuildwheel", "wheel"] +dev = ["autoflake (==1.4)", "black (==22.8.0)", "coverage (==6.4.1)", "cryptography (==38.0.3)", "flake8 (==6.0.0)", "flake8-bugbear (==22.7.1)", "flake8-pie (==0.15.0)", "httpx (==0.23.1)", "isort (==5.10.1)", "mypy (==0.971)", "pytest (==7.1.2)", "pytest-asyncio (==0.19.0)", "pytest-trio (==0.7.0)", "trio (==0.21.0)", "trio-typing (==0.7.0)", "trustme (==0.9.0)", "types-certifi (==2021.10.8.2)", "uvicorn (==0.18.3)"] +test = ["cryptography (==38.0.3)", "httpx (==0.23.1)", "pytest (==7.1.2)", "pytest-asyncio (==0.19.0)", "pytest-trio (==0.7.0)", "trio (==0.21.0)", "trio-typing (==0.7.0)", "trustme (==0.9.0)", "types-certifi (==2021.10.8.2)", "uvicorn (==0.18.3)"] + +[[package]] +name = "exceptiongroup" +version = "1.1.1" +description = "Backport of PEP 654 (exception groups)" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "exceptiongroup-1.1.1-py3-none-any.whl", hash = "sha256:232c37c63e4f682982c8b6459f33a8981039e5fb8756b2074364e5055c498c9e"}, + {file = "exceptiongroup-1.1.1.tar.gz", hash = "sha256:d484c3090ba2889ae2928419117447a14daf3c1231d5e30d0aae34f354f01785"}, +] + +[package.extras] +test = ["pytest (>=6)"] + +[[package]] +name = "fake-useragent" +version = "1.1.3" +description = "Up-to-date simple useragent faker with real world database" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "fake-useragent-1.1.3.tar.gz", hash = "sha256:1c06f0aa7d6e4894b919b30b9c7ebd72ff497325191057fbb5df3d5db06b93fc"}, + {file = "fake_useragent-1.1.3-py3-none-any.whl", hash = "sha256:695d3b1bf7d11d04ab0f971fb73b0ca8de98b78bbadfbc8bacbc9a48423f7531"}, +] + +[package.dependencies] +importlib-metadata = {version = ">=4.0,<5.0", markers = "python_version < \"3.8\""} +importlib-resources = {version = ">=5.0", markers = "python_version < \"3.10\""} + +[[package]] +name = "h11" +version = "0.14.0" +description = "A pure-Python, bring-your-own-I/O implementation of HTTP/1.1" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "h11-0.14.0-py3-none-any.whl", hash = "sha256:e3fe4ac4b851c468cc8363d500db52c2ead036020723024a109d37346efaa761"}, + {file = "h11-0.14.0.tar.gz", hash = "sha256:8f19fbbe99e72420ff35c00b27a34cb9937e902a8b810e2c88300c6f0a3b699d"}, +] + +[package.dependencies] +typing-extensions = {version = "*", markers = "python_version < \"3.8\""} + +[[package]] +name = "idna" +version = "3.4" +description = "Internationalized Domain Names in Applications (IDNA)" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "idna-3.4-py3-none-any.whl", hash = "sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2"}, + {file = "idna-3.4.tar.gz", hash = "sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4"}, +] + +[[package]] +name = "importlib-metadata" +version = "4.13.0" +description = "Read metadata from Python packages" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "importlib_metadata-4.13.0-py3-none-any.whl", hash = "sha256:8a8a81bcf996e74fee46f0d16bd3eaa382a7eb20fd82445c3ad11f4090334116"}, + {file = "importlib_metadata-4.13.0.tar.gz", hash = "sha256:dd0173e8f150d6815e098fd354f6414b0f079af4644ddfe90c71e2fc6174346d"}, +] + +[package.dependencies] +typing-extensions = {version = ">=3.6.4", markers = "python_version < \"3.8\""} +zipp = ">=0.5" + +[package.extras] +docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker (>=1.9)", "sphinx (>=3.5)"] +perf = ["ipython"] +testing = ["flake8 (<5)", "flufl.flake8", "importlib-resources (>=1.3)", "packaging", "pyfakefs", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)", "pytest-perf (>=0.9.2)"] + +[[package]] +name = "importlib-resources" +version = "5.12.0" +description = "Read resources from Python packages" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "importlib_resources-5.12.0-py3-none-any.whl", hash = "sha256:7b1deeebbf351c7578e09bf2f63fa2ce8b5ffec296e0d349139d43cca061a81a"}, + {file = "importlib_resources-5.12.0.tar.gz", hash = "sha256:4be82589bf5c1d7999aedf2a45159d10cb3ca4f19b2271f8792bc8e6da7b22f6"}, +] + +[package.dependencies] +zipp = {version = ">=3.1.0", markers = "python_version < \"3.10\""} + +[package.extras] +docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-lint"] +testing = ["flake8 (<5)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)"] + +[[package]] +name = "names" +version = "0.3.0" +description = "Generate random names" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "names-0.3.0.tar.gz", hash = "sha256:726e46254f2ed03f1ffb5d941dae3bc67c35123941c29becd02d48d0caa2a671"}, +] + +[[package]] +name = "outcome" +version = "1.2.0" +description = "Capture the outcome of Python function calls." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "outcome-1.2.0-py2.py3-none-any.whl", hash = "sha256:c4ab89a56575d6d38a05aa16daeaa333109c1f96167aba8901ab18b6b5e0f7f5"}, + {file = "outcome-1.2.0.tar.gz", hash = "sha256:6f82bd3de45da303cf1f771ecafa1633750a358436a8bb60e06a1ceb745d2672"}, +] + +[package.dependencies] +attrs = ">=19.2.0" + +[[package]] +name = "pycparser" +version = "2.21" +description = "C parser in Python" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"}, + {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"}, +] + +[[package]] +name = "pydantic" +version = "1.10.7" +description = "Data validation and settings management using python type hints" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pydantic-1.10.7-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e79e999e539872e903767c417c897e729e015872040e56b96e67968c3b918b2d"}, + {file = "pydantic-1.10.7-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:01aea3a42c13f2602b7ecbbea484a98169fb568ebd9e247593ea05f01b884b2e"}, + {file = "pydantic-1.10.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:516f1ed9bc2406a0467dd777afc636c7091d71f214d5e413d64fef45174cfc7a"}, + {file = "pydantic-1.10.7-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae150a63564929c675d7f2303008d88426a0add46efd76c3fc797cd71cb1b46f"}, + {file = "pydantic-1.10.7-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ecbbc51391248116c0a055899e6c3e7ffbb11fb5e2a4cd6f2d0b93272118a209"}, + {file = "pydantic-1.10.7-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:f4a2b50e2b03d5776e7f21af73e2070e1b5c0d0df255a827e7c632962f8315af"}, + {file = "pydantic-1.10.7-cp310-cp310-win_amd64.whl", hash = "sha256:a7cd2251439988b413cb0a985c4ed82b6c6aac382dbaff53ae03c4b23a70e80a"}, + {file = "pydantic-1.10.7-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:68792151e174a4aa9e9fc1b4e653e65a354a2fa0fed169f7b3d09902ad2cb6f1"}, + {file = "pydantic-1.10.7-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:dfe2507b8ef209da71b6fb5f4e597b50c5a34b78d7e857c4f8f3115effaef5fe"}, + {file = "pydantic-1.10.7-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:10a86d8c8db68086f1e30a530f7d5f83eb0685e632e411dbbcf2d5c0150e8dcd"}, + {file = "pydantic-1.10.7-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d75ae19d2a3dbb146b6f324031c24f8a3f52ff5d6a9f22f0683694b3afcb16fb"}, + {file = "pydantic-1.10.7-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:464855a7ff7f2cc2cf537ecc421291b9132aa9c79aef44e917ad711b4a93163b"}, + {file = "pydantic-1.10.7-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:193924c563fae6ddcb71d3f06fa153866423ac1b793a47936656e806b64e24ca"}, + {file = "pydantic-1.10.7-cp311-cp311-win_amd64.whl", hash = "sha256:b4a849d10f211389502059c33332e91327bc154acc1845f375a99eca3afa802d"}, + {file = "pydantic-1.10.7-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:cc1dde4e50a5fc1336ee0581c1612215bc64ed6d28d2c7c6f25d2fe3e7c3e918"}, + {file = "pydantic-1.10.7-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e0cfe895a504c060e5d36b287ee696e2fdad02d89e0d895f83037245218a87fe"}, + {file = "pydantic-1.10.7-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:670bb4683ad1e48b0ecb06f0cfe2178dcf74ff27921cdf1606e527d2617a81ee"}, + {file = "pydantic-1.10.7-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:950ce33857841f9a337ce07ddf46bc84e1c4946d2a3bba18f8280297157a3fd1"}, + {file = "pydantic-1.10.7-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:c15582f9055fbc1bfe50266a19771bbbef33dd28c45e78afbe1996fd70966c2a"}, + {file = "pydantic-1.10.7-cp37-cp37m-win_amd64.whl", hash = "sha256:82dffb306dd20bd5268fd6379bc4bfe75242a9c2b79fec58e1041fbbdb1f7914"}, + {file = "pydantic-1.10.7-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8c7f51861d73e8b9ddcb9916ae7ac39fb52761d9ea0df41128e81e2ba42886cd"}, + {file = "pydantic-1.10.7-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:6434b49c0b03a51021ade5c4daa7d70c98f7a79e95b551201fff682fc1661245"}, + {file = "pydantic-1.10.7-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:64d34ab766fa056df49013bb6e79921a0265204c071984e75a09cbceacbbdd5d"}, + {file = "pydantic-1.10.7-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:701daea9ffe9d26f97b52f1d157e0d4121644f0fcf80b443248434958fd03dc3"}, + {file = "pydantic-1.10.7-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:cf135c46099ff3f919d2150a948ce94b9ce545598ef2c6c7bf55dca98a304b52"}, + {file = "pydantic-1.10.7-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b0f85904f73161817b80781cc150f8b906d521fa11e3cdabae19a581c3606209"}, + {file = "pydantic-1.10.7-cp38-cp38-win_amd64.whl", hash = "sha256:9f6f0fd68d73257ad6685419478c5aece46432f4bdd8d32c7345f1986496171e"}, + {file = "pydantic-1.10.7-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c230c0d8a322276d6e7b88c3f7ce885f9ed16e0910354510e0bae84d54991143"}, + {file = "pydantic-1.10.7-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:976cae77ba6a49d80f461fd8bba183ff7ba79f44aa5cfa82f1346b5626542f8e"}, + {file = "pydantic-1.10.7-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7d45fc99d64af9aaf7e308054a0067fdcd87ffe974f2442312372dfa66e1001d"}, + {file = "pydantic-1.10.7-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d2a5ebb48958754d386195fe9e9c5106f11275867051bf017a8059410e9abf1f"}, + {file = "pydantic-1.10.7-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:abfb7d4a7cd5cc4e1d1887c43503a7c5dd608eadf8bc615413fc498d3e4645cd"}, + {file = "pydantic-1.10.7-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:80b1fab4deb08a8292d15e43a6edccdffa5377a36a4597bb545b93e79c5ff0a5"}, + {file = "pydantic-1.10.7-cp39-cp39-win_amd64.whl", hash = "sha256:d71e69699498b020ea198468e2480a2f1e7433e32a3a99760058c6520e2bea7e"}, + {file = "pydantic-1.10.7-py3-none-any.whl", hash = "sha256:0cd181f1d0b1d00e2b705f1bf1ac7799a2d938cce3376b8007df62b29be3c2c6"}, + {file = "pydantic-1.10.7.tar.gz", hash = "sha256:cfc83c0678b6ba51b0532bea66860617c4cd4251ecf76e9846fa5a9f3454e97e"}, +] + +[package.dependencies] +typing-extensions = ">=4.2.0" + +[package.extras] +dotenv = ["python-dotenv (>=0.10.4)"] +email = ["email-validator (>=1.0.3)"] + +[[package]] +name = "pydub" +version = "0.25.1" +description = "Manipulate audio with an simple and easy high level interface" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"}, + {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"}, +] + +[[package]] +name = "pypasser" +version = "0.0.5" +description = "Bypassing reCaptcha V3 by sending HTTP requests & solving reCaptcha V2 using speech to text." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "PyPasser-0.0.5.tar.gz", hash = "sha256:72b0ded34edcfa885a13ecc825c5a058503b68521ab87294205d7ff5cd569515"}, +] + +[package.dependencies] +pydub = "0.25.1" +PySocks = "1.7.1" +requests = ">=2.25.1,<3.0" +selenium = "*" +SpeechRecognition = "3.8.1" + +[[package]] +name = "pysocks" +version = "1.7.1" +description = "A Python SOCKS client module. See https://github.com/Anorov/PySocks for more information." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "PySocks-1.7.1-py27-none-any.whl", hash = "sha256:08e69f092cc6dbe92a0fdd16eeb9b9ffbc13cadfe5ca4c7bd92ffb078b293299"}, + {file = "PySocks-1.7.1-py3-none-any.whl", hash = "sha256:2725bd0a9925919b9b51739eea5f9e2bae91e83288108a9ad338b2e3a4435ee5"}, + {file = "PySocks-1.7.1.tar.gz", hash = "sha256:3f8804571ebe159c380ac6de37643bb4685970655d3bba243530d6558b799aa0"}, +] + +[[package]] +name = "requests" +version = "2.29.0" +description = "Python HTTP for Humans." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "requests-2.29.0-py3-none-any.whl", hash = "sha256:e8f3c9be120d3333921d213eef078af392fba3933ab7ed2d1cba3b56f2568c3b"}, + {file = "requests-2.29.0.tar.gz", hash = "sha256:f2e34a75f4749019bb0e3effb66683630e4ffeaf75819fb51bebef1bf5aef059"}, +] + +[package.dependencies] +certifi = ">=2017.4.17" +charset-normalizer = ">=2,<4" +idna = ">=2.5,<4" +urllib3 = ">=1.21.1,<1.27" + +[package.extras] +socks = ["PySocks (>=1.5.6,!=1.5.7)"] +use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] + +[[package]] +name = "selenium" +version = "4.9.0" +description = "" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "selenium-4.9.0-py3-none-any.whl", hash = "sha256:4c19e6aac202719373108d53a5a8e9336ba8d2b25822ca32ae6ff37acbabbdbe"}, + {file = "selenium-4.9.0.tar.gz", hash = "sha256:478fae77cdfaec32adb1e68d59632c8c191f920535282abcaa2d1a3d98655624"}, +] + +[package.dependencies] +certifi = ">=2021.10.8" +trio = ">=0.17,<1.0" +trio-websocket = ">=0.9,<1.0" +urllib3 = {version = ">=1.26,<2.0", extras = ["socks"]} + +[[package]] +name = "sniffio" +version = "1.3.0" +description = "Sniff out which async library your code is running under" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "sniffio-1.3.0-py3-none-any.whl", hash = "sha256:eecefdce1e5bbfb7ad2eeaabf7c1eeb404d7757c379bd1f7e5cce9d8bf425384"}, + {file = "sniffio-1.3.0.tar.gz", hash = "sha256:e60305c5e5d314f5389259b7f22aaa33d8f7dee49763119234af3755c55b9101"}, +] + +[[package]] +name = "sortedcontainers" +version = "2.4.0" +description = "Sorted Containers -- Sorted List, Sorted Dict, Sorted Set" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0"}, + {file = "sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88"}, +] + +[[package]] +name = "speechrecognition" +version = "3.8.1" +description = "Library for performing speech recognition, with support for several engines and APIs, online and offline." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "SpeechRecognition-3.8.1-py2.py3-none-any.whl", hash = "sha256:4d8f73a0c05ec70331c3bacaa89ecc06dfa8d9aba0899276664cda06ab597e8e"}, +] + +[[package]] +name = "tls-client" +version = "0.2" +description = "Advanced Python HTTP Client." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "tls_client-0.2-py3-none-any.whl", hash = "sha256:26012084e753d7531d32960ec706f81188bb7d825586675fa300b2b44f791412"}, + {file = "tls_client-0.2.tar.gz", hash = "sha256:eef3860c6f186fa866dc782f1b9e43ae837e40e831f50831c3515cee7c84fd0f"}, +] + +[[package]] +name = "trio" +version = "0.22.0" +description = "A friendly Python library for async concurrency and I/O" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "trio-0.22.0-py3-none-any.whl", hash = "sha256:f1dd0780a89bfc880c7c7994519cb53f62aacb2c25ff487001c0052bd721cdf0"}, + {file = "trio-0.22.0.tar.gz", hash = "sha256:ce68f1c5400a47b137c5a4de72c7c901bd4e7a24fbdebfe9b41de8c6c04eaacf"}, +] + +[package.dependencies] +async-generator = ">=1.9" +attrs = ">=19.2.0" +cffi = {version = ">=1.14", markers = "os_name == \"nt\" and implementation_name != \"pypy\""} +exceptiongroup = {version = ">=1.0.0rc9", markers = "python_version < \"3.11\""} +idna = "*" +outcome = "*" +sniffio = "*" +sortedcontainers = "*" + +[[package]] +name = "trio-websocket" +version = "0.10.2" +description = "WebSocket library for Trio" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "trio-websocket-0.10.2.tar.gz", hash = "sha256:af13e9393f9051111300287947ec595d601758ce3d165328e7d36325135a8d62"}, + {file = "trio_websocket-0.10.2-py3-none-any.whl", hash = "sha256:0908435e4eecc49d830ae1c4d6c47b978a75f00594a2be2104d58b61a04cdb53"}, +] + +[package.dependencies] +exceptiongroup = "*" +trio = ">=0.11" +wsproto = ">=0.14" + +[[package]] +name = "twocaptcha" +version = "0.0.1" +description = "2Captcha Python3 API Wrapper" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "TwoCaptcha-0.0.1.tar.gz", hash = "sha256:fd04127de71ca4bd31c22add84a5bcb7c683cf9ee5bf503ca14a8f372ac76a0e"}, +] + +[package.dependencies] +requests = "*" + +[[package]] +name = "typing-extensions" +version = "4.5.0" +description = "Backported and Experimental Type Hints for Python 3.7+" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"}, + {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"}, +] + +[[package]] +name = "urllib3" +version = "1.26.15" +description = "HTTP library with thread-safe connection pooling, file post, and more." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*, !=3.5.*" +files = [ + {file = "urllib3-1.26.15-py2.py3-none-any.whl", hash = "sha256:aa751d169e23c7479ce47a0cb0da579e3ede798f994f5816a74e4f4500dcea42"}, + {file = "urllib3-1.26.15.tar.gz", hash = "sha256:8a388717b9476f934a21484e8c8e61875ab60644d29b9b39e11e4b9dc1c6b305"}, +] + +[package.dependencies] +PySocks = {version = ">=1.5.6,<1.5.7 || >1.5.7,<2.0", optional = true, markers = "extra == \"socks\""} + +[package.extras] +brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)", "brotlipy (>=0.6.0)"] +secure = ["certifi", "cryptography (>=1.3.4)", "idna (>=2.0.0)", "ipaddress", "pyOpenSSL (>=0.14)", "urllib3-secure-extra"] +socks = ["PySocks (>=1.5.6,!=1.5.7,<2.0)"] + +[[package]] +name = "websocket-client" +version = "1.5.1" +description = "WebSocket client for Python with low level API options" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "websocket-client-1.5.1.tar.gz", hash = "sha256:3f09e6d8230892547132177f575a4e3e73cfdf06526e20cc02aa1c3b47184d40"}, + {file = "websocket_client-1.5.1-py3-none-any.whl", hash = "sha256:cdf5877568b7e83aa7cf2244ab56a3213de587bbe0ce9d8b9600fc77b455d89e"}, +] + +[package.extras] +docs = ["Sphinx (>=3.4)", "sphinx-rtd-theme (>=0.5)"] +optional = ["python-socks", "wsaccel"] +test = ["websockets"] + +[[package]] +name = "wsproto" +version = "1.2.0" +description = "WebSockets state-machine based protocol implementation" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "wsproto-1.2.0-py3-none-any.whl", hash = "sha256:b9acddd652b585d75b20477888c56642fdade28bdfd3579aa24a4d2c037dd736"}, + {file = "wsproto-1.2.0.tar.gz", hash = "sha256:ad565f26ecb92588a3e43bc3d96164de84cd9902482b130d0ddbaa9664a85065"}, +] + +[package.dependencies] +h11 = ">=0.9.0,<1" + +[[package]] +name = "zipp" +version = "3.15.0" +description = "Backport of pathlib-compatible object wrapper for zip files" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "zipp-3.15.0-py3-none-any.whl", hash = "sha256:48904fc76a60e542af151aded95726c1a5c34ed43ab4134b597665c86d7ad556"}, + {file = "zipp-3.15.0.tar.gz", hash = "sha256:112929ad649da941c23de50f356a2b5570c954b65150642bccdd66bf194d224b"}, +] + +[package.extras] +docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-lint"] +testing = ["big-O", "flake8 (<5)", "jaraco.functools", "jaraco.itertools", "more-itertools", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)"] + +[metadata] +lock-version = "2.0" +python-versions = "^3.7" +content-hash = "d22aa72ee6bd554c0676a3b6b723090d156a720d03d3b05422a01aa9bf22dda2" diff --git a/pyproject.toml b/pyproject.toml new file mode 100644 index 00000000..83df5dc5 --- /dev/null +++ b/pyproject.toml @@ -0,0 +1,28 @@ +[tool.poetry] +name = "openai-rev" +version = "0.1.0" +description = "" +authors = [] +license = "GPL-3.0" +readme = "README.md" +packages = [{ include = "gpt4free" }] +exclude = ["**/*.txt"] + +[tool.poetry.dependencies] +python = "^3.7" +websocket-client = "^1.5.1" +requests = "2.29.0" +tls-client = "^0.2" +pypasser = "^0.0.5" +names = "^0.3.0" +colorama = "^0.4.6" +curl-cffi = "^0.5.5" +selenium = "^4.9.0" +fake-useragent = "^1.1.3" +twocaptcha = "^0.0.1" +pydantic = "^1.10.7" + + +[build-system] +requires = ["poetry-core"] +build-backend = "poetry.core.masonry.api" diff --git a/requirements.txt b/requirements.txt index a5673119..d30b29aa 100644 --- a/requirements.txt +++ b/requirements.txt @@ -9,3 +9,5 @@ streamlit==1.21.0 selenium fake-useragent twocaptcha +https://github.com/AI-Yash/st-chat/archive/refs/pull/24/head.zip +pydantic diff --git a/test.py b/test.py new file mode 100644 index 00000000..0fd2ec8b --- /dev/null +++ b/test.py @@ -0,0 +1,5 @@ +from gpt4free import theb + +for token in theb.Completion.create('hello world'): + print(token, end='', flush=True) + print('asdsos')
\ No newline at end of file diff --git a/testing/forefront_test.py b/testing/forefront_test.py index 4d5f4bc1..b7b5c57c 100644 --- a/testing/forefront_test.py +++ b/testing/forefront_test.py @@ -1,10 +1,9 @@ -import forefront +from gpt4free import forefront # create an account token = forefront.Account.create(logging=True) print(token) # get a response -for response in forefront.StreamingCompletion.create(token=token, - prompt='hello world', model='gpt-4'): - print(response.completion.choices[0].text, end='') +for response in forefront.StreamingCompletion.create(token=token, prompt='hello world', model='gpt-4'): + print(response.text, end='') diff --git a/testing/poe_account_create_test.py b/testing/poe_account_create_test.py index 5d435b1f..ace2306e 100644 --- a/testing/poe_account_create_test.py +++ b/testing/poe_account_create_test.py @@ -1,27 +1,28 @@ from hashlib import md5 from json import dumps from re import findall +from typing import Optional from tls_client import Session as TLS from twocaptcha import TwoCaptcha -from quora import extract_formkey -from quora.mail import Emailnator +from gpt4free.quora import extract_formkey +from gpt4free.quora.mail import Emailnator solver = TwoCaptcha('72747bf24a9d89b4dcc1b24875efd358') class Account: - def create(proxy: None or str = None, logging: bool = False, enable_bot_creation: bool = False): + @staticmethod + def create(proxy: Optional[str] = None, logging: bool = False, enable_bot_creation: bool = False): client = TLS(client_identifier='chrome110') - client.proxies = { - 'http': f'http://{proxy}', - 'https': f'http://{proxy}'} if proxy else None + client.proxies = {'http': f'http://{proxy}', 'https': f'http://{proxy}'} if proxy else None mail_client = Emailnator() mail_address = mail_client.get_mail() - if logging: print('email', mail_address) + if logging: + print('email', mail_address) client.headers = { 'authority': 'poe.com', @@ -39,29 +40,30 @@ class Account: 'sec-fetch-dest': 'empty', 'sec-fetch-mode': 'cors', 'sec-fetch-site': 'same-origin', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36' + 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', } client.headers["poe-formkey"] = extract_formkey(client.get('https://poe.com/login').text) client.headers["poe-tchannel"] = client.get('https://poe.com/api/settings').json()['tchannelData']['channel'] # token = reCaptchaV3('https://www.recaptcha.net/recaptcha/enterprise/anchor?ar=1&k=6LflhEElAAAAAI_ewVwRWI9hsyV4mbZnYAslSvlG&co=aHR0cHM6Ly9wb2UuY29tOjQ0Mw..&hl=en&v=4PnKmGB9wRHh1i04o7YUICeI&size=invisible&cb=bi6ivxoskyal') - token = solver.recaptcha(sitekey='6LflhEElAAAAAI_ewVwRWI9hsyV4mbZnYAslSvlG', - url='https://poe.com/login?redirect_url=%2F', - version='v3', - enterprise=1, - invisible=1, - action='login', )['code'] - - payload = dumps(separators=(',', ':'), obj={ - 'queryName': 'MainSignupLoginSection_sendVerificationCodeMutation_Mutation', - 'variables': { - 'emailAddress': mail_address, - 'phoneNumber': None, - 'recaptchaToken': token + token = solver.recaptcha( + sitekey='6LflhEElAAAAAI_ewVwRWI9hsyV4mbZnYAslSvlG', + url='https://poe.com/login?redirect_url=%2F', + version='v3', + enterprise=1, + invisible=1, + action='login', + )['code'] + + payload = dumps( + separators=(',', ':'), + obj={ + 'queryName': 'MainSignupLoginSection_sendVerificationCodeMutation_Mutation', + 'variables': {'emailAddress': mail_address, 'phoneNumber': None, 'recaptchaToken': token}, + 'query': 'mutation MainSignupLoginSection_sendVerificationCodeMutation_Mutation(\n $emailAddress: String\n $phoneNumber: String\n $recaptchaToken: String\n) {\n sendVerificationCode(verificationReason: login, emailAddress: $emailAddress, phoneNumber: $phoneNumber, recaptchaToken: $recaptchaToken) {\n status\n errorMessage\n }\n}\n', }, - 'query': 'mutation MainSignupLoginSection_sendVerificationCodeMutation_Mutation(\n $emailAddress: String\n $phoneNumber: String\n $recaptchaToken: String\n) {\n sendVerificationCode(verificationReason: login, emailAddress: $emailAddress, phoneNumber: $phoneNumber, recaptchaToken: $recaptchaToken) {\n status\n errorMessage\n }\n}\n', - }) + ) base_string = payload + client.headers["poe-formkey"] + 'WpuLMiXEKKE98j56k' client.headers["poe-tag-id"] = md5(base_string.encode()).hexdigest() @@ -74,31 +76,34 @@ class Account: print('please try using a proxy / wait for fix') if 'Bad Request' in response.text: - if logging: print('bad request, retrying...', response.json()) + if logging: + print('bad request, retrying...', response.json()) quit() - if logging: print('send_code', response.json()) + if logging: + print('send_code', response.json()) mail_content = mail_client.get_message() mail_token = findall(r';">(\d{6,7})</div>', mail_content)[0] - if logging: print('code', mail_token) + if logging: + print('code', mail_token) - payload = dumps(separators=(',', ':'), obj={ - "queryName": "SignupOrLoginWithCodeSection_signupWithVerificationCodeMutation_Mutation", - "variables": { - "verificationCode": str(mail_token), - "emailAddress": mail_address, - "phoneNumber": None + payload = dumps( + separators=(',', ':'), + obj={ + "queryName": "SignupOrLoginWithCodeSection_signupWithVerificationCodeMutation_Mutation", + "variables": {"verificationCode": str(mail_token), "emailAddress": mail_address, "phoneNumber": None}, + "query": "mutation SignupOrLoginWithCodeSection_signupWithVerificationCodeMutation_Mutation(\n $verificationCode: String!\n $emailAddress: String\n $phoneNumber: String\n) {\n signupWithVerificationCode(verificationCode: $verificationCode, emailAddress: $emailAddress, phoneNumber: $phoneNumber) {\n status\n errorMessage\n }\n}\n", }, - "query": "mutation SignupOrLoginWithCodeSection_signupWithVerificationCodeMutation_Mutation(\n $verificationCode: String!\n $emailAddress: String\n $phoneNumber: String\n) {\n signupWithVerificationCode(verificationCode: $verificationCode, emailAddress: $emailAddress, phoneNumber: $phoneNumber) {\n status\n errorMessage\n }\n}\n" - }) + ) base_string = payload + client.headers["poe-formkey"] + 'WpuLMiXEKKE98j56k' client.headers["poe-tag-id"] = md5(base_string.encode()).hexdigest() response = client.post('https://poe.com/api/gql_POST', data=payload) - if logging: print('verify_code', response.json()) + if logging: + print('verify_code', response.json()) Account.create(proxy='xtekky:wegwgwegwed_streaming-1@geo.iproyal.com:12321', logging=True) diff --git a/testing/poe_test.py b/testing/poe_test.py index 8d527879..22d95f5f 100644 --- a/testing/poe_test.py +++ b/testing/poe_test.py @@ -1,13 +1,11 @@ from time import sleep -import quora +from gpt4free import quora token = quora.Account.create(proxy=None, logging=True) print('token', token) sleep(2) -for response in quora.StreamingCompletion.create(model='gpt-3.5-turbo', - prompt='hello world', - token=token): - print(response.completion.choices[0].text, end="", flush=True) +for response in quora.StreamingCompletion.create(model='ChatGPT', prompt='hello world', token=token): + print(response.text, flush=True) diff --git a/testing/quora_test_2.py b/testing/quora_test_2.py index d5316946..297ca7a1 100644 --- a/testing/quora_test_2.py +++ b/testing/quora_test_2.py @@ -1,17 +1,12 @@ -import quora +from gpt4free import quora token = quora.Account.create(logging=True, enable_bot_creation=True) model = quora.Model.create( - token=token, - model='gpt-3.5-turbo', # or claude-instant-v1.0 - system_prompt='you are ChatGPT a large language model ...' + token=token, model='ChatGPT', system_prompt='you are ChatGPT a large language model ...' # or claude-instant-v1.0 ) print(model.name) -for response in quora.StreamingCompletion.create( - custom_model=model.name, - prompt='hello world', - token=token): - print(response.completion.choices[0].text) +for response in quora.StreamingCompletion.create(custom_model=model.name, prompt='hello world', token=token): + print(response.text) diff --git a/testing/sqlchat_test.py b/testing/sqlchat_test.py index 577d85a6..1db71be2 100644 --- a/testing/sqlchat_test.py +++ b/testing/sqlchat_test.py @@ -1,6 +1,4 @@ import sqlchat -for response in sqlchat.StreamCompletion.create( - prompt='write python code to reverse a string', - messages=[]): +for response in sqlchat.StreamCompletion.create(prompt='write python code to reverse a string', messages=[]): print(response.completion.choices[0].text, end='') diff --git a/testing/t3nsor_test.py b/testing/t3nsor_test.py index 1506a1b4..6d36400d 100644 --- a/testing/t3nsor_test.py +++ b/testing/t3nsor_test.py @@ -1,6 +1,4 @@ import t3nsor -for response in t3nsor.StreamCompletion.create( - prompt='write python code to reverse a string', - messages=[]): +for response in t3nsor.StreamCompletion.create(prompt='write python code to reverse a string', messages=[]): print(response.completion.choices[0].text) diff --git a/testing/test_main.py b/testing/test_main.py new file mode 100644 index 00000000..7c28f1d2 --- /dev/null +++ b/testing/test_main.py @@ -0,0 +1,27 @@ +import gpt4free +from gpt4free import Provider, quora, forefront + +# usage You +response = gpt4free.Completion.create(Provider.You, prompt='Write a poem on Lionel Messi') +print(response) + +# usage Poe +token = quora.Account.create(logging=False) +response = gpt4free.Completion.create(Provider.Poe, prompt='Write a poem on Lionel Messi', token=token, model='ChatGPT') +print(response) + +# usage forefront +token = forefront.Account.create(logging=False) +response = gpt4free.Completion.create( + Provider.ForeFront, prompt='Write a poem on Lionel Messi', model='gpt-4', token=token +) +print(response) +print(f'END') + +# usage theb +response = gpt4free.Completion.create(Provider.Theb, prompt='Write a poem on Lionel Messi') +print(response) + +# usage cocalc +response = gpt4free.Completion.create(Provider.CoCalc, prompt='Write a poem on Lionel Messi', cookie_input='') +print(response) diff --git a/testing/writesonic_test.py b/testing/writesonic_test.py index 5c68bbe0..aff847f0 100644 --- a/testing/writesonic_test.py +++ b/testing/writesonic_test.py @@ -4,17 +4,14 @@ import writesonic # create account (3-4s) account = writesonic.Account.create(logging=True) -# with loging: +# with loging: # 2023-04-06 21:50:25 INFO __main__ -> register success : '{"id":"51aa0809-3053-44f7-922a...' (2s) # 2023-04-06 21:50:25 INFO __main__ -> id : '51aa0809-3053-44f7-922a-2b85d8d07edf' # 2023-04-06 21:50:25 INFO __main__ -> token : 'eyJhbGciOiJIUzI1NiIsInR5cCI6Ik...' # 2023-04-06 21:50:28 INFO __main__ -> got key : '194158c4-d249-4be0-82c6-5049e869533c' (2s) # simple completion -response = writesonic.Completion.create( - api_key=account.key, - prompt='hello world' -) +response = writesonic.Completion.create(api_key=account.key, prompt='hello world') print(response.completion.choices[0].text) # Hello! How may I assist you today? @@ -24,16 +21,7 @@ response = writesonic.Completion.create( api_key=account.key, prompt='what is my name ?', enable_memory=True, - history_data=[ - { - 'is_sent': True, - 'message': 'my name is Tekky' - }, - { - 'is_sent': False, - 'message': 'hello Tekky' - } - ] + history_data=[{'is_sent': True, 'message': 'my name is Tekky'}, {'is_sent': False, 'message': 'hello Tekky'}], ) print(response.completion.choices[0].text) # Your name is Tekky. @@ -41,9 +29,7 @@ print(response.completion.choices[0].text) # Your name is Tekky. # enable internet response = writesonic.Completion.create( - api_key=account.key, - prompt='who won the quatar world cup ?', - enable_google_results=True + api_key=account.key, prompt='who won the quatar world cup ?', enable_google_results=True ) print(response.completion.choices[0].text) # Argentina won the 2022 FIFA World Cup tournament held in Qatar ... diff --git a/testing/you_test.py b/testing/you_test.py index 6ba12f57..1e9f6205 100644 --- a/testing/you_test.py +++ b/testing/you_test.py @@ -1,4 +1,4 @@ -import you +from gpt4free import you # simple request with links and details response = you.Completion.create(prompt="hello world", detailed=True, include_links=True) @@ -22,6 +22,6 @@ while True: response = you.Completion.create(prompt=prompt, chat=chat) - print("Bot:", response["response"]) + print("Bot:", response.text) - chat.append({"question": prompt, "answer": response["response"]}) + chat.append({"question": prompt, "answer": response.text}) diff --git a/unfinished/chatpdf/__init__.py b/unfinished/chatpdf/__init__.py new file mode 100644 index 00000000..4c9d2d3e --- /dev/null +++ b/unfinished/chatpdf/__init__.py @@ -0,0 +1,75 @@ +import requests +import json + +class Completion: + + def request(prompt: str): + '''TODO: some sort of authentication + upload PDF from URL or local file + Then you should get the atoken and chat ID + ''' + + token = "your_token_here" + chat_id = "your_chat_id_here" + + url = "https://chat-pr4yueoqha-ue.a.run.app/" + + payload = json.dumps({ + "v": 2, + "chatSession": { + "type": "join", + "chatId": chat_id + }, + "history": [ + { + "id": "VNsSyJIq_0", + "author": "p_if2GPSfyN8hjDoA7unYe", + "msg": "<start>", + "time": 1682672009270 + }, + { + "id": "Zk8DRUtx_6", + "author": "uplaceholder", + "msg": prompt, + "time": 1682672181339 + } + ] + }) + + # TODO: fix headers, use random user-agent, streaming response, etc + headers = { + 'authority': 'chat-pr4yueoqha-ue.a.run.app', + 'accept': '*/*', + 'accept-language': 'en-US,en;q=0.9', + 'atoken': token, + 'content-type': 'application/json', + 'origin': 'https://www.chatpdf.com', + 'referer': 'https://www.chatpdf.com/', + 'sec-ch-ua': '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"', + 'sec-ch-ua-mobile': '?0', + 'sec-ch-ua-platform': '"Windows"', + 'sec-fetch-dest': 'empty', + 'sec-fetch-mode': 'cors', + 'sec-fetch-site': 'cross-site', + 'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36' + } + + response = requests.request("POST", url, headers=headers, data=payload).text + Completion.stream_completed = True + return {'response': response} + + @staticmethod + def create(prompt: str): + Thread(target=Completion.request, args=[prompt]).start() + + while Completion.stream_completed != True or not Completion.message_queue.empty(): + try: + message = Completion.message_queue.get(timeout=0.01) + for message in findall(Completion.regex, message): + yield loads(Completion.part1 + message + Completion.part2)['delta'] + + except Empty: + pass + + @staticmethod + def handle_stream_response(response): + Completion.message_queue.put(response.decode()) diff --git a/unfinished/openai/README.md b/unfinished/openai/README.md deleted file mode 100644 index 67e8645c..00000000 --- a/unfinished/openai/README.md +++ /dev/null @@ -1,2 +0,0 @@ -to do: -- code refractoring
\ No newline at end of file diff --git a/unfinished/openai/__ini__.py b/unfinished/openai/__ini__.py deleted file mode 100644 index f0894e1b..00000000 --- a/unfinished/openai/__ini__.py +++ /dev/null @@ -1,87 +0,0 @@ -# Import required libraries -from uuid import uuid4 - -from browser_cookie3 import chrome -from tls_client import Session - - -class OpenAIChat: - def __init__(self): - self.client = Session(client_identifier='chrome110') - self._load_cookies() - self._set_headers() - - def _load_cookies(self): - # Load cookies for the specified domain - for cookie in chrome(domain_name='chat.openai.com'): - self.client.cookies[cookie.name] = cookie.value - - def _set_headers(self): - # Set headers for the client - self.client.headers = { - 'authority': 'chat.openai.com', - 'accept': 'text/event-stream', - 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', - 'authorization': 'Bearer ' + self.session_auth()['accessToken'], - 'cache-control': 'no-cache', - 'content-type': 'application/json', - 'origin': 'https://chat.openai.com', - 'pragma': 'no-cache', - 'referer': 'https://chat.openai.com/chat', - 'sec-ch-ua': '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"', - 'sec-ch-ua-mobile': '?0', - 'sec-ch-ua-platform': '"macOS"', - 'sec-fetch-dest': 'empty', - 'sec-fetch-mode': 'cors', - 'sec-fetch-site': 'same-origin', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', - } - - def session_auth(self): - headers = { - 'authority': 'chat.openai.com', - 'accept': '*/*', - 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', - 'cache-control': 'no-cache', - 'pragma': 'no-cache', - 'referer': 'https://chat.openai.com/chat', - 'sec-ch-ua': '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"', - 'sec-ch-ua-mobile': '?0', - 'sec-ch-ua-platform': '"macOS"', - 'sec-fetch-dest': 'empty', - 'sec-fetch-mode': 'cors', - 'sec-fetch-site': 'same-origin', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', - } - - return self.client.get('https://chat.openai.com/api/auth/session', headers=headers).json() - - def send_message(self, message): - response = self.client.post('https://chat.openai.com/backend-api/conversation', json={ - 'action': 'next', - 'messages': [ - { - 'id': str(uuid4()), - 'author': { - 'role': 'user', - }, - 'content': { - 'content_type': 'text', - 'parts': [ - message, - ], - }, - }, - ], - 'parent_message_id': '9b4682f7-977c-4c8a-b5e6-9713e73dfe01', - 'model': 'text-davinci-002-render-sha', - 'timezone_offset_min': -120, - }) - - return response.text - - -if __name__ == "__main__": - chat = OpenAIChat() - response = chat.send_message("hello world") - print(response) diff --git a/unfinished/vercelai/__init__.py b/unfinished/vercelai/__init__.py new file mode 100644 index 00000000..1dcb5b39 --- /dev/null +++ b/unfinished/vercelai/__init__.py @@ -0,0 +1,41 @@ +import requests + +class Completion: + def create(prompt: str, + model: str = 'openai:gpt-3.5-turbo', + temperature: float = 0.7, + max_tokens: int = 200, + top_p: float = 1, + top_k: int = 1, + frequency_penalty: float = 1, + presence_penalty: float = 1, + stopSequences: list = []): + + token = requests.get('https://play.vercel.ai/openai.jpeg', headers={ + 'authority': 'play.vercel.ai', + 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', + 'referer': 'https://play.vercel.ai/', + 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36'}).text.replace('=','') + + print(token) + + headers = { + 'authority': 'play.vercel.ai', + 'custom-encoding': token, + 'origin': 'https://play.vercel.ai', + 'referer': 'https://play.vercel.ai/', + 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36' + } + + for chunk in requests.post('https://play.vercel.ai/api/generate', headers=headers, stream=True, json={ + 'prompt': prompt, + 'model': model, + 'temperature': temperature, + 'maxTokens': max_tokens, + 'topK': top_p, + 'topP': top_k, + 'frequencyPenalty': frequency_penalty, + 'presencePenalty': presence_penalty, + 'stopSequences': stopSequences}).iter_lines(): + + yield (chunk)
\ No newline at end of file diff --git a/unfinished/vercelai/test.js b/unfinished/vercelai/test.js new file mode 100644 index 00000000..0f822cfd --- /dev/null +++ b/unfinished/vercelai/test.js @@ -0,0 +1,33 @@ +(async () => { + + let response = await fetch("https://play.vercel.ai/openai.jpeg", { + "headers": { + "accept": "*/*", + "accept-language": "en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3", + "sec-ch-ua": "\"Chromium\";v=\"112\", \"Google Chrome\";v=\"112\", \"Not:A-Brand\";v=\"99\"", + "sec-ch-ua-mobile": "?0", + "sec-ch-ua-platform": "\"macOS\"", + "sec-fetch-dest": "empty", + "sec-fetch-mode": "cors", + "sec-fetch-site": "same-origin" + }, + "referrer": "https://play.vercel.ai/", + "referrerPolicy": "strict-origin-when-cross-origin", + "body": null, + "method": "GET", + "mode": "cors", + "credentials": "omit" + }); + + + let data = JSON.parse(atob(await response.text())) + let ret = eval("(".concat(data.c, ")(data.a)")); + + botPreventionToken = btoa(JSON.stringify({ + r: ret, + t: data.t + })) + + console.log(botPreventionToken); + +})()
\ No newline at end of file diff --git a/unfinished/vercelai/test.py b/unfinished/vercelai/test.py new file mode 100644 index 00000000..318e71c3 --- /dev/null +++ b/unfinished/vercelai/test.py @@ -0,0 +1,67 @@ +import requests +from base64 import b64decode, b64encode +from json import loads +from json import dumps + +headers = { + 'Accept': '*/*', + 'Accept-Language': 'en-GB,en-US;q=0.9,en;q=0.8', + 'Connection': 'keep-alive', + 'Referer': 'https://play.vercel.ai/', + 'Sec-Fetch-Dest': 'empty', + 'Sec-Fetch-Mode': 'cors', + 'Sec-Fetch-Site': 'same-origin', + 'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/110.0.0.0 Safari/537.36', + 'sec-ch-ua': '"Chromium";v="110", "Google Chrome";v="110", "Not:A-Brand";v="99"', + 'sec-ch-ua-mobile': '?0', + 'sec-ch-ua-platform': '"macOS"', +} + +response = requests.get('https://play.vercel.ai/openai.jpeg', headers=headers) + +token_data = loads(b64decode(response.text)) +print(token_data) + +raw_token = { + 'a': token_data['a'] * .1 * .2, + 't': token_data['t'] +} + +print(raw_token) + +new_token = b64encode(dumps(raw_token, separators=(',', ':')).encode()).decode() +print(new_token) + +import requests + +headers = { + 'authority': 'play.vercel.ai', + 'accept': '*/*', + 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', + 'content-type': 'application/json', + 'custom-encoding': new_token, + 'origin': 'https://play.vercel.ai', + 'referer': 'https://play.vercel.ai/', + 'sec-ch-ua': '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"', + 'sec-ch-ua-mobile': '?0', + 'sec-ch-ua-platform': '"macOS"', + 'sec-fetch-dest': 'empty', + 'sec-fetch-mode': 'cors', + 'sec-fetch-site': 'same-origin', + 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', +} + +json_data = { + 'prompt': 'hello\n', + 'model': 'openai:gpt-3.5-turbo', + 'temperature': 0.7, + 'maxTokens': 200, + 'topK': 1, + 'topP': 1, + 'frequencyPenalty': 1, + 'presencePenalty': 1, + 'stopSequences': [], +} + +response = requests.post('https://play.vercel.ai/api/generate', headers=headers, json=json_data) +print(response.text)
\ No newline at end of file diff --git a/unfinished/vercelai/token.py b/unfinished/vercelai/token.py new file mode 100644 index 00000000..e69de29b --- /dev/null +++ b/unfinished/vercelai/token.py diff --git a/unfinished/vercelai/v2.py b/unfinished/vercelai/v2.py deleted file mode 100644 index 176ee342..00000000 --- a/unfinished/vercelai/v2.py +++ /dev/null @@ -1,27 +0,0 @@ -import requests - -token = requests.get('https://play.vercel.ai/openai.jpeg', headers={ - 'authority': 'play.vercel.ai', - 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', - 'referer': 'https://play.vercel.ai/', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36'}).text + '.' - -headers = { - 'authority': 'play.vercel.ai', - 'custom-encoding': token, - 'origin': 'https://play.vercel.ai', - 'referer': 'https://play.vercel.ai/', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36' -} - -for chunk in requests.post('https://play.vercel.ai/api/generate', headers=headers, stream=True, json={ - 'prompt': 'hi', - 'model': 'openai:gpt-3.5-turbo', - 'temperature': 0.7, - 'maxTokens': 200, - 'topK': 1, - 'topP': 1, - 'frequencyPenalty': 1, - 'presencePenalty': 1, - 'stopSequences': []}).iter_lines(): - print(chunk) diff --git a/unfinished/vercelai/vercelai_test.py b/unfinished/vercelai/vercelai_test.py new file mode 100644 index 00000000..24cbe0bc --- /dev/null +++ b/unfinished/vercelai/vercelai_test.py @@ -0,0 +1,5 @@ +import vercelai + +for token in vercelai.Completion.create('summarize the gnu gpl 1.0'): + print(token, end='', flush=True) + diff --git a/unfinished/writesonic/README.md b/unfinished/writesonic/README.md deleted file mode 100644 index a658a87c..00000000 --- a/unfinished/writesonic/README.md +++ /dev/null @@ -1,53 +0,0 @@ -### Example: `writesonic` (use like openai pypi package) <a name="example-writesonic"></a> - -```python -# import writesonic -import writesonic - -# create account (3-4s) -account = writesonic.Account.create(logging = True) - -# with loging: - # 2023-04-06 21:50:25 INFO __main__ -> register success : '{"id":"51aa0809-3053-44f7-922a...' (2s) - # 2023-04-06 21:50:25 INFO __main__ -> id : '51aa0809-3053-44f7-922a-2b85d8d07edf' - # 2023-04-06 21:50:25 INFO __main__ -> token : 'eyJhbGciOiJIUzI1NiIsInR5cCI6Ik...' - # 2023-04-06 21:50:28 INFO __main__ -> got key : '194158c4-d249-4be0-82c6-5049e869533c' (2s) - -# simple completion -response = writesonic.Completion.create( - api_key = account.key, - prompt = 'hello world' -) - -print(response.completion.choices[0].text) # Hello! How may I assist you today? - -# conversation - -response = writesonic.Completion.create( - api_key = account.key, - prompt = 'what is my name ?', - enable_memory = True, - history_data = [ - { - 'is_sent': True, - 'message': 'my name is Tekky' - }, - { - 'is_sent': False, - 'message': 'hello Tekky' - } - ] -) - -print(response.completion.choices[0].text) # Your name is Tekky. - -# enable internet - -response = writesonic.Completion.create( - api_key = account.key, - prompt = 'who won the quatar world cup ?', - enable_google_results = True -) - -print(response.completion.choices[0].text) # Argentina won the 2022 FIFA World Cup tournament held in Qatar ... -```
\ No newline at end of file diff --git a/unfinished/writesonic/__init__.py b/unfinished/writesonic/__init__.py deleted file mode 100644 index ce684912..00000000 --- a/unfinished/writesonic/__init__.py +++ /dev/null @@ -1,163 +0,0 @@ -from random import choice -from time import time - -from colorama import Fore, init; -from names import get_first_name, get_last_name -from requests import Session -from requests import post - -init() - - -class logger: - @staticmethod - def info(string) -> print: - import datetime - now = datetime.datetime.now() - return print( - f"{Fore.CYAN}{now.strftime('%Y-%m-%d %H:%M:%S')} {Fore.BLUE}INFO {Fore.MAGENTA}__main__ -> {Fore.RESET}{string}") - - -class SonicResponse: - class Completion: - class Choices: - def __init__(self, choice: dict) -> None: - self.text = choice['text'] - self.content = self.text.encode() - self.index = choice['index'] - self.logprobs = choice['logprobs'] - self.finish_reason = choice['finish_reason'] - - def __repr__(self) -> str: - return f'''<__main__.APIResponse.Completion.Choices(\n text = {self.text.encode()},\n index = {self.index},\n logprobs = {self.logprobs},\n finish_reason = {self.finish_reason})object at 0x1337>''' - - def __init__(self, choices: dict) -> None: - self.choices = [self.Choices(choice) for choice in choices] - - class Usage: - def __init__(self, usage_dict: dict) -> None: - self.prompt_tokens = usage_dict['prompt_chars'] - self.completion_tokens = usage_dict['completion_chars'] - self.total_tokens = usage_dict['total_chars'] - - def __repr__(self): - return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>''' - - def __init__(self, response_dict: dict) -> None: - self.response_dict = response_dict - self.id = response_dict['id'] - self.object = response_dict['object'] - self.created = response_dict['created'] - self.model = response_dict['model'] - self.completion = self.Completion(response_dict['choices']) - self.usage = self.Usage(response_dict['usage']) - - def json(self) -> dict: - return self.response_dict - - -class Account: - session = Session() - session.headers = { - "connection": "keep-alive", - "sec-ch-ua": "\"Not_A Brand\";v=\"99\", \"Google Chrome\";v=\"109\", \"Chromium\";v=\"109\"", - "accept": "application/json, text/plain, */*", - "content-type": "application/json", - "sec-ch-ua-mobile": "?0", - "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36", - "sec-ch-ua-platform": "\"Windows\"", - "sec-fetch-site": "same-origin", - "sec-fetch-mode": "cors", - "sec-fetch-dest": "empty", - # "accept-encoding" : "gzip, deflate, br", - "accept-language": "en-GB,en-US;q=0.9,en;q=0.8", - "cookie": "" - } - - @staticmethod - def get_user(): - password = f'0opsYouGoTme@1234' - f_name = get_first_name() - l_name = get_last_name() - hosts = ['gmail.com', 'protonmail.com', 'proton.me', 'outlook.com'] - - return { - "email": f"{f_name.lower()}.{l_name.lower()}@{choice(hosts)}", - "password": password, - "confirm_password": password, - "full_name": f'{f_name} {l_name}' - } - - @staticmethod - def create(logging: bool = False): - while True: - try: - user = Account.get_user() - start = time() - response = Account.session.post("https://app.writesonic.com/api/session-login", json=user | { - "utmParams": "{}", - "visitorId": "0", - "locale": "en", - "userAgent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36", - "signInWith": "password", - "request_type": "signup", - }) - - if logging: - logger.info(f"\x1b[31mregister success\x1b[0m : '{response.text[:30]}...' ({int(time() - start)}s)") - logger.info(f"\x1b[31mid\x1b[0m : '{response.json()['id']}'") - logger.info(f"\x1b[31mtoken\x1b[0m : '{response.json()['token'][:30]}...'") - - start = time() - response = Account.session.post("https://api.writesonic.com/v1/business/set-business-active", - headers={"authorization": "Bearer " + response.json()['token']}) - key = response.json()["business"]["api_key"] - if logging: logger.info(f"\x1b[31mgot key\x1b[0m : '{key}' ({int(time() - start)}s)") - - return Account.AccountResponse(user['email'], user['password'], key) - - except Exception as e: - if logging: logger.info(f"\x1b[31merror\x1b[0m : '{e}'") - continue - - class AccountResponse: - def __init__(self, email, password, key): - self.email = email - self.password = password - self.key = key - - -class Completion: - def create( - api_key: str, - prompt: str, - enable_memory: bool = False, - enable_google_results: bool = False, - history_data: list = []) -> SonicResponse: - response = post('https://api.writesonic.com/v2/business/content/chatsonic?engine=premium', - headers={"X-API-KEY": api_key}, - json={ - "enable_memory": enable_memory, - "enable_google_results": enable_google_results, - "input_text": prompt, - "history_data": history_data}).json() - - return SonicResponse({ - 'id': f'cmpl-premium-{int(time())}', - 'object': 'text_completion', - 'created': int(time()), - 'model': 'premium', - - 'choices': [{ - 'text': response['message'], - 'index': 0, - 'logprobs': None, - 'finish_reason': 'stop' - }], - - 'usage': { - 'prompt_chars': len(prompt), - 'completion_chars': len(response['message']), - 'total_chars': len(prompt) + len(response['message']) - } - }) |