diff options
Diffstat (limited to '')
-rw-r--r-- | README.md | 207 | ||||
-rw-r--r-- | docs/client.md | 12 | ||||
-rw-r--r-- | docs/docker.md | 19 | ||||
-rw-r--r-- | docs/git.md | 66 | ||||
-rw-r--r-- | docs/interference.md | 69 | ||||
-rw-r--r-- | docs/leagcy.md | 18 | ||||
-rw-r--r-- | docs/requirements.md | 10 | ||||
-rw-r--r-- | g4f/Provider/needs_auth/Gemini.py | 38 |
8 files changed, 250 insertions, 189 deletions
@@ -100,72 +100,43 @@ or set the api base in your client to: [http://localhost:1337/v1](http://localho ##### Install using pypi: -Install all supported tools / all used packages: ``` pip install -U g4f[all] ``` -Or use: [Partially Requirements](/docs/requirements.md) +Or use partial requirements. +See: [/docs/requirements](/docs/requirements.md) ##### Install from source: -1. Clone the GitHub repository: +See: [/docs/git](/docs/git.md) -``` -git clone https://github.com/xtekky/gpt4free.git -``` - -2. Navigate to the project directory: - -``` -cd gpt4free -``` -3. (Recommended) Create a Python virtual environment: -You can follow the [Python official documentation](https://docs.python.org/3/tutorial/venv.html) for virtual environments. - - -``` -python3 -m venv venv -``` +##### Install using Docker -4. Activate the virtual environment: - - On Windows: - ``` - .\venv\Scripts\activate - ``` - - On macOS and Linux: - ``` - source venv/bin/activate - ``` -5. Install minimum requirements: + See: [/docs/docker](/docs/docker.md) -``` -pip install -r requirements-min.txt -``` -6. Or install all used Python packages from `requirements.txt`: +## 💡 Usage -``` -pip install -r requirements.txt -``` +#### Text Generation +**with Python** -7. Create a `test.py` file in the root folder and start using the repo, further Instructions are below +```python +from g4f.client import Client -```py -import g4f -... +client = Client() +response = client.chat.completions.create( + model="gpt-3.5-turbo", + messages=[{"role": "user", "content": "Say this is a test"}], + ... +) +print(response.choices[0].message.content) ``` -##### Install using Docker - - Or use: [Build Docker](/docs/docker.md) - - -## 💡 Usage - #### Image Generation +**with Python** ```python from g4f.client import Client @@ -182,9 +153,7 @@ Result: [![Image with cat](/docs/cat.jpeg)](/docs/client.md) -#### Text Generation - -and more: +**See also for Python:** - [Documentation for new Client](/docs/client.md) - [Documentation for leagcy API](/docs/leagcy.md) @@ -192,19 +161,31 @@ and more: #### Web UI -To start the web interface, type the following codes in the command line. +To start the web interface, type the following codes in python: ```python from g4f.gui import run_gui run_gui() ``` +or type in command line: +```bash +python -m g4f.cli gui -port 8080 -debug +``` + +### Interference API + +You can use the Interference API to serve other OpenAI integrations with G4F. + +See: [/docs/interference](/docs/interference.md) + +### Configuration ##### Cookies / Access Token -For generating images with Bing and for the OpenAi Chat you need cookies or a token from your browser session. From Bing you need the "_U" cookie and from OpenAI you need the "access_token". You can pass the cookies / the access token in the create function or you use the `set_cookies` setter: +For generating images with Bing and for the OpenAi Chat you need cookies or a token from your browser session. From Bing you need the "_U" cookie and from OpenAI you need the "access_token". You can pass the cookies / the access token in the create function or you use the `set_cookies` setter before you run G4F: ```python -from g4f import set_cookies +from g4f.cookies import set_cookies set_cookies(".bing.com", { "_U": "cookie value" @@ -212,124 +193,30 @@ set_cookies(".bing.com", { set_cookies("chat.openai.com", { "access_token": "token value" }) +set_cookies(".google.com", { + "__Secure-1PSID": "cookie value" +}) -from g4f.gui import run_gui -run_gui() -``` - -Alternatively, g4f reads the cookies with “browser_cookie3” from your browser -or it starts a browser instance with selenium "webdriver" for logging in. -If you use the pip package, you have to install “browser_cookie3” or "webdriver" by yourself. - -```bash -pip install browser_cookie3 -pip install g4f[webdriver] +... ``` -##### Proxy and Timeout Support +Alternatively, G4F reads the cookies with `browser_cookie3` from your browser +or it starts a browser instance with selenium `webdriver` for logging in. -All providers support specifying a proxy and increasing timeout in the create functions. +##### Using Proxy -```python -import g4f +If you want to hide or change your IP address for the providers, you can set a proxy globally via an environment variable: -response = g4f.ChatCompletion.create( - model=g4f.models.default, - messages=[{"role": "user", "content": "Hello"}], - proxy="http://host:port", - # or socks5://user:pass@host:port - timeout=120, # in secs -) - -print(f"Result:", response) -``` - -You can also set a proxy globally via an environment variable: - -```sh +- On macOS and Linux: +```bash export G4F_PROXY="http://host:port" ``` -### Interference openai-proxy API (Use with openai python package) - -#### Run interference API from PyPi package - -```python -from g4f.api import run_api - -run_api() -``` - -#### Run interference API from repo - -If you want to use the embedding function, you need to get a Hugging Face token. You can get one at [Hugging Face Tokens](https://huggingface.co/settings/tokens). Make sure your role is set to write. If you have your token, just use it instead of the OpenAI api-key. - -Run server: - -```sh -g4f api -``` - -or - -```sh -python -m g4f.api.run -``` - -```python -from openai import OpenAI - -client = OpenAI( - # Set your Hugging Face token as the API key if you use embeddings - api_key="YOUR_HUGGING_FACE_TOKEN", - - # Set the API base URL if needed, e.g., for a local development environment - base_url="http://localhost:1337/v1" -) - - -def main(): - chat_completion = client.chat.completions.create( - model="gpt-3.5-turbo", - messages=[{"role": "user", "content": "write a poem about a tree"}], - stream=True, - ) - - if isinstance(chat_completion, dict): - # Not streaming - print(chat_completion.choices[0].message.content) - else: - # Streaming - for token in chat_completion: - content = token.choices[0].delta.content - if content is not None: - print(content, end="", flush=True) - - -if __name__ == "__main__": - main() -``` - -## API usage (POST) -#### Chat completions -Send the POST request to /v1/chat/completions with body containing the `model` method. This example uses python with requests library: -```python -import requests -url = "http://localhost:1337/v1/chat/completions" -body = { - "model": "gpt-3.5-turbo-16k", - "stream": False, - "messages": [ - {"role": "assistant", "content": "What can you do?"} - ] -} -json_response = requests.post(url, json=body).json().get('choices', []) - -for choice in json_response: - print(choice.get('message', {}).get('content', '')) +- On Windows: +```bash +set G4F_PROXY=http://host:port ``` - ## 🚀 Providers and Models ### GPT-4 diff --git a/docs/client.md b/docs/client.md index 8e02b581..f2ba9bcd 100644 --- a/docs/client.md +++ b/docs/client.md @@ -44,10 +44,22 @@ client = Client( You can use the `ChatCompletions` endpoint to generate text completions as follows: ```python +response = client.chat.completions.create( + model="gpt-3.5-turbo", + messages=[{"role": "user", "content": "Say this is a test"}], + ... +) +print(response.choices[0].message.content) +``` + +Also streaming are supported: + +```python stream = client.chat.completions.create( model="gpt-4", messages=[{"role": "user", "content": "Say this is a test"}], stream=True, + ... ) for chunk in stream: if chunk.choices[0].delta.content: diff --git a/docs/docker.md b/docs/docker.md index 6baf386a..db33b925 100644 --- a/docs/docker.md +++ b/docs/docker.md @@ -1,38 +1,37 @@ -### G4F - Docker +### G4F - Docker Setup -If you have Docker installed, you can easily set up and run the project without manually installing dependencies. - -1. First, ensure you have both Docker and Docker Compose installed. +Easily set up and run the G4F project using Docker without the hassle of manual dependency installation. +1. **Prerequisites:** - [Install Docker](https://docs.docker.com/get-docker/) - [Install Docker Compose](https://docs.docker.com/compose/install/) -2. Clone the GitHub repo: +2. **Clone the Repository:** ```bash git clone https://github.com/xtekky/gpt4free.git ``` -3. Navigate to the project directory: +3. **Navigate to the Project Directory:** ```bash cd gpt4free ``` -4. Build the Docker image: +4. **Build the Docker Image:** ```bash docker pull selenium/node-chrome docker-compose build ``` -5. Start the service using Docker Compose: +5. **Start the Service:** ```bash docker-compose up ``` -Your server will now be running at `http://localhost:1337`. You can interact with the API or run your tests as you would normally. +Your server will now be accessible at `http://localhost:1337`. Interact with the API or run tests as usual. To stop the Docker containers, simply run: @@ -41,6 +40,6 @@ docker-compose down ``` > [!Note] -> When using Docker, any changes you make to your local files will be reflected in the Docker container thanks to the volume mapping in the `docker-compose.yml` file. If you add or remove dependencies, however, you'll need to rebuild the Docker image using `docker-compose build`. +> Changes made to local files reflect in the Docker container due to volume mapping in `docker-compose.yml`. However, if you add or remove dependencies, rebuild the Docker image using `docker-compose build`. [Return to Home](/)
\ No newline at end of file diff --git a/docs/git.md b/docs/git.md new file mode 100644 index 00000000..89137ffc --- /dev/null +++ b/docs/git.md @@ -0,0 +1,66 @@ +### G4F - Installation Guide + +Follow these steps to install G4F from the source code: + +1. **Clone the Repository:** + +```bash +git clone https://github.com/xtekky/gpt4free.git +``` + +2. **Navigate to the Project Directory:** + +```bash +cd gpt4free +``` + +3. **(Optional) Create a Python Virtual Environment:** + +It's recommended to isolate your project dependencies. You can follow the [Python official documentation](https://docs.python.org/3/tutorial/venv.html) for virtual environments. + +```bash +python3 -m venv venv +``` + +4. **Activate the Virtual Environment:** + +- On Windows: + +```bash +.\venv\Scripts\activate +``` + +- On macOS and Linux: + +```bash +source venv/bin/activate +``` + +5. **Install Minimum Requirements:** + +Install the minimum required packages: + +```bash +pip install -r requirements-min.txt +``` + +6. **Or Install All Packages from `requirements.txt`:** + +If you prefer, you can install all packages listed in `requirements.txt`: + +```bash +pip install -r requirements.txt +``` + +7. **Start Using the Repository:** + +You can now create Python scripts and utilize the G4F functionalities. Here's a basic example: + +Create a `test.py` file in the root folder and start using the repository: + +```python +import g4f +# Your code here +``` + +[Return to Home](/)
\ No newline at end of file diff --git a/docs/interference.md b/docs/interference.md new file mode 100644 index 00000000..b140f66a --- /dev/null +++ b/docs/interference.md @@ -0,0 +1,69 @@ +### Interference openai-proxy API + +#### Run interference API from PyPi package + +```python +from g4f.api import run_api + +run_api() +``` + +#### Run interference API from repo + +Run server: + +```sh +g4f api +``` + +or + +```sh +python -m g4f.api.run +``` + +```python +from openai import OpenAI + +client = OpenAI( + api_key="", + # Change the API base URL to the local interference API + base_url="http://localhost:1337/v1" +) + + response = client.chat.completions.create( + model="gpt-3.5-turbo", + messages=[{"role": "user", "content": "write a poem about a tree"}], + stream=True, + ) + + if isinstance(response, dict): + # Not streaming + print(response.choices[0].message.content) + else: + # Streaming + for token in response: + content = token.choices[0].delta.content + if content is not None: + print(content, end="", flush=True) +``` + +#### API usage (POST) +Send the POST request to /v1/chat/completions with body containing the `model` method. This example uses python with requests library: +```python +import requests +url = "http://localhost:1337/v1/chat/completions" +body = { + "model": "gpt-3.5-turbo-16k", + "stream": False, + "messages": [ + {"role": "assistant", "content": "What can you do?"} + ] +} +json_response = requests.post(url, json=body).json().get('choices', []) + +for choice in json_response: + print(choice.get('message', {}).get('content', '')) +``` + +[Return to Home](/)
\ No newline at end of file diff --git a/docs/leagcy.md b/docs/leagcy.md index 224bc098..e8808381 100644 --- a/docs/leagcy.md +++ b/docs/leagcy.md @@ -179,4 +179,22 @@ async def run_all(): asyncio.run(run_all()) ``` +##### Proxy and Timeout Support + +All providers support specifying a proxy and increasing timeout in the create functions. + +```python +import g4f + +response = g4f.ChatCompletion.create( + model=g4f.models.default, + messages=[{"role": "user", "content": "Hello"}], + proxy="http://host:port", + # or socks5://user:pass@host:port + timeout=120, # in secs +) + +print(f"Result:", response) +``` + [Return to Home](/)
\ No newline at end of file diff --git a/docs/requirements.md b/docs/requirements.md index 7715a403..a4137a64 100644 --- a/docs/requirements.md +++ b/docs/requirements.md @@ -6,15 +6,19 @@ You can install requirements partially or completely. So G4F can be used as you #### Options -Install required packages for the OpenaiChat provider: +Install g4f with all possible dependencies: +``` +pip install -U g4f[all] +``` +Or install only g4f and the required packages for the OpenaiChat provider: ``` pip install -U g4f[openai] ``` -Install required packages for the interference api: +Install required packages for the Interference API: ``` pip install -U g4f[api] ``` -Install required packages for the web interface: +Install required packages for the Web UI: ``` pip install -U g4f[gui] ``` diff --git a/g4f/Provider/needs_auth/Gemini.py b/g4f/Provider/needs_auth/Gemini.py index 0650942e..bd30d8b2 100644 --- a/g4f/Provider/needs_auth/Gemini.py +++ b/g4f/Provider/needs_auth/Gemini.py @@ -50,7 +50,6 @@ class Gemini(AsyncGeneratorProvider): url = "https://gemini.google.com" needs_auth = True working = True - supports_stream = False @classmethod async def create_async_generator( @@ -64,10 +63,9 @@ class Gemini(AsyncGeneratorProvider): **kwargs ) -> AsyncResult: prompt = format_prompt(messages) - - if not cookies: - cookies = get_cookies(".google.com", False, True) - if "__Secure-1PSID" not in cookies or "__Secure-1PSIDCC" not in cookies: + cookies = cookies if cookies else get_cookies(".google.com", False, True) + snlm0e = await cls.fetch_snlm0e(cookies, proxy) if cookies else None + if not snlm0e: driver = None try: driver = get_browser(proxy=proxy) @@ -90,8 +88,12 @@ class Gemini(AsyncGeneratorProvider): if driver: driver.close() - if "__Secure-1PSID" not in cookies: - raise MissingAuthError('Missing "__Secure-1PSID" cookie') + if not snlm0e: + if "__Secure-1PSID" not in cookies: + raise MissingAuthError('Missing "__Secure-1PSID" cookie') + snlm0e = await cls.fetch_snlm0e(cookies, proxy) + if not snlm0e: + raise RuntimeError("Invalid auth. SNlM0e not found") image_url = await cls.upload_image(to_bytes(image), image_name, proxy) if image else None @@ -99,14 +101,6 @@ class Gemini(AsyncGeneratorProvider): cookies=cookies, headers=REQUEST_HEADERS ) as session: - async with session.get(cls.url, proxy=proxy) as response: - text = await response.text() - match = re.search(r'SNlM0e\":\"(.*?)\"', text) - if match: - snlm0e = match.group(1) - else: - raise RuntimeError("SNlM0e not found") - params = { 'bl': REQUEST_BL_PARAM, '_reqid': random.randint(1111, 9999), @@ -204,4 +198,16 @@ class Gemini(AsyncGeneratorProvider): upload_url, headers=headers, data=image, proxy=proxy ) as response: response.raise_for_status() - return await response.text()
\ No newline at end of file + return await response.text() + + @classmethod + async def fetch_snlm0e(cls, cookies: Cookies, proxy: str = None): + async with ClientSession( + cookies=cookies, + headers=REQUEST_HEADERS + ) as session: + async with session.get(cls.url, proxy=proxy) as response: + text = await response.text() + match = re.search(r'SNlM0e\":\"(.*?)\"', text) + if match: + return match.group(1)
\ No newline at end of file |