- Move all image-generation content into docs/image-generation-and-editing/ - Add ComfyUI image-editing instructions (Image Size, Steps, Image Prompt Generation) - Explain how to set node IDs in Open WebUI - Add example setups for Qwen and FLUX.1 models - Fix broken links and tidy formatting
3.4 KiB
sidebar_position, title
| sidebar_position | title |
|---|---|
| 5 | 🎨 Gemini |
:::warning This tutorial is a community contribution and is not supported by the Open WebUI team. It serves only as a demonstration on how to customize Open WebUI for your specific use case. Want to contribute? Check out the contributing tutorial. :::
Open WebUI also supports image generation through the Google AI Studio API also known as the Gemini API.
Initial Setup
- Obtain an API key from Google AI Studio.
- You may need to create a project and enable the
Generative Language APIin addition to adding billing information.
Configuring Open WebUI
- In Open WebUI, navigate to the Admin Panel > Settings > Images menu.
- Set the
Image Generation Enginefield toGemini. - Set the
API Base URLtohttps://generativelanguage.googleapis.com/v1beta. - Enter your Google AI Studio API key.
- Enter the model you wish to use from these available models.
- Set the image size to one of the available image sizes.
:::info
This feature appears to only work for models supported with this endpoint: https://generativelanguage.googleapis.com/v1beta/models/<MODEL_NAME>:predict.
This is the OpenAI BETA endpoint, which Google provides for experimental OpenAI compatibility.
Google Imagen models use this endpoint while Gemini models use a different endpoint ending with :generateContent
Imagen model endpoint example:
https://generativelanguage.googleapis.com/v1beta/models/imagen-4.0-generate-001:predict.- Documentation for Imagen models
Gemini model endpoint example:
https://generativelanguage.googleapis.com/v1beta/models/gemini-2.5-flash-image:generateContent.- Documentation for Gemini models
Trying to call a Gemini model, such as gemini-2.5-flash-image aka Nano Banana would result in an error due to the difference in supported endpoints for Image Generation.
400: [ERROR: models/gemini-2.5-flash-image is not found for API version v1beta, or is not supported for predict. Call ListModels to see the list of available models and their supported methods.]
:::
LiteLLM Proxy with Gemini Endpoints
Image generation with a LiteLLM proxy using Gemini or Imagen endpoints is supported with Open WebUI. Configure the Image Generation as follows:
- In Open WebUI, navigate to the Admin Panel > Settings > Images menu.
- Set the
Image Generation Enginefield toOpen AI. - Change the API endpoint URL to
https://<your-litellm-url>:<port>/v1. - Enter your LiteLLM API key.
- The API version can be left blank.
- Enter the image model name as it appears in your LiteLLM configuration.
- Set the image size to one of the available sizes for the selected model.
:::tip
To find your LiteLLM connection information, navigate to the Admin Panel > Settings > Connections menu. Your connection information will be listed under the Gemini API connection.
:::
