InvokeAI icon

InvokeAI

A professional, creative-friendly interface for Stable Diffusion.

InvokeAI: Open-Source AI Image Generation and Creative Engine

Introduction

InvokeAI is an open-source, AI-powered creative engine designed to empower professionals, artists, and enthusiasts in generating and creating visual media using Stable Diffusion and other AI models. It provides a robust and user-friendly platform, including an industry-leading WebUI and a powerful backend, that allows for extensive customization and control over the image generation process. Developed by the InvokeAI team and a vibrant open-source community, its mission is to offer a stable, powerful, and accessible tool for AI-driven creativity, serving as a foundation for both personal exploration and commercial products.

The platform is known for its sophisticated features like the Unified Canvas for advanced editing and compositing, and a node-based Workflow Editor for creating complex image generation pipelines, catering to users who want more than just basic text-to-image capabilities.

Key Features

InvokeAI offers a comprehensive suite of features for advanced AI image generation and editing:

  • Core Stable Diffusion Functionality:
    • Text-to-Image (txt2img): Generate images from detailed text prompts.
    • Image-to-Image (img2img): Transform existing images based on text prompts and image inputs, with control over the degree of transformation.
  • Unified Canvas:
    • A powerful, integrated, and "infinite" canvas that allows for seamless inpainting (modifying specific parts of an image), outpainting (extending image boundaries), and compositing multiple AI-generated or uploaded elements into a single cohesive artwork.
    • Includes brush tools and intuitive controls for a fluid editing experience.
  • Workflow Editor (Nodes):
    • A node-based system for creating, customizing, and sharing complex image generation and processing pipelines without traditional coding.
    • Each node represents a specific operation (e.g., loading a model, inputting prompts, applying samplers, upscaling, image operations).
    • Allows for advanced and repeatable generation workflows.
  • Model Management:
    • Easy installation, importing, and switching between various Stable Diffusion models (checkpoints in .ckpt or .safetensors formats).
    • Support for SDXL (Stable Diffusion XL) models for higher resolution and detail.
    • Integration and management of LoRAs (Low-Rank Adaptations), Textual Inversions (Embeddings), and VAEs (Variational Auto-Encoders) to customize styles, subjects, and image quality.
    • Support for ControlNet models, allowing users to guide image generation with precise control over pose, depth, edges, and other structural elements from reference images.
    • Model manager for organizing and selecting installed models.
  • Image Generation Parameters:
    • Granular control over prompts, negative prompts, seed, steps, CFG scale, samplers/schedulers (e.g., Euler, DPM++ 2M SDE Karras), image dimensions, and aspect ratios.
  • Image Enhancement & Post-Processing:
    • Upscaling Tools: Various algorithms to increase image resolution and detail.
    • Face Restoration: Tools to improve or correct faces in AI-generated portraits.
  • Image & Workflow Gallery/Library:
    • A local gallery to manage, view, and organize generated images.
    • Save and load custom workflows created in the Node Editor.
    • Generated images often save metadata (prompts, model, seed) for reproducibility.
  • Cross-Platform Availability:
    • User-friendly installers for Windows, macOS (including Apple Silicon), and Linux.
    • Docker support for containerized deployment.
  • Command-Line Interface (CLI):
    • Provides CLI access for users who prefer or need to interact with InvokeAI programmatically or for automation.
  • Open Source & Community Driven:
    • InvokeAI is open-source software, fostering active development and contributions from its community.
  • Focus on Stability & User Experience: Aims to provide a polished and reliable user experience for both beginners and advanced users.
  • Invocation API: For more advanced users and developers, InvokeAI has an underlying API structure that allows interaction with its services (e.g., saving images, loading models, logging) when building custom nodes or integrations.

Specific Use Cases

InvokeAI's powerful and flexible toolkit is suitable for a wide array of creative and professional endeavors:

  • AI Art Creation: Generating unique digital art, illustrations, and abstract pieces in diverse styles.
  • Professional Image Editing & Compositing: Using the Unified Canvas for detailed inpainting, outpainting, and seamlessly blending multiple AI-generated or existing image elements.
  • Concept Art & Character Design: Creating detailed concept art for games, films, and other media, including character designs, environments, and props.
  • Workflow Automation for Artists: Designing and saving complex, repeatable image generation pipelines using the Node Editor.
  • Photorealistic Image Generation: Crafting highly realistic images for various purposes.
  • Texture Generation: Creating unique textures for 3D models or graphic design.
  • Experimentation with Stable Diffusion: A robust platform for exploring the capabilities of different Stable Diffusion models, LoRAs, ControlNets, and generation parameters.
  • Educational Purposes: Helping users learn advanced AI image generation techniques through a visual and interactive interface.

Usage Guide

Setting up and using InvokeAI typically involves the following:

  1. System Requirements:
    • GPU (Highly Recommended):
      • NVIDIA: 10xx series or later with 4GB+ VRAM (SD1.5 at 512x512). For SDXL (1024x1024) or FLUX models, 8GB-10GB+ VRAM (NVIDIA 20xx series or later) is recommended.
      • AMD: Supported on Linux only, with VRAM requirements similar to NVIDIA.
    • Apple Silicon (M1/M2/M3): Works well, with 16GB+ unified memory recommended.
    • RAM: At least 8GB system RAM (for SD1.5), with 16GB-32GB+ recommended for larger models like SDXL or FLUX.
    • Disk Space: 10GB for base installation, plus significant additional space for models (e.g., 30GB to 200GB+ depending on how many models you download). SSD is highly recommended.
    • Operating System: Windows 10+, macOS 14+, Linux (Ubuntu 20.04+ is well-tested).
  2. Installation:
    • Official Installers (Recommended for most users): Download the latest launcher for your operating system (Windows, macOS, Linux) from the InvokeAI official website (https://invoke.ai/download or via GitHub releases). Run the launcher and follow the on-screen instructions. This typically handles dependencies and initial model downloads.
    • Manual Installation / Docker: For advanced users or specific environments, manual installation instructions (involving Python, uv, and pip) and Docker deployment options are available in the official documentation (https://invoke-ai.github.io/InvokeAI/installation/manual/).
  3. Launching InvokeAI:
    • After installation, run the InvokeAI application. It will start a local web server.
    • Access the WebUI by opening the provided local URL (usually http://localhost:9090 or http://127.0.0.1:9090) in your browser.
  4. Navigating the User Interface:
    • Text-to-Image / Image-to-Image Tabs: Main areas for basic image generation. Input prompts, negative prompts, select models, adjust parameters (sampler, steps, CFG, seed, dimensions).
    • Unified Canvas: An interactive workspace for inpainting, outpainting, extending images, sketching, and combining elements. Upload images or send generated images to the canvas.
    • Workflow (Node) Editor: A graph-based editor to create and run complex image generation pipelines by connecting various nodes (model loaders, prompt encoders, samplers, image operations, upscalers, etc.).
    • Model Manager: Install, manage, and import Stable Diffusion models (checkpoints), LoRAs, ControlNets, Textual Inversions, and VAEs. InvokeAI can download "Starter Models" or you can import your own.
    • Image Gallery: Browse, view, and manage your previously generated images. Metadata (prompts, settings) is often saved with images.
  5. Generating Images:
    • Basic Generation: In the txt2img tab, type a detailed prompt, select a model, choose a sampler and other settings, and click "Invoke."
    • Using the Unified Canvas: Upload an image or send one from a generation. Use tools like the brush for masking (inpainting), the bounding box to define areas for outpainting or generation, and invoke the AI with new prompts for selected regions.
    • Using the Node Editor: Build a flow by adding and connecting nodes. Define inputs (prompts, models, images) and execute the workflow to generate images.
  6. Prompt Crafting:
    • Be descriptive: Include subject, style (e.g., "photograph," "oil painting," "pop art"), quality ("masterpiece," "award-winning"), aesthetics (colors, mood, lighting, setting).
    • Use negative prompts to exclude undesired elements.
    • Experiment with different models and schedulers (samplers) as they significantly affect the output.

Pricing & Plans

InvokeAI is free and open-source software, licensed under the Apache 2.0 License.

  • There are no subscription fees or charges for using the InvokeAI software itself.
  • Users can download, install, and run it on their own hardware without cost from InvokeAI.
  • Costs are associated with:
    • Your hardware: The computer and especially the GPU needed to run Stable Diffusion models effectively.
    • Electricity consumption.
    • Optional cloud compute: If you choose to run InvokeAI on a cloud GPU instance (e.g., via Vultr Marketplace or other cloud providers), you will pay for the cloud resources.

Commercial Use & Licensing

  • InvokeAI Software: The InvokeAI software itself is licensed under the Apache License 2.0, which is a permissive open-source license that allows for commercial use, modification, and distribution (subject to the license terms, such as including notices).
  • Generated Images: You generally own the images you create with InvokeAI using open-source Stable Diffusion models. However, the specific license of the base Stable Diffusion model (e.g., SDXL, SD 1.5) and any LoRAs, Textual Inversions, or other custom models you use will govern the rights to the output. Many Stable Diffusion models are released under permissive licenses (like CreativeML OpenRAIL++-M or similar) that allow commercial use, but it's crucial to check the license for each model asset.
  • Responsibility: Users are responsible for the content they generate, ensuring it does not infringe on copyrights, trademarks, or other rights, and adheres to ethical guidelines.

Frequently Asked Questions (FAQ)

Q1: What is InvokeAI? A1: InvokeAI is a free, open-source, user-friendly platform and creative engine for generating images and other visual media using Stable Diffusion and similar AI models. It offers a powerful WebUI with features like a Unified Canvas and a node-based Workflow Editor for advanced control.

Q2: Is InvokeAI free? A2: Yes, InvokeAI is free and open-source software licensed under Apache 2.0. You can download and use it without paying for the software itself.

Q3: What hardware do I need to run InvokeAI? A3: A dedicated GPU (NVIDIA recommended, with 4GB VRAM minimum, 8-10GB+ for SDXL/FLUX) is highly recommended for good performance. You'll also need sufficient system RAM (8GB minimum, 16-32GB+ recommended) and disk space for the software and models. It can run on Windows, macOS (including Apple Silicon), and Linux.

Q4: What makes InvokeAI different from Automatic1111 Stable Diffusion WebUI or ComfyUI? A4: While all are popular UIs for Stable Diffusion: * InvokeAI is often praised for its polished user experience, integrated Unified Canvas for advanced editing, and a user-friendly approach to its node-based Workflow Editor, aiming to balance power with ease of use. It provides official installers for easier setup. * Automatic1111 is known for its vast number of features, settings, and a massive extension ecosystem, offering extreme customizability but can have a steeper learning curve for some. * ComfyUI is purely node-based, offering the most granular control over the generation pipeline, appealing to users who want to deeply customize every step.

Q5: What types of AI models can I use with InvokeAI? A5: InvokeAI primarily supports Stable Diffusion models (various versions like SD1.5, SDXL, and community fine-tunes) and other compatible diffusion models (like FLUX, CogView4 mentioned in releases). It handles checkpoints (.ckpt, .safetensors), LoRAs, Textual Inversions, VAEs, and ControlNet models.

Q6: What is the Unified Canvas in InvokeAI? A6: The Unified Canvas is an integrated, infinite canvas-like interface within InvokeAI that allows for advanced image editing and generation tasks. You can perform inpainting (editing parts of an image), outpainting (extending an image), combine multiple images, sketch, and use AI to generate content within specific regions seamlessly.

Q7: What is the Workflow (Node) Editor in InvokeAI? A7: The Workflow Editor provides a node-based interface where users can visually construct complex image generation and processing pipelines by connecting different operational nodes (e.g., model loaders, prompt inputs, samplers, upscalers, image filters). This allows for powerful customization and automation of creative workflows.

Q8: Where can I get support for InvokeAI? A8: The primary support channel is the official InvokeAI Discord server. The GitHub repository (Discussions and Issues sections) and the official documentation website are also key resources.

Here are examples of helpful resources and tutorials for InvokeAI:

Community & Support

  • Discord Server: The official InvokeAI Discord server is the main hub for community interaction, user support, sharing creations, and getting help from developers and experienced users. (Link usually prominent on their website and GitHub).
  • GitHub Repository: For source code, issue tracking, discussions, and contributions. (https://github.com/invoke-ai/InvokeAI)
  • InvokeAI Support Portal: https://support.invoke.ai/ (Provides access to knowledge base and ways to contact support).

Ethical Considerations & Safety

  • User Responsibility: As InvokeAI is a tool that runs locally and utilizes open-source models, users are responsible for the content they generate and must adhere to the licenses of any models they use.
  • Content Generation: The tool itself does not inherently filter content beyond what the chosen models are trained for. Users should be mindful of creating ethical and non-harmful imagery.
  • Open Source Nature: Being open source allows for community scrutiny and development of safer practices if needed.

Last updated: May 16, 2025

Found an error in our documentation?Email us for assistance