ComfyUI  >  Workflows  >  BAGEL AI | T2I + I2T + I2I

BAGEL AI | T2I + I2T + I2I

BAGEL AI is an open-source multimodal foundation model featuring 7B active parameters (14B total) and a Mixture-of-Transformer-Experts (MoT) design. Built for multimodal tasks like text-to-image generation, image editing, and visual question answering, BAGEL AI outperforms top-tier open VLMs such as Qwen2.5-VL and InternVL-2.5 in benchmark tests. It also provides high-quality generative capabilities on par with specialist models like SD3. With support for natural language prompting, complex reasoning, and optional transparency into the model's decision-making process, BAGEL AI offers an all-in-one solution for advanced multimodal workflows in ComfyUI.

ComfyUI BAGEL AI Workflow

BAGEL AI | Advanced Text-to-Image & Visual Chat
Want to run this workflow?
  • Fully operational workflows
  • No missing nodes or models
  • No manual setups required
  • Features stunning visuals

ComfyUI BAGEL AI Examples

bagel-ai-advanced-text-to-image-visual-chat-1229-example_01.webp
bagel-ai-advanced-text-to-image-visual-chat-1229-example_02.webp
bagel-ai-advanced-text-to-image-visual-chat-1229-example_03.webp
bagel-ai-advanced-text-to-image-visual-chat-1229-example_04.webp
bagel-ai-advanced-text-to-image-visual-chat-1229-example_05.webp
bagel-ai-advanced-text-to-image-visual-chat-1229-example_06.webp
bagel-ai-advanced-text-to-image-visual-chat-1229-example_07.webp

ComfyUI BAGEL AI Description

BAGEL AI: Multimodal Foundation Model for ComfyUI

BAGEL (BAndwidth-efficient Generalist Expert Learner) AI is a powerful multimodal foundation model designed for both image generation and vision-language understanding. Based on a 14B parameter Mixture-of-Transformer-Experts (MoT) architecture—with 7B active at inference—BAGEL AI delivers state-of-the-art performance across text-to-image generation, image editing, and image understanding tasks.

Integrated directly into ComfyUI, BAGEL AI allows creators to generate detailed images from natural language prompts, edit visuals with textual instructions, and perform multimodal tasks like visual Q&A, captioning, and step-by-step reasoning. BAGEL AI combines the quality of diffusion models (like Stable Diffusion 3) with the analytical power of leading VLMs (outperforming models like Qwen2.5-VL and InternVL-2.5).

Why Use BAGEL AI?

BAGEL AI

The BAGEL AI workflow offers:

  • Text-to-Image Generation: Create high-quality images from natural language prompts using BAGEL AI
  • Image Editing via Text: Modify existing images using descriptive instructions with BAGEL AI
  • Image Understanding: Perform image captioning, Q&A, and visual analysis tasks in BAGEL AI
  • Multimodal Reasoning: Enable step-by-step explanation or analysis of visual inputs through BAGEL AI
  • All-in-One Foundation Model: Use a single 14B MoT-based architecture for diverse multimodal tasks within BAGEL AI

With BAGEL AI, artists, researchers, and developers can explore both the generative and analytical capabilities of multimodal AI using a unified and extensible ComfyUI interface powered by BAGEL AI technology.

1 - Text-to-Image Generation with BAGEL AI

BAGEL AI

Generate Images Using Natural Language Prompts

BAGEL AI allows you to create high-quality images directly from text inputs. To get started with BAGEL AI:

  1. Enter a detailed text prompt into the Prompt input node in BAGEL AI.
  2. Optionally configure parameters like seed, aspect ratio, or decoding steps within BAGEL AI.
  3. Run the BAGEL AI workflow to generate a new image from the BAGEL model.

This BAGEL AI function is ideal for concept art, visual ideation, storytelling, or rapid prototyping using purely natural language descriptions.

2 - Image Understanding and Visual Q&A with BAGEL AI

BAGEL AI

Analyze and Understand Images Using Language

BAGEL AI includes advanced multimodal reasoning and comprehension features, making BAGEL AI ideal for image captioning, analysis, and Q&A:

  1. Upload an image to analyze in BAGEL AI.
  2. Type a question or prompt about the image in BAGEL AI (e.g., "What is the man holding?", "Describe this scene.").
  3. The BAGEL AI system returns a visual answer or reasoning trace based on the image content. This BAGEL AI feature is particularly useful for education, content tagging, accessibility workflows, or AI agents needing visual grounding through BAGEL AI capabilities.

3 - Image Editing with Textual Instructions in BAGEL AI

BAGEL AI

Modify Existing Images via Prompt-Based Editing

BAGEL AI also supports prompt-based image editing through its advanced BAGEL AI interface. Here's how to use BAGEL AI:

  1. Upload your original image in the BAGEL AI input node.
  2. Provide a text instruction describing the modification you want in BAGEL AI (e.g., "add a sunset background", "make it snow", etc.).
  3. Run the node group to apply your desired edits using BAGEL AI processing.

This allows artists and designers to non-destructively transform images through simple text without needing manual photo editing, all powered by BAGEL AI technology.

Acknowledgement

The BAGEL AI workflow for ComfyUI is based on the open-source BAGEL-7B-MoT model by ByteDance Seed. ComfyUI integration and BAGEL AI workflow setup were developed by neverbiasu, providing seamless access to image generation, editing, and understanding capabilities within a single unified BAGEL AI interface.

GitHub Repository:

BAGEL AI Model Information

  • Model Name: ComfyUI BAGEL-7B-MoT
  • Architecture: Mixture-of-Transformer-Experts (MoT) optimized for BAGEL AI
  • Total Parameters: 14B (7B Active) in BAGEL AI
  • ComfyUI Path: models/bagel/ComfyUI-BAGEL-7B-MoT/
  • Automatic Download: Enabled for BAGEL AI
  • Manual Download:

Want More ComfyUI Workflows?

RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.