Run ComfyUI as an API

Production-grade, zero-ops, auto-scaling

ComfyUI Interface
ComfyUI API Request

How to Run the ComfyUI API

1. Build/Test in ComfyUI Cloud

Create your own ComfyUI workflows in cloud, export the workflow's API JSON and choose which parameters you want to adjust at runtime.

Then, use Cloud Save to bundle your nodes, models, dependencies, and runtime into one reproducible container, ready to deploy as a production-grade ComfyUI API.

Build/Test in ComfyUI Cloud

2. Deploy Workflows as an API

Pick a saved workflow, choose the hardware you need, and set simple autoscaling rules. Once deployed, your ComfyUI API gets a unique deployment_id that your apps can use to send requests.

Monitor performance, scale up or down as needed, and manage multiple API versions seamlessly.

Deploy Workflows as an API

3. Scale On-Demand

Your ComfyUI API automatically scales up when requests come in and scales down to zero when things are quiet, no extra work needed.

After deployment, you can use the API endpoints to send requests, check progress, get results, or cancel jobs.

cURL
curl --request POST \ --url https://api.runcomfy.net/prod/v1/deployments/{DEPLOYMENT_ID}/inference \ --header "Authorization: Bearer YOUR_API_KEY" \ --header "Content-Type: application/json" \ --data '{' "overrides": { "6": { "inputs": { "text": "futuristic cityscape" } } "189": { "inputs": { "image": "https://example.com/new-image.jpg" } } } }

The Easiest Way to Use ComfyUI API

No-Hassle Deployment

Launch your ComfyUI API in one click from a Cloud Save. No Docker, no CUDA setup, no Kubernetes. Everything runs with the exact nodes, models, and libraries you saved, so results are always consistent.

High-Performance GPUs

Pick the GPU power you need, from 16GB (T4/A4000) to 80GB (A100/H100) and up to 141GB (H200), so you can run heavy models smoothly and reliably.

Scale On-Demand

Your API automatically scales up for traffic bursts and down to zero when idle. Control queue sizes and keep-warm settings to keep latency low and costs in check.

Workflow Versioning

Update with confidence. Manage workflow versions and use rolling updates to add features or roll back without interrupting running jobs.

Real-Time Monitoring

Stay on top of performance with a live dashboard. See request counts, queue times, cold starts, execution speed, and usage patterns to optimize your setup.

200+ Ready-to-Deploy Templates

Start fast with over 200 ready-made community workflows. Explore and customize them to fit your needs, save your version to the cloud, and deploy it as your own ComfyUI API in just minutes.

From Prototype to Production,RunComfy Makes ComfyUI APIEasier Than Ever.

자주 묻는 질문

RunComfy란 무엇이며, ComfyUI API에서 로컬 ComfyUI와 어떻게 다른가요?

RunComfy Serverless API turns your ComfyUI workflows into production-grade ComfyUI APIs with auto-scaling and no operations needed. This lets you focus on building generative AI without infrastructure worries. Unlike local ComfyUI setups that require hardware management, CUDA setup, and ongoing monitoring, RunComfy Serverless API handles deployment, scaling, and consistency in the cloud. Your ComfyUI API runs reliably on high-performance GPUs, making it easy to go from prototype to production. For more details, please read the RunComfy Serverless API documentation.

ComfyUI 워크플로우를 ComfyUI API 서비스로 배포하려면 어떻게 해야 하나요?

To deploy a ComfyUI workflow as a ComfyUI API service on RunComfy, start by building it in ComfyUI Cloud and saving it along with your nodes, models, and dependencies. Then, select GPU hardware, set autoscaling rules, and deploy with a few clicks. This creates a serverless ComfyUI API that scales automatically, processes requests asynchronously, and provides endpoints for inferences. You'll have a ready-to-use ComfyUI API without dealing with Docker, Kubernetes, or manual configurations, everything is reproducible and consistent.

ComfyUI 워크플로우를 ComfyUI API로 배포를 시작하려면 어떻게 해야 하나요?

To deploy your ComfyUI workflow as a ComfyUI API on RunComfy, start in ComfyUI Cloud where you can easily create or edit your workflow. Once it's ready, export it as a simple API JSON file and pick the parts you want to tweak during runs, like prompts or seeds—this keeps things flexible. From there, just click Cloud Save. RunComfy takes care of the rest by bundling your workflow, nodes, models, and full setup into a ready-to-use container, so you skip all the technical headaches. Finally, deploy it by selecting your preferred GPU and basic scaling options. You'll instantly get a unique deployment ID to connect your ComfyUI API to your apps or projects. The whole thing is designed to be quick and hassle-free, letting you focus on your creative ideas while getting a scalable ComfyUI API without any DevOps work. For more details, check RunComfy Serverless API - Quickstart documentation.

ComfyUI 워크플로우를 ComfyUI API 형식으로 내보내려면 어떻게 해야 하나요?

For the latest version of ComfyUI, open the ComfyUI interface, locate the Workflow menu in the upper-left corner, and select "Export (API)" from the options. This will generate a JSON file that includes all your nodes, inputs, default values, and connections. For older versions, you need to enable dev mode in the settings (click the gear icon next to Queue Size or in the menu box, then check the "Enable Dev mode Options" box), which will make the "Save (API Format)" button appear in the menu.

ComfyUI API에서 사용할 수 있는 GPU는 무엇이며, 워크플로우에 적합한 것을 어떻게 선택하나요?

RunComfy offers a range of high-performance GPUs for your ComfyUI API deployments, with VRAM from 16GB for basic workflows to 141GB for intensive models. To choose the right one for your ComfyUI API workflow, consider your model's size and memory needs, start with around 48GB (like X-Large or X-Large Plus) for most typical tasks to ensure smooth performance, then scale up or down based on testing. Monitor usage in the dashboard to optimize. For full details, visit the RunComfy Pricing page.

배포된 ComfyUI API에서 사용자 정의 노드, 모델 또는 종속성을 사용할 수 있나요?

Yes, you can easily include custom nodes, models, or dependencies in your deployed ComfyUI API. Simply add them when saving your workflow in ComfyUI Cloud, such as custom nodes, models, or specific libraries, and they'll be bundled into the container. RunComfy automatically recreates your exact environment for consistent, reliable results every time. No extra setup is required after deployment, so you can build advanced ComfyUI APIs that fit your specific needs.

RunComfy 템플릿을 사용하여 ComfyUI API를 배포할 수 있으며, 사용자 정의할 수 있나요?

Yes, RunComfy's 200+ templates let you deploy a ComfyUI API quickly, providing workflows corresponding to the latest models. Browse community workflows, fork one, tweak nodes or parameters, and save it as your own. Then deploy it as a customized ComfyUI API. All your changes stay private.

ComfyUI API를 배포한 후 API 엔드포인트는 무엇이며, 어떻게 사용하나요?

After deploying your ComfyUI API, you have endpoints for key actions: POST to queue inferences, GET to check job status or progress, GET to retrieve results like images or videos, and POST to cancel jobs. Use your deployment_id in HTTP/REST requests, with API keys for security. This asynchronous design keeps your ComfyUI API efficient, so you can track jobs easily. For full details, visit the RunComfy Serverless API - Async Queue Endpoints documentation.

ComfyUI API를 기존 기술 스택과 통합할 수 있나요?

Yes, you can easily integrate the ComfyUI API with your existing tech stack. It uses simple HTTP/REST calls and JSON data, so it works with common tools like curl, Python, or JavaScript. Check the Quickstart for ready-to-use code snippets to get started fast.

ComfyUI API의 자동 스케일링은 어떻게 작동하며, 비용 관리를 위해 제어할 수 있나요?

Auto-scaling for your ComfyUI API increases instances during busy times and scales to zero when idle, keeping things efficient. You can set min/max instances, queue sizes, and keep-warm times to fine-tune latency and costs. You're only charged for active GPU time, with no fees for downtime. This flexible control helps you run a cost-effective ComfyUI API that matches your traffic patterns.

ComfyUI API의 성능을 어떻게 모니터링하고 최적화할 수 있나요?

You can monitor your ComfyUI API with a real-time dashboard that shows request counts, queue times, cold starts, execution speeds, and usage patterns. You can also review billing data in the dashboard to track and optimize costs based on GPU time. Use these insights to adjust GPUs, scaling rules. This helps you keep your ComfyUI API running smoothly, fix issues fast, and manage expenses effectively.

다운타임 없이 ComfyUI 워크플로우를 업데이트해야 하는 경우 어떻게 되나요?

To update your ComfyUI workflow without downtime, save changes as a new version under the same name, this bundles the updates into a fresh container while keeping your live ComfyUI API running on the current version. When ready, edit the deployment to switch to the new version, which rolls out gradually: existing jobs complete on the old one, and new requests use the update. Roll back anytime by selecting a previous version. This ensures your ComfyUI API stays stable and available. For more details, refer to RunComfy Serverless API - Workflow Versions and RunComfy Serverless API - Edit a Deployment.

RunComfy에서 내 데이터는 어떻게 안전하게 보관되나요?

Your workflows run on dedicated, isolated GPUs, which guarantees complete resource separation so that no processes or memory are ever shared with other users. This ensures that your computation environment remains private and independent, providing both stability and security. Each ComfyUI execution environment, including the operating system, Python runtime, ComfyUI core, workflow definitions, models, and custom nodes, is encapsulated in its own secure cloud container. These containers are persistent, allowing your entire setup to be reliably reproduced across sessions while remaining fully private to you. Access to these environments is strictly controlled: only you can manage or expose your containerized setup, and no third party, including RunComfy, can access it unless you explicitly choose to share.

ComfyUI 워크플로우 복잡성이나 ComfyUI API 사용에 제한이 있나요?

Most ComfyUI workflows run smoothly with the ComfyUI API. However, very large models may require GPUs with higher VRAM to avoid memory-related issues. The number of concurrent jobs you can run depends on your scaling configuration, and queue limits can be adjusted to fit your workload. For high-volume or specialized needs, enterprise support is available, please reach out to us at hi@runcomfy.com.

ComfyUI API의 청구는 어떻게 작동하나요?

Billing for the ComfyUI API follows a pay-per-use model. You are only charged for the exact number of seconds your GPU is actively running, giving you full cost efficiency and flexibility. For more details, please see the RunComfy Serverless API - Billing documentation.

ComfyUI API에서 문제가 발생하면 어떤 종류의 지원을 받을 수 있나요?

If you encounter issues while using the ComfyUI API, we recommend first checking the official documentation RunComfy Serverless API – Error Handling, which covers common error codes and troubleshooting steps. If the problem persists or you need additional assistance, you can always contact us at hi@runcomfy.com.

기업이나 팀을 위한 서비스를 제공하나요?

Yes, we provide solutions tailored for enterprises and teams. For more details and customized support, please contact us directly at hi@runcomfy.com.
RunComfy
저작권 2025 RunComfy. All Rights Reserved.

RunComfy는 최고의 ComfyUI 플랫폼으로서 ComfyUI 온라인 환경과 서비스를 제공하며 ComfyUI 워크플로우 멋진 비주얼을 제공합니다. RunComfy는 또한 제공합니다 AI Playground, 예술가들이 최신 AI 도구를 활용하여 놀라운 예술을 창조할 수 있도록 지원합니다.