Skip to main content
RunComfy provides three API products. They share the same high-level flow (submit > get a request_id > fetch status/results), but they solve different problems.

Which API should I use?

Use this as a quick decision guide:
What you are trying to doRecommended APIWhat you call withDeployment required?
Run a model from the RunComfy Models catalog (or a hosted pipeline)Model APImodel_idNo
Run inference with a LoRA without deploying anythingModel APImodel_id + LoRA inputsNo
Turn a ComfyUI workflow into a production endpoint (versions, autoscaling, webhooks, instance proxy)Serverless API (ComfyUI)deployment_idYes
Serve a LoRA behind a dedicated, scalable endpointServerless API (LoRA)deployment_idYes
One important mental model:
Both Serverless API (LoRA) and Serverless API (ComfyUI) are built on the same serverless deployment system. The difference is what you deploy and therefore what the request schema looks like.

Getting started


Common request pattern

Most RunComfy endpoints are asynchronous:
  1. Submit a job (POST …/inference) > get a request_id
  2. Poll status (GET …/status) until it completes
  3. Fetch outputs (GET …/result) or use webhooks for push-based updates
If you are deploying workflows (Serverless API), you can also manage the deployment lifecycle (create/update/delete) and interact with live instances through the Instance Proxy.