Base Architecture
live

Illustrious XL

Illustrious XL is an SDXL-based architecture specifically trained for anime, illustration, and stylized art generation. It excels at accurate character rendering, dynamic poses, and detailed line art with strong prompt adherence for artistic styles.

Illustrious XL

API Options

Platform-level options for task execution and delivery.

taskType

string required value: imageInference

Identifier for the type of task being performed

taskUUID

string required UUID v4

UUID v4 identifier for tracking tasks and matching async responses. Must be unique per task.

outputType

string default: URL

Image output type.

Allowed values 3 values

outputFormat

string default: JPG

Specifies the file format of the generated output. The available values depend on the task type and the specific model's capabilities.

  • `JPG`: Best for photorealistic images with smaller file sizes (no transparency).
  • `PNG`: Lossless compression, supports high quality and transparency (alpha channel).
  • `WEBP`: Modern format providing superior compression and transparency support.
**Transparency**: If you are using features like background removal or LayerDiffuse that require transparency, you must select a format that supports an alpha channel (e.g., `PNG`, `WEBP`, `TIFF`). `JPG` does not support transparency.
Allowed values 3 values

outputQuality

integer min: 20 max: 99 default: 95

Compression quality of the output. Higher values preserve quality but increase file size.

webhookURL

string URI

Specifies a webhook URL where JSON responses will be sent via HTTP POST when generation tasks complete. For batch requests with multiple results, each completed item triggers a separate webhook call as it becomes available.

Learn more 1 resource

deliveryMethod

string default: sync

Determines how the API delivers task results.

Allowed values 2 values
Returns complete results directly in the API response.
Returns an immediate acknowledgment with the task UUID. Poll for results using getResponse.
Learn more 1 resource

uploadEndpoint

string URI

Specifies a URL where the generated content will be automatically uploaded using the HTTP PUT method. The raw binary data of the media file is sent directly as the request body. For secure uploads to cloud storage, use presigned URLs that include temporary authentication credentials.

Common use cases:

  • Cloud storage: Upload directly to S3 buckets, Google Cloud Storage, or Azure Blob Storage using presigned URLs.
  • CDN integration: Upload to content delivery networks for immediate distribution.
// S3 presigned URL for secure upload
https://your-bucket.s3.amazonaws.com/generated/content.mp4?X-Amz-Signature=abc123&X-Amz-Expires=3600

// Google Cloud Storage presigned URL
https://storage.googleapis.com/your-bucket/content.jpg?X-Goog-Signature=xyz789

// Custom storage endpoint
https://storage.example.com/uploads/generated-image.jpg

The content data will be sent as the request body to the specified URL when generation is complete.

safety

object

Content safety checking configuration for image generation.

Properties 2 properties
safety » checkContent

checkContent

boolean default: false

Enable or disable content safety checking. When enabled, defaults to fast mode.

safety » mode

mode

string default: none

Safety checking mode for image generation.

Allowed values 2 values
Disables checking.
Performs a single check.

ttl

integer min: 60

Time-to-live (TTL) in seconds for generated content. Only applies when outputType is URL.

includeCost

boolean default: false

Include task cost in the response.

numberResults

integer min: 1 max: 20 default: 1

Number of results to generate. Each result uses a different seed, producing variations of the same parameters.

Advanced caching mechanisms to speed up generation.

Properties 10 properties
acceleratorOptions » cacheEndStep

cacheEndStep

integer min: 1

Absolute step number to end caching. Must be greater than cacheStartStep and less than or equal to steps.

acceleratorOptions » cacheEndStepPercentage

cacheEndStepPercentage

integer min: 1 max: 100

Percentage of steps to end caching. Alternative to cacheEndStep. Must be greater than cacheStartStepPercentage.

acceleratorOptions » cacheMaxConsecutiveSteps

cacheMaxConsecutiveSteps

integer min: 1 max: 5 default: 3

Limits the maximum number of consecutive steps that can use cached computations before forcing a fresh computation.

acceleratorOptions » cacheStartStep

cacheStartStep

integer min: 0

Absolute step number to start caching. Must be less than cacheEndStep.

acceleratorOptions » cacheStartStepPercentage

cacheStartStepPercentage

integer min: 0 max: 99

Percentage of steps to start caching. Alternative to cacheStartStep. Must be less than cacheEndStepPercentage.

acceleratorOptions » teaCache

teaCache

boolean default: false

TeaCache acceleration for transformer-based models. Estimates step differences to skip redundant computations.

acceleratorOptions » teaCacheDistance

teaCacheDistance

float min: 0 max: 1 step: 0.01 default: 0.5

Controls the aggressiveness of the TeaCache feature. Lower values prioritize quality, higher values prioritize speed.

acceleratorOptions » deepCache

deepCache

boolean

DeepCache acceleration. Skips transformer computations in certain steps to speed up generation.

acceleratorOptions » deepCacheInterval

deepCacheInterval

integer min: 1

Interval for DeepCache acceleration. A value of 2 skips every other step, 3 skips two out of three, etc.

acceleratorOptions » deepCacheBranchId

deepCacheBranchId

integer min: 0

Branch ID for DeepCache acceleration. Determines which U-Net layers are skipped.

Inputs

Input resources for the task (images, audio, etc). These must be nested inside the inputs object.

inputs » seedImage

seedImage

string

Image used as a starting point for the generation (UUID, URL, Data URI, or Base64).

Learn more 3 resources
inputs » maskImage

maskImage

string

Image used to specify which areas of the seed image should be edited (UUID, URL, Data URI, or Base64).

Learn more 1 resource

Generation Parameters

Core parameters for controlling the generated content.

model

string required

Identifier of the model to use for generation.

Learn more 3 resources

positivePrompt

string min: 2 max: 3000

Text prompt describing elements to include in the generated output.

Learn more 2 resources

negativePrompt

string min: 2 max: 3000

Prompt to guide what to exclude from generation. Ignored when guidance is disabled (CFGScale ≤ 1).

Learn more 1 resource

width

integer min: 128 max: 2048 step: 8

Width of the generated media in pixels.

Learn more 2 resources

height

integer min: 128 max: 2048 step: 8

Height of the generated media in pixels.

Learn more 2 resources

seed

integer min: 0 max: 9223372036854776000

Random seed for reproducible generation. When not provided, a random seed is generated in the unsigned 32-bit range.

Learn more 1 resource

steps

integer min: 1 max: 50

Total number of denoising steps. Higher values generally produce more detailed results but take longer.

Learn more 1 resource

scheduler

string

Scheduler to use for the diffusion process.

Allowed values 75 values
Learn more 2 resources

CFGScale

float min: 0 max: 30 step: 0.01

Guidance scale representing how closely the output will resemble the prompt. Higher values produce results more aligned with the prompt.

Learn more 1 resource

strength

float min: 0 max: 1 step: 0.01 default: 0.8

Strength of the transformation. Lower values result in more influence from the original input.

Learn more 3 resources

maskMargin

integer min: 32 max: 128

Extra context pixels around the masked region during inpainting. The model zooms into the masked area with these additional pixels for better integration.

Learn more 1 resource

clipSkip

integer min: 0 max: 4

Number of layers to skip in the CLIP model.

Learn more 2 resources

Defines the syntax to be used for prompt weighting.

Prompt weighting allows you to adjust how strongly different parts of your prompt influence the generated image. Choose between compel notation with advanced weighting operations or sdEmbeds for simple emphasis adjustments.

View Compel syntax

Adds 0.2 seconds to image inference time and incurs additional costs.

When compel syntax is selected, you can use the following notation in prompts:

Weighting

Syntax: + - (word)0.9

Increase or decrease the attention given to specific words or phrases.

Examples:

  • Single words: small+ dog, pixar style
  • Multiple words: small dog, (pixar style)-
  • Multiple symbols for more effect: small+++ dog, pixar style
  • Nested weighting: (small+ dog)++, pixar style
  • Explicit weight percentage: small dog, (pixar)1.2 style

Blend

Syntax: .blend()

Merge multiple conditioning prompts.

Example: ("small dog", "robot").blend(1, 0.8)

Conjunction

Syntax: .and()

Break a prompt into multiple clauses and pass them separately.

Example: ("small dog", "pixar style").and()

View sdEmbeds syntax

When sdEmbeds syntax is selected, you can use the following notation in prompts:

Weighting

Syntax: (text) (text:number) [text]

Use parentheses () to increase attention, square brackets [] to decrease it. Add a number after the text to specify a custom multiplier.

Examples:

  • Single words: (small) dog, pixar style
  • Multiple words: small dog, [pixar style]
  • Higher emphasis: (small:2.5) dog, pixar style
  • Combined emphasis: (small dog:1.5), pixar style
Allowed values 2 values

outpaint

object

Extends image boundaries in specified directions. Final width/height must account for original image plus extensions.

Learn more 1 resource
Properties 4 properties
outpaint » bottom

bottom

integer min: 0

Number of pixels to extend to the bottom.

outpaint » left

left

integer min: 0

Number of pixels to extend to the left.

outpaint » right

right

integer min: 0

Number of pixels to extend to the right.

outpaint » top

top

integer min: 0

Number of pixels to extend to the top.

lora

array of objects min items: 1

With LoRA (Low-Rank Adaptation), you can adapt a model to specific styles or features by emphasizing particular aspects of the data. This technique enhances the quality and relevance of generated content and can be especially useful when the output needs to adhere to a specific artistic style or follow particular guidelines.

Multiple LoRA models can be used simultaneously to achieve different adaptation goals.

Examples 1 example
"lora": [
  {
    "model": "<lora-model-air>",
    "weight": 0.8
  }
]
Learn more 1 resource
Properties 3 properties
lora » model

model

string required

LoRA model identifier.

lora » weight

weight

float min: -4 max: 4 step: 0.01 default: 1

Strength of the LoRA influence. A value of 0 means no influence. Higher values increase the influence, and negative values can be used to steer away from the LoRA's style.

lora » transformer

transformer

string default: both

Transformer stages to apply LoRA. Some video models use separate high-noise and low-noise processing stages, and LoRAs can be selectively applied to optimize their effectiveness.

Allowed values 3 values
Apply LoRA only to the high-noise processing stage (coarse structure and early generation steps).
Apply LoRA only to the low-noise processing stage (fine details and later generation steps).
Apply LoRA to both stages for full coverage.

controlNet

array of objects min items: 1

With ControlNet, you can provide a guide image to help the model generate images that align with the desired structure. This guide image can be generated with our ControlNet preprocessing tool, extracting guidance information from an input image. The guide image can be in the form of an edge map, a pose, a depth estimation or any other type of control image that guides the generation process via the ControlNet model.

Multiple ControlNet models can be used at the same time to provide different types of guidance information to the model.

Examples 1 example
"controlNet": [
  {
    "model": "<controlnet-model-air>",
    "guideImage": "c64351d5-4c59-42f7-95e1-eace013eddab",
    "weight": 0.7,
    "startStep": 0,
    "endStep": 20,
    "controlMode": "controlnet"
  }
]
Learn more 2 resources
Properties 8 properties
controlNet » model

model

string required

ControlNet model identifier.

controlNet » weight

weight

float min: -4 max: 4 step: 0.01 default: 1

Strength of the ControlNet influence. A value of 0 means no influence. Higher values increase the influence, and negative values can be used to steer away from the guide image.

controlNet » guideImage

guideImage

string required

Reference image for ControlNet guidance (UUID, URL, Data URI, or Base64).

controlNet » controlMode

controlMode

string default: balanced

ControlNet guidance mode.

Allowed values 3 values
Equal weight between ControlNet and prompt.
Prioritize ControlNet guidance.
Prioritize prompt guidance.
controlNet » endStep

endStep

integer min: 1

Absolute step number to end ControlNet influence. Must be greater than startStep and less than or equal to steps.

controlNet » endStepPercentage

endStepPercentage

integer min: 1 max: 100

Percentage of steps to end ControlNet influence. Must be greater than startStepPercentage.

controlNet » startStep

startStep

integer min: 0

Absolute step number to start ControlNet influence. Must be less than endStep.

controlNet » startStepPercentage

startStepPercentage

integer min: 0 max: 99

Percentage of steps to start ControlNet influence. Must be less than endStepPercentage.

ipAdapters

array of objects min items: 1

IP-Adapters enable image-prompted generation, allowing you to use reference images to guide the style and content of your generations. Multiple IP Adapters can be used simultaneously.

Examples 1 example
"ipAdapters": [
  {
    "model": "<ip-adapter-model-air>",
    "guideImages": ["c64351d5-4c59-42f7-95e1-eace013eddab"],
    "weight": 0.75
  },
  {
    "model": "<ip-adapter-model-air>",
    "guideImages": ["d7e8f9a0-2b5c-4e7f-a1d3-9c8b7a6e5d4f"],
    "weight": 0.5
  }
]
Learn more 1 resource
Properties 7 properties
ipAdapters » model

model

string required

We make use of the AIR system to identify IP-Adapter models. This identifier is a unique string that represents a specific model.

Supported models list
AIR IDModel Name
runware:55@1 IP Adapter SDXL
runware:55@2 IP Adapter SDXL Plus
runware:55@3 IP Adapter SDXL Plus Face
runware:55@4 IP Adapter SDXL Vit-H
runware:55@5 IP Adapter SD 1.5
runware:55@6 IP Adapter SD 1.5 Plus
runware:55@7 IP Adapter SD 1.5 Light
runware:55@8 IP Adapter SD 1.5 Plus Face
runware:55@10 IP Adapter SD 1.5 Vit-G
ipAdapters » weight

weight

float min: -4 max: 4 step: 0.01 default: 1

Strength of the IP-Adapter influence. A value of 0 means no influence. Higher values increase the influence, and negative values can be used to steer away from the reference.

ipAdapters » guideImages

guideImages

array of strings required min items: 1

Images to guide the IP-Adapter (UUID, URL, Data URI, or Base64).

ipAdapters » combineMethod

combineMethod

string default: concat

Controls how multiple reference images are combined.

Allowed values 5 values
ipAdapters » embedScaling

embedScaling

string default: kv

Determines which embedding components are used and their strength.

Allowed values 4 values
ipAdapters » weightType

weightType

string default: normal

Shapes how influence evolves during generation.

Allowed values 13 values
ipAdapters » weightComposition

weightComposition

float min: 0 max: 1 step: 0.01

Controls composition/layout influence specifically.

Features

Standalone addons and post-processing features.

hiresFix

boolean | object

Two-stage generation for improved resolution and detail. The model generates at a lower resolution first, then upscales and refines the result in a second pass. Can be enabled with true for default settings, or configured as an object for fine-grained control over the upscaling model, steps, and strength.

When using the object form, the model parameter is required. Available upscaling models:

ModelNameUpscale Factor
runware:504@1 RealESRGAN_x4plus 4x
runware:realesrgan@anime-6b RealESRGAN_x4plus_anime_6B 4x
runware:esrgan@animesharp 4x-AnimeSharp 4x
runware:esrgan@ultrasharp 4x-UltraSharp 4x
"hiresFix": true
"hiresFix": {
  "model": "runware:esrgan@ultrasharp",
  "steps": 15,
  "strength": 0.6
}
Allowed values 1 value

ultralytics

object

Configuration object for Ultralytics face enhancement during generation. This feature uses face detection and inpainting to improve facial details in the same generation step, without requiring post-processing.

Face enhancement is available for Stable Diffusion 1.X, SDXL, and FLUX models. The system automatically detects faces and applies targeted refinement to improve quality while maintaining consistency with the overall generation.

Properties 8 properties
ultralytics » CFGScale

CFGScale

float min: 0 max: 50 step: 0.1 default: 8

Face refinement guidance scale.

ultralytics » confidence

confidence

float min: 0 max: 1 step: 0.01 default: 0.9

Confidence threshold for detection.

ultralytics » maskBlur

maskBlur

integer min: 0 max: 100 default: 5

Mask feathering amount. Higher values create softer transitions between the enhanced face region and surrounding areas.

ultralytics » maskPadding

maskPadding

integer min: 0 max: 20 default: 5

Padding around detected face in pixels. Expands the refinement area to include surrounding context like hair and neck.

ultralytics » negativePrompt

Negative prompt for detection.

ultralytics » positivePrompt

Positive prompt for detection.

ultralytics » steps

steps

integer min: 1 max: 100 default: 20

Number of face refinement steps.

ultralytics » strength

strength

float min: 0 max: 1 step: 0.01 default: 0.3

Refinement strength. Lower values preserve more of the original, higher values allow more aggressive reconstruction.