ComfyUI > Nodes > ComfyUI-QuantOps > Load Checkpoint (Quantized)

ComfyUI Node: Load Checkpoint (Quantized)

Class Name

QuantizedModelLoader

Category
loaders/quantized
Author
silveroxides (Account age: 0days)
Extension
ComfyUI-QuantOps
Latest Updated
2026-03-22
Github Stars
0.04K

How to Install ComfyUI-QuantOps

Install this extension via the ComfyUI Manager by searching for ComfyUI-QuantOps
  • 1. Click the Manager button in the main menu
  • 2. Select Custom Nodes Manager button
  • 3. Enter ComfyUI-QuantOps in the search bar
After installation, click the Restart button to restart ComfyUI. Then, manually refresh your browser to clear the cache and access the updated list of nodes.

Visit ComfyUI Online for ready-to-use ComfyUI environment

  • Free trial available
  • 16GB VRAM to 80GB VRAM GPU machines
  • 400+ preloaded models/nodes
  • Freedom to upload custom models/nodes
  • 200+ ready-to-run workflows
  • 100% private workspace with up to 200GB storage
  • Dedicated Support

Run ComfyUI Online

Load Checkpoint (Quantized) Description

Facilitates loading of quantized models in ComfyUI, optimizing performance on limited hardware.

Load Checkpoint (Quantized):

The QuantizedModelLoader is a specialized node designed to facilitate the loading of quantized models within the ComfyUI framework. Its primary purpose is to handle models that have been quantized to reduce their size and improve performance, particularly in environments where computational resources are limited. This node supports various quantization formats, including INT8 and FP8, and can automatically detect the format of a given model file, allowing for seamless integration and operation. By leveraging custom operations tailored to specific quantization types, the QuantizedModelLoader ensures that models are loaded efficiently and accurately, maintaining the integrity of the model's performance while optimizing for speed and resource usage. This node is particularly beneficial for AI artists and developers who need to work with large models on hardware with limited capabilities, as it provides a streamlined and automated approach to model loading and execution.

Load Checkpoint (Quantized) Input Parameters:

quant_format

The quant_format parameter determines the quantization format of the model being loaded. It can be set to specific formats such as "int8_tensorwise", "int8_blockwise", "float8_e4m3fn_blockwise", "float8_e4m3fn_rowwise", "mxfp8", "nvfp4", or "auto" for automatic detection. This parameter impacts the selection of custom operations used during model loading, which can affect the model's performance and compatibility. The default value is "auto", which allows the node to automatically detect the appropriate format based on the model file.

ckpt_path

The ckpt_path parameter specifies the file path to the model checkpoint that needs to be loaded. This parameter is crucial as it directs the node to the exact location of the model file, enabling the loading process. The path must be accurate and accessible to ensure successful model loading. There are no specific minimum or maximum values, but it must be a valid file path.

Load Checkpoint (Quantized) Output Parameters:

model

The model output parameter represents the loaded quantized model. This output is crucial as it provides the fully constructed model ready for inference or further processing. The model is built from the state dictionary extracted from the checkpoint file, and its structure and operations are tailored based on the detected or specified quantization format. This ensures that the model operates efficiently and effectively within the constraints of the quantization method used.

Load Checkpoint (Quantized) Usage Tips:

  • Ensure that the ckpt_path is correctly specified and points to a valid model checkpoint file to avoid loading errors.
  • Use the "auto" setting for quant_format to let the node automatically detect the best-suited quantization format, which can simplify the loading process and reduce the risk of compatibility issues.
  • Familiarize yourself with the different quantization formats supported by the node to better understand how they might impact model performance and resource usage.

Load Checkpoint (Quantized) Common Errors and Solutions:

Format detection failed

  • Explanation: This error occurs when the node is unable to automatically detect the quantization format of the model file.
  • Solution: Verify that the ckpt_path is correct and that the model file is not corrupted. If the issue persists, manually specify the quant_format to bypass automatic detection.

HybridINT8Ops not available

  • Explanation: This error indicates that the necessary operations for handling INT8 quantization are not available.
  • Solution: Ensure that the required dependencies for INT8 operations are installed and correctly configured in your environment.

HybridFP8Ops not available

  • Explanation: This error suggests that the operations needed for FP8 quantization are missing.
  • Solution: Check that all necessary libraries and dependencies for FP8 operations are installed and properly set up in your system.

Load Checkpoint (Quantized) Related Nodes

Go back to the extension to check out more related nodes.
ComfyUI-QuantOps
RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Models, enabling artists to harness the latest AI tools to create incredible art.

Load Checkpoint (Quantized)