ComfyUI Node: Remote Text Encoder

Class Name

RemoteTextEncoder

Category
Shrug Nodes/Text
Author
fblissjr (Account age: 4014days)
Extension
Shrug-Prompter: Unified VLM Integration for ComfyUI
Latest Updated
2025-09-30
Github Stars
0.02K

How to Install Shrug-Prompter: Unified VLM Integration for ComfyUI

Install this extension via the ComfyUI Manager by searching for Shrug-Prompter: Unified VLM Integration for ComfyUI
  • 1. Click the Manager button in the main menu
  • 2. Select Custom Nodes Manager button
  • 3. Enter Shrug-Prompter: Unified VLM Integration for ComfyUI in the search bar
After installation, click the Restart button to restart ComfyUI. Then, manually refresh your browser to clear the cache and access the updated list of nodes.

Visit ComfyUI Online for ready-to-use ComfyUI environment

  • Free trial available
  • 16GB VRAM to 80GB VRAM GPU machines
  • 400+ preloaded models/nodes
  • Freedom to upload custom models/nodes
  • 200+ ready-to-run workflows
  • 100% private workspace with up to 200GB storage
  • Dedicated Support

Run ComfyUI Online

Remote Text Encoder Description

Transforms text into embeddings via external service for precise AI model enhancement.

Remote Text Encoder:

The RemoteTextEncoder is a specialized node designed to transform text into meaningful embeddings using an external service. This node leverages the embeddings endpoint provided by heylookitsanllm to obtain real model embeddings, ensuring that the data returned is accurate and not artificially generated. By sending text to the /v1/embeddings endpoint, it retrieves embeddings that can be used for various applications, such as enhancing AI models or improving text analysis. The node is particularly beneficial for those who require precise and reliable text embeddings, as it integrates seamlessly with external APIs to fetch and process data efficiently. Its ability to handle batch processing and optional normalization of embeddings makes it a versatile tool for AI artists looking to incorporate advanced text encoding into their workflows.

Remote Text Encoder Input Parameters:

context

The context parameter is essential as it provides the necessary configuration and environment settings required for the node to function correctly. It typically includes provider-specific configurations, such as API keys and base URLs, which are crucial for establishing a connection with the embeddings endpoint. This parameter ensures that the node operates within the correct context and accesses the appropriate resources.

text

The text parameter is the primary input for the node, representing the string of text that you wish to encode into embeddings. It supports multiline input, allowing for the encoding of complex or lengthy text passages. This parameter is crucial as it directly influences the embeddings generated by the node.

normalize

The normalize parameter is a boolean option that determines whether the output vectors should be normalized. By default, it is set to True, which means the embeddings will be adjusted to have a consistent scale, potentially improving the performance of downstream tasks that utilize these embeddings.

batch_texts

The batch_texts parameter allows you to input a list of texts for batch processing. This is particularly useful when you need to encode multiple texts simultaneously, as it can significantly reduce processing time and improve efficiency. It is an optional parameter, and if not provided, the node will process the single text input.

dimensions

The dimensions parameter specifies the number of dimensions to which the embeddings should be truncated. It accepts integer values ranging from 0 to 4096, with a default of 0, indicating that the full dimension of the embeddings should be used. Adjusting this parameter can help manage the size and complexity of the embeddings, depending on your specific needs.

cache_embeddings

The cache_embeddings parameter is a boolean option that controls whether the generated embeddings should be cached for future use. By default, it is set to True, enabling caching to improve performance by avoiding redundant computations for the same input text.

debug_mode

The debug_mode parameter is a boolean option that, when enabled, provides additional debugging information during the execution of the node. This can be helpful for troubleshooting and understanding the internal workings of the node, especially if you encounter issues or unexpected results.

Remote Text Encoder Output Parameters:

conditioning

The conditioning output represents the processed state of the input text, ready for use in further AI model conditioning tasks. It is a crucial component for models that require pre-processed input data.

latent

The latent output provides the latent representation of the input text, capturing its underlying features and characteristics. This output is valuable for tasks that involve deep learning models, where latent spaces are often utilized.

embedding_tensor

The embedding_tensor output is the core result of the node, containing the numerical embeddings of the input text. These embeddings are used in various applications, such as similarity analysis, clustering, and more.

dimension

The dimension output indicates the dimensionality of the generated embeddings, providing insight into the size and complexity of the data. This information is useful for understanding the structure of the embeddings and ensuring compatibility with other components.

debug_info

The debug_info output contains detailed debugging information, which can be invaluable for diagnosing issues and understanding the node's execution process. It includes status messages, response data, and other relevant details.

Remote Text Encoder Usage Tips:

  • Ensure that the context parameter is correctly configured with the necessary provider settings, such as API keys and base URLs, to establish a successful connection with the embeddings endpoint.
  • Utilize the batch_texts parameter to encode multiple texts at once, which can significantly improve processing efficiency and reduce the time required for large-scale text encoding tasks.
  • Consider enabling debug_mode if you encounter issues or need to understand the node's internal processes better. This will provide additional insights and help identify potential problems.

Remote Text Encoder Common Errors and Solutions:

Embeddings API error (404): Embeddings endpoint not found

  • Explanation: This error occurs when the node cannot locate the /v1/embeddings endpoint, possibly due to incorrect configuration or the endpoint not being implemented.
  • Solution: Verify that the heylookitsanllm service is running and that the endpoint is correctly implemented. Check the provider configuration in the context parameter to ensure the base URL is accurate.

Could not connect to embeddings endpoint

  • Explanation: This error indicates a connection issue between the node and the embeddings endpoint, which could be due to network problems or incorrect endpoint configuration.
  • Solution: Ensure that the heylookitsanllm service is operational and accessible. Check your network connection and verify that the base URL in the context parameter is correct.

Request to embeddings endpoint timed out

  • Explanation: This error suggests that the request to the embeddings endpoint took too long to complete, possibly due to network latency or server issues.
  • Solution: Try increasing the timeout setting if possible, or check the server's performance and network conditions to identify any bottlenecks.

Encoding failed: <specific error message>

  • Explanation: This is a generic error message indicating that the encoding process encountered an unexpected issue.
  • Solution: Enable debug_mode to gather more detailed information about the error. Review the debug_info output for clues and consult the node's documentation or support resources for further assistance.

Remote Text Encoder Related Nodes

Go back to the extension to check out more related nodes.
Shrug-Prompter: Unified VLM Integration for ComfyUI
RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Models, enabling artists to harness the latest AI tools to create incredible art.

Remote Text Encoder