ComfyUI > Nodes > ComfyUI-FLOAT_Optimized > FLOAT Extract Emotion from Features (VA)

ComfyUI Node: FLOAT Extract Emotion from Features (VA)

Class Name

FloatExtractEmotionWithCustomModel

Category
FLOAT/Very Advanced
Author
set-soft (Account age: 3450days)
Extension
ComfyUI-FLOAT_Optimized
Latest Updated
2026-03-20
Github Stars
0.03K

How to Install ComfyUI-FLOAT_Optimized

Install this extension via the ComfyUI Manager by searching for ComfyUI-FLOAT_Optimized
  • 1. Click the Manager button in the main menu
  • 2. Select Custom Nodes Manager button
  • 3. Enter ComfyUI-FLOAT_Optimized in the search bar
After installation, click the Restart button to restart ComfyUI. Then, manually refresh your browser to clear the cache and access the updated list of nodes.

Visit ComfyUI Online for ready-to-use ComfyUI environment

  • Free trial available
  • 16GB VRAM to 80GB VRAM GPU machines
  • 400+ preloaded models/nodes
  • Freedom to upload custom models/nodes
  • 200+ ready-to-run workflows
  • 100% private workspace with up to 200GB storage
  • Dedicated Support

Run ComfyUI Online

FLOAT Extract Emotion from Features (VA) Description

Generates emotion latents from audio using a custom model for AI-driven artistic applications.

FLOAT Extract Emotion from Features (VA):

The FloatExtractEmotionWithCustomModel node is designed to generate emotion conditioning latents from audio features, which can be used in various AI-driven artistic applications. This node leverages a custom emotion recognition model to either predict emotions from preprocessed audio features or utilize a specified emotion to create a one-hot encoded tensor. This flexibility allows you to either let the model infer emotions dynamically from the audio input or specify an emotion directly, providing a versatile tool for emotion-based audio processing. The node's primary goal is to facilitate the extraction of emotional content from audio data, enabling the creation of emotionally aware AI models that can enhance user experiences in applications such as interactive storytelling, virtual assistants, and more.

FLOAT Extract Emotion from Features (VA) Input Parameters:

processed_audio_features

This parameter represents a batch of preprocessed audio features, typically output by a feature extractor like FloatAudioPreprocessAndFeatureExtract. It is a tensor that contains the audio data after it has been processed to highlight features relevant for emotion recognition. The quality and accuracy of the emotion extraction depend significantly on the quality of these features. There are no explicit minimum or maximum values, but the data should be appropriately preprocessed for optimal results.

emotion_model_pipe

This parameter is a tuple that includes the loaded emotion recognition model, a reference to the feature extractor used for the emotion model, and its configuration. It serves as the backbone for emotion prediction, allowing the node to utilize a pre-trained model to interpret the audio features. The model pipe must be correctly configured and loaded to ensure accurate emotion extraction.

emotion

This parameter allows you to specify a particular emotion or choose 'none' to let the model predict the emotion from the audio features. The available options are defined by the EMOTIONS list, and the default value is 'none'. Selecting a specific emotion will bypass the prediction step and directly use the specified emotion to generate the one-hot encoded tensor. This parameter provides flexibility in how emotions are handled, either by manual specification or automatic inference.

FLOAT Extract Emotion from Features (VA) Output Parameters:

we_latent

The we_latent output is a tensor that represents the emotion conditioning latent. This tensor can be used in subsequent processing steps to influence the behavior of AI models based on the detected or specified emotion. It is crucial for applications that require emotion-aware processing, as it encapsulates the emotional content derived from the audio features.

emotion_model_pipe_out

This output returns the emotion model pipe, which is essentially the same as the input emotion_model_pipe. It ensures that the model pipe is available for further processing or reuse in other nodes, maintaining consistency and continuity in the workflow.

FLOAT Extract Emotion from Features (VA) Usage Tips:

  • Ensure that the audio features are properly preprocessed using a compatible feature extractor to maximize the accuracy of emotion prediction.
  • When specifying an emotion, ensure it matches one of the available options in the EMOTIONS list to avoid prediction errors and ensure the correct one-hot encoding.

FLOAT Extract Emotion from Features (VA) Common Errors and Solutions:

Failed to map '<specified_emotion>'. Predicting emotion from audio features.

  • Explanation: This error occurs when the specified emotion does not match any of the available options in the EMOTIONS list.
  • Solution: Verify that the specified emotion is correctly spelled and matches one of the available options. If unsure, use 'none' to allow the model to predict the emotion.

Predicting emotion from audio features using custom model.

  • Explanation: This message indicates that the node is defaulting to emotion prediction because 'none' was selected or the specified emotion could not be mapped.
  • Solution: If this behavior is unintended, ensure that the emotion specified is valid. Otherwise, this message is informational and indicates normal operation when 'none' is selected.

FLOAT Extract Emotion from Features (VA) Related Nodes

Go back to the extension to check out more related nodes.
ComfyUI-FLOAT_Optimized
RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Models, enabling artists to harness the latest AI tools to create incredible art.