ComfyUI > Nodes > ComfyUI_DiffRhythm_MW > DiffRhythm Run

ComfyUI Node: DiffRhythm Run

Class Name

DiffRhythmRun

Category
🎤MW/MW-DiffRhythm
Author
mw (Account age: 2258days)
Extension
ComfyUI_DiffRhythm_MW
Latest Updated
2025-05-09
Github Stars
0.09K

How to Install ComfyUI_DiffRhythm_MW

Install this extension via the ComfyUI Manager by searching for ComfyUI_DiffRhythm_MW
  • 1. Click the Manager button in the main menu
  • 2. Select Custom Nodes Manager button
  • 3. Enter ComfyUI_DiffRhythm_MW in the search bar
After installation, click the Restart button to restart ComfyUI. Then, manually refresh your browser to clear the cache and access the updated list of nodes.

Visit ComfyUI Online for ready-to-use ComfyUI environment

  • Free trial available
  • 16GB VRAM to 80GB VRAM GPU machines
  • 400+ preloaded models/nodes
  • Freedom to upload custom models/nodes
  • 200+ ready-to-run workflows
  • 100% private workspace with up to 200GB storage
  • Dedicated Support

Run ComfyUI Online

DiffRhythm Run Description

Node for generating music with AI models, lyrics, and style prompts for seamless music creation.

DiffRhythm Run:

DiffRhythmRun is a node designed to facilitate the generation of music by leveraging advanced machine learning models. It integrates various components such as a tokenizer, a model for generating music, and a mechanism for processing lyrics and style prompts. The primary goal of DiffRhythmRun is to create a seamless experience for users who wish to generate music that aligns with specific lyrical content and stylistic preferences. By utilizing a combination of text and audio prompts, this node can produce music that not only matches the lyrical timing but also adheres to a desired style, whether derived from text descriptions or audio samples. This capability is particularly beneficial for AI artists looking to explore creative musical compositions without needing extensive technical knowledge of the underlying processes.

DiffRhythm Run Input Parameters:

model

The model parameter specifies the pre-trained model to be used for music generation. It determines the maximum number of frames that can be processed, with options like "cfm_full_model.pt" allowing for a larger frame size of 6144, compared to a default of 2048. This parameter impacts the complexity and length of the generated music, with larger models potentially offering more detailed and nuanced outputs.

lyrics_prompt

The lyrics_prompt parameter is a string input that provides the lyrical content to be synchronized with the generated music. It is crucial for aligning the musical output with specific lyrics, ensuring that the timing and rhythm of the music match the provided text. This parameter directly influences the lyrical structure of the generated music.

style_audio

The style_audio parameter allows users to input an audio sample that serves as a stylistic reference for the music generation process. When provided, the node extracts stylistic elements from the audio to influence the generated music, ensuring that the output aligns with the desired audio characteristics. This parameter is optional but can significantly enhance the stylistic coherence of the music.

style_prompt

The style_prompt parameter is an alternative to style_audio, allowing users to input a textual description of the desired music style. This text-based prompt guides the model in generating music that fits the described style, offering flexibility for users who may not have an audio sample but have a clear stylistic vision. It is essential for defining the overall mood and genre of the music.

steps

The steps parameter controls the number of inference steps during the music generation process. It affects the quality and detail of the output, with higher values typically resulting in more refined music. However, setting this parameter too high may increase computation time without significant quality improvements. A value below 32 may trigger a sway sampling coefficient adjustment.

DiffRhythm Run Output Parameters:

generated_song

The generated_song output is the final music track produced by the node. It encapsulates the combined influence of the lyrics, style prompts, and model parameters, resulting in a unique musical composition. This output is the primary deliverable of the node, providing users with a ready-to-use music track that aligns with their creative inputs.

DiffRhythm Run Usage Tips:

  • Ensure that either style_audio or style_prompt is provided to guide the stylistic direction of the music, as omitting both will result in an error.
  • Experiment with different model options to find the balance between computational efficiency and output quality that best suits your needs.
  • Use the steps parameter to fine-tune the detail and quality of the generated music, but be mindful of the increased computation time with higher values.

DiffRhythm Run Common Errors and Solutions:

ValueError: Style prompt or style audio must be provided

  • Explanation: This error occurs when neither a style_audio nor a style_prompt is provided, leaving the node without guidance for the music's style.
  • Solution: Ensure that you provide either a style_audio sample or a style_prompt text to define the desired style of the music.

Exception during tokenizer encoding

  • Explanation: This error may arise if there is an issue with encoding the lyrics using the tokenizer, possibly due to unsupported characters or formatting.
  • Solution: Check the lyrics_prompt for any unusual characters or formatting issues and ensure it is compatible with the tokenizer's expected input format.

DiffRhythm Run Related Nodes

Go back to the extension to check out more related nodes.
ComfyUI_DiffRhythm_MW
RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.