This is a simplified guide to an AI model called Musicgen-Stereo-Chord maintained by Sakemin. If you like these kinds of analysis, you should join AImodels.fyi or follow us on Twitter.
Model overview
musicgen-stereo-chord is a Cog implementation of Meta's MusicGen Melody model, created by sakemin. It can generate music based on audio-based chord conditions or text-based chord conditions, with the key difference being that it is restricted to chord sequences and tempo. This contrasts with the original MusicGen model, which can generate music from a prompt or melody.
Model inputs and outputs
The musicgen-stereo-chord model takes a variety of inputs to condition the generated music, including a text-based prompt, chord progression, tempo, and time signature. It outputs a generated audio file in either WAV or MP3 format.
Inputs
- Prompt: A description of the music you want to generate.
- Text Chords: A text-based chord progression condition, with each chord specified by a root note and optional chord type.
- BPM: The tempo of the generated music, in beats per minute.
- Time Signature: The time signature of the generated music, in the format "numerator/denominator".
- Audio Chords: An optional audio file that will be used to condition the chord progression.
- Audio Start/End: The start and end times within the audio file to use for chord conditioning.
- Duration: The length of the generated audio, in seconds.
- Continuation: Whether to continue the music from the provided audio file, or to generate new music based on the chord conditions.
- Multi-Band Diffusion: Whether to use the Multi-Band Diffusion technique to decode the generated audio.
- Normalization Strategy: The strategy to use for normalizing the output audio.
- Sampling Parameters: Various parameters to control the sampling process, such as temperature, top-k, and top-p.
Outputs
- Generated Audio: The generated music in WAV or MP3 format.
Capabilities
musicgen-stereo-chord can generate c...
Top comments (0)