Generative AI (aka GenAI) is a fast moving field, and the provision of generative tools for music specifically is no exception to this. To make sense of that, several teams have provided summaries. Now there are a lot of summaries, so here’s a summary of the summaries, and a few signposts that might be useful.

No knowledge is assumed, and we begin at the more fundamental level of defining the different musical tasks to which you GenAI might be applied.

In brief:

  • Title: ‘Generative AI Tools in Music’
  • By: Mark Gotham, 2026
  • Licence: CC-By-SA
  • Suggestions/contributions: Are welcome! PR or see email contact details here

On this page:


Tasks

There are many different tasks which GenAI can engage. Often, breaking down a large creative task into smaller component parts allows the human to have a more active role. Here are some example tasks, with notes and corresponding tools.

In this section:

“Text to Music” (text → audio)

Going from a user’s text prompt to an audio output (also called “text to music” or “TTM”) is probably the most familiar GenAI use case. As always, the results depend on many factors chief among them, the model and the prompt.

Among the many apps providing this, is Stable Audio, which we highlight for a few reasons including:

  • It comes top of many metrics for openness (see below)
  • The team have provided “Prompt Guide” here.
  • The option to combine TTM with an audio prompt.*

*This model (among others) allows the user to combine (or even replace) the text prompt with audio. In that case, we’re not talking about “text → audio” so much as “audio → audio” or even “text+audio → audio”. Use cases include the creation of new audio to fix with an existing template, where “fit” could include matching the tempo (to align in the timings) and/or the style.

Audio from scratch (nothing → audio)

What really is audio? As a data type, audio is a form of time-series data. To understand the nature of this data, here is a very basic demo for exploring raw audio sound data.

A Digital Audio Workstation (DAW, pronounced to sound like “door”) is a type of software application for handling audio, in which you can record, edit, mix, and more. These days, many DAW’s have AI features internally. Here is Wiki’s comparison of digital audio editors and here is a new open source DAW with the most pleasingly apposite name of “Open DAW”:

Further hints:

  • To create your own audio from nothing, use sound synthesis (using any number of tools including many right there in the DAW).
  • Additionally and/or alternatively, you might want to start working with existing audio files e.g., as samples and building music with these blocks, Lego-style. Again, all DAWs include some samples and here are some external sources of samples with open licences:

Symbolic from scratch (nothing → symbolic)

To engage with “the notes”, you need to access those notes. DAW’s often feature some basic (limiter) controls of this kind. Again, to understand this data a bit better, here is a very basic demo for exploring symbolic data.

To create your own symbolic data from nothing, you might use:

Alternatively, start working with existing symbolic data. There are many sites hosting musical scores, of varying and often unclear quantity and quality. By contrast, “Datasets for Symbolic Music Processing” are typically clearer. I list some datasets at the end of my “Keeping Score” book

Some systems include Generative AI for symbolic music. WeaveMuse includes that functionality (among others spanning understanding and generation).

Transcription (audio → symbolic)

Automatic Music Transcription (AMT) is the task of extracting symbolic data from raw audio.

Tools include:

Source separation (mixed audio → audio in separate parts)

As mentioned above, audio is time-series data. All the signal and noise (literally) are mixed together. For many use cases, it is necessary/helpful to separate the component parts of this sound (e.g. guitar based drums). For example, source separation can help with,

  • transcription (of music and/or lyrics),
  • alignment,
  • recognition/detection (of instruments, musicians, lyrics, …).

Models include:

See also


Use Cases

In this section:

Studies of previous use cases include experimental studies in the “lab” and surveys of uses in “the wild”. Both can give ideas for how you might use these tools.

Even where there is an experimental setup (as in Ronchini et al.), these are not typically released publicly, so this is for information and ideation only.

AI-Assisted Music Production

Dated: Nov 2025

Ronchini et al.’s “AI-Assisted Music Production” paper is an experimental study of how users engage with AI tools, especially “text to music” (TTM). studies

AI Song Contestants

Accessed: September 2025

The AI Song Contest is probably the premiere venue for AI-based composition today. Many winning and shortlisted entries from recent years are on the site, including an account of how the piece was made.

Dated: July 2025

The team at Stability AI (Jordi Pons et al.) recently reviewed how GenAI has been used in recent music artistic work “in the wild”. This covers “A collection (337 artworks) of AI music released before July 31, 2025” and includes model details.


Summaries

In this section:

AI Song Contest

Accessed: September 2025

The AI Song Contest (mentioned above) also includes a list of models, organised around use case (lyrics, composition, and more). The largest of those lists is here.

Fairly Trained

Accessed: September 2025

“Fairly trained” offers accreditation for models that are trained fairly (the clue’s in the name).

Their list of certified models is provided here.

As of September 2025, there are 19 total:

  • 12 from Fairly trained’s “Company certification”,
  • 2 from “Product certification”,
  • 5 from “Model certification”,

MusGO

Accessed: Jan 2026 (but the authors plan continuous updates)

The “MusGO Framework” from the Music Technology Group in Barcelona serves to assess openness in gen AI model for music.

The author highlight the following attributes (quote):

  • multidimensional with 13 categories
  • graded with different levels of openness
  • divided into essential and desirable categories
  • refined with feedback from +100 MIR members
  • continuously updated

The “Stable Audio Open” model (mentioned above) currently comes out on top.

Responsible AI

Dated: July 2025

The Responsible AI (RAI) team at UAL (University of the Arts, London, UK) has produced “A Short Review of Responsible AI Music Generation”. (Elizabeth Wilson et al., Proceedings of the 6th Conference on AI Music Creativity, AIMC 2025). The paper has a tabular summary (no need to replicate that here) and the team have also provided a model explorer online.

Survey on the Evaluation of Generative Models in Music

Dated: October 2025

And finally, the thorny issue of evaluation. Do these systems “work”? What does “work” mean in this context?

Here is an important “Survey on the Evaluation of Generative Models in Music” by Lerch et al.