← the notebook
8 min · by tim grossmann ·

the ai music landscape in 2026: which tool is for what.

the ai music category in 2026 is the most crowded outside of chatbots, and people keep lumping together tools that are nothing alike. a friend asking “which ai music tool should i use?” could be told to use suno, mubert, or flowy. three tools that solve completely different problems.

this is the map as i see it, with no marketing-speak.

the categories

there are four real categories. most tools sit in one of them. occasionally a tool tries to straddle two and the result is usually weaker than either dedicated tool.

  1. song creators. you describe a song, you get a song you can iterate on, save, and release. single-track output. for creators.
  2. royalty-free stock-music tools. background audio for content creators. youtube, podcasts, twitch. subscription-based with explicit commercial licensing.
  3. listener tools. ai-generated streams for personal listening. the music keeps going. you don't deliver anything, you just listen.
  4. research and model platforms. apis and models for engineers to build their own products on top of.

song creators

suno is the category leader. their model handles vocals across most major languages, has fine-grained controls for structure and lyrics, and the community around it is large enough to be its own surface. if you want a single song, this is the right tool.

udio is the close mirror. similar capabilities, different audio character. some people prefer udio for instrumentals and orchestral, suno for vocal-forward stuff. the difference is smaller than partisans claim.

riffusion still exists and still has its niche, particularly for genre experiments and oddball outputs the more polished models smooth over.

royalty-free stock-music tools

mubert is the established player. they target youtubers and podcasters who need an explicit royalty-free license. tag-based controls (bpm, intensity, mood, genre) rather than prose prompts.

soundraw targets the same audience with a focus on customization. duration, intensity curves, section editing. popular with youtubers specifically.

aiva sits adjacent: ai-composed orchestral and instrumental music with a focus on cinematic and game soundtracks. they target a slightly more pro audience than mubert.

listener tools

this is the smallest and youngest category, and it's where flowy lives.

brain.fm is the most established listener tool. their pitch is functional music engineered for focus, sleep, and relaxation, with a neuroscience story behind it. you pick a preset (focus, relax, sleep) and the music plays. it's well-tuned for the jobs it picks.

endel is the other established player. ai-generated soundscapes tied to context signals (time of day, weather, heart rate via wearables). more ambient-focused than flowy or brain.fm.

flowy is the prose-input bet. instead of picking a preset or feeding context signals, you describe the moment you're in. “late night kitchen after a long day”. “morning run, sunrise tempo”. “saturday warmup before guests arrive”. the stream tunes to whatever you typed.

spotify's algorithmic radio (focus playlists, lofi station, chill mix) is the de-facto competitor for all three. it's cheaper, the catalog is huge, and it's familiar. the playlist-loop problem is what each listener tool is trying to solve in a different way.

research and model platforms

stability ai (stable audio) publishes models and runs a consumer product, but the deeper play is the model being downloadable for developers to host. meta's musicgen / audiocraft is open-weights and powers a lot of indie experiments. google's lyria is the underpinning of their youtube generative experiments.

for a builder, this layer is where the action is. the consumer products are often thin wrappers around one of these models, and the moat is the experience around the model rather than the model itself.

which one is for you

the honest answer depends entirely on what you're trying to do:

  • make a song to release on spotify: suno or udio.
  • background music for a monetized youtube video: mubert or soundraw, because of the explicit license.
  • focus music for deep work, neuroscience-style: brain.fm.
  • cinematic / orchestral for video projects: aiva.
  • continuous background music for any moment you can describe: flowy.
  • building your own music product: stable audio open weights, or one of the apis.

where the category is heading

a few predictions, marked as opinions:

  1. consolidation in song creation. suno and udio will keep absorbing the head terms. niche song generators won't survive unless they're wildly better at a specific style.
  2. the royalty-free category will commoditize. generative music is going to be a feature in adobe premiere, capcut, and canva within 18 months. the standalone tools will need a stronger product story than “same thing, different login.”
  3. listener tools will grow. the streaming generation got used to having music in the background of everything. they want streaming music, not single-song delivery. the only question is which input model wins (presets vs prose vs context).
  4. licensing will harden. the current ambiguity around ai-generated music copyright isn't stable. courts will settle on something. probably more permissive than the major labels want, less permissive than some users assume.
  5. real artist collaboration models (paying a human artist for their style, training a model on it, revenue-sharing on generated output) will become a viable business model alongside pure generation.

the part nobody is really doing yet

the gap i still see, and which flowy is trying to fill: ai music tools have mostly been built for creators (make a song) or for content producers (background music with a license). almost none have been built for the listener: the person who wants music in the background of their life and would rather describe it than pick from a catalog.

the bet is that this is the biggest category by a wide margin. most music consumption isn't active listening, it's soundtracking a day. if that's right, the listener-tool category is where the volume actually lives.

i could be wrong. people might just keep using spotify forever. but the pattern of complaint i hear, especially from long-time streamers, is the repetition problem. and that's exactly the problem a continuous generated stream solves.

keep reading

the ai music landscape in 2026: which tool is for what · Flowy