Something interesting... the first 10 seconds or so of the "Death Growl" example[1] is basically copied verbatim from "Ov Fire And The Void" by Behemoth.
More specifically, I think the part that seems copied is at 2:13 of the original[2], as it leads into a solo-ish bit which in the AI version sounds similar still, but goes on to do its own thing:
> Additionally, our memorization-effect experiments in Section 11 demonstrate that our design maintains creativity without plagiarizing, even under strong training set conditioning.
That decision was ridiculous. It's pretty obvious that the Robin Thicke song is a $1.50 Great Value version of "Got To Give It Up" because of the aesthetic similarities but they have nothing to do with each other melodically or harmonically... "Blurred Lines" sounds like I V with a walk at the end whereas "Got To Give It Up" is more like a I IV V. The vocal melodies aren't the same nor is the bass. They have different arrangements. The percussion isn't the same.
The only things they have in common are vibes (in the contemporary sense, not vibraphones). Two dudes singing about sex in falsetto at 120bpm over prototypical R&B/funk elements isn't special. If that's the bar for copyright infringement then 99% of the popular music canon is illegally-derivative. Marvin Gaye was a singular talent but that doesn't mean that his heirs should be able to collect money every time somebody plays an electric piano bassline and sings about making whoopie in alto II.
It was ridiculous but I think it’s important to note it was a jury trial:
1) They had a musician come in and deconstruct the songs, and she showed that many of Blurred Lines elements copied the “rhythm” and “feel” of the sheet music, not the master recording.
2) Robin Thicke said in an interview he told Pharrell they should make something with the groove of Got To Give It Up.
As a non-musical person in a jury, those points are convincing enough that there was intent to copy the song even if the final song is clearly different. Though, it should never have been down to a jury and judge to decide.
Even more hilarious, a couple of years ago the Gayes tried to sue Pharrell because of an interview where he mentioned again Got To Give It Up was inspiration for Blurred Lines. Luckily that failed, but they definitely have it in for Pharrell it seems.
That same jury would have a very hard time processing the fact that all music is based upon the music that came before. Before you know it the JS Bach estate (if there is such a thing) ends up owning all of the music made after the 16th century and we should all be very lucky that Hildegard von Bingen was a nun from a very early age.
Does it matter? If the AI "comes up" with "Let it Be" melody on kazoo, it wont match "Let it Be" the Beatles single either, but it will still be plagiarized.
I get the incentives for full-song generation models, but it looks like it's the only thing that pops up. Where are the audio models I can use with a positive effect while working on music? Style transfer for instruments, restoration (but not just trained on over-compressed mp3s, talking about bad recording conditions), audio-to-audio editing?
You'd think those would be easier to achieve than something that tries to just replace me completely.
Very nice. Anyone know of projects that aren't tackling the full-song problem but rather instrument parts/loops/stems/acapellas? I'd like something that's more like "infinite AI Loopcloud/Splice" most of these full-song models don't do well to be asked for individual parts in my experience (though I will have to try it with this one).
This gets discussed a lot but unfortunately there's just not much out there around this.
The closest thing I've seen is virtual drummers in Logic X which will follow along with the structure of your song and generate a percussive accompaniment. It's no substitute for a real drummer but it's serviceable.
Yeah. Or like, a loop that plays continuously and has style parameters exposed you can tweak with a controller like a Midi Fighter Twister and get feedback from in real-time. Then you could do something akin to DJ/live production by having two of these going in sync with each other into a mixer. (Tweak params of the cue track until you like it, transition at a phrase point, repeat).
Yesssss. Something I'd really like is some kind of virtual drummer pedal with an expression pedal (pedals?) so that it can be tweaked in real time.
One of the most fun things about jamming with other musicians is that things change with the ebb and flow of the band's energy. i always miss it when I'm just fooling around at home with a looper.
What is the use case for music generation models? I see usecases for alot of the other foundation models like text, image, tts, sst, but why do I want AI generated music?
> What is the use case for music generation models?
New types of electronic instruments.
We’ve been able to use analog circuits, digital logic, and then computers to generate sounds for decades… aka synthesizers.
I would love to see synthesizers which use music generation models under the hood to create new sounds. And / or new interfaces to create different types of instruments.
There’s a lot to explore here, in the same way there was (is) lots of exploration of electronic music starting I suppose with the Theremin in the 1930s.
A major subset is people pretending to be musicians but not care to make music themselves. They just want the title for free. That's the demographic that used loops off of Splice (not even crate digging) + note generators and such and called it a day.
Another more valid subset would be something like a music bed for a video or podcast etc.
A third use is for spamming streaming platforms and making money off undiscerning suckers.
I’ve mostly used them for laughs with my friends. Sometimes generating “custom” songs with funny lyrics, but most fun so far is editing lyrics of existing songs to say ridiculous things for fun.
No real clue how someone would use them for a more serious endeavor, only thing I could imagine would be to quickly iterate/prototype with song structures on a fixed seed to generate ideas for a real composition. Consider the case of an indie game developer or film maker getting some placeholder music to test the experience during early throwaway iterations.
streamers and youtubers are constantly looking for royalty free music options.
i could see something like this baked into an editing tool that allowers video editors to specify a tone or style of music in plain language to serve as background music.
It's the only way for Spotify to turn a profit without all the human work of having to scout, sign and promote flesh-and-blood artists, the things that real music labels do.
An actual serious answer is to help musicians brainstorm while writing. It's so good at helping me come up with ideas, or converting an idea to another genre.
If the computer is fast enough to keep up with real time music, one could imagine a model that can jam with you with other instruments, reacting to and complementing what you're doing.
Not using streaming services help somewhat, but I'm not looking forward to having to vet artists I come across for whether they did any substantial original thinking or work for themselves. Tired of snake oil foolishness.
I'd imagine that would not get much traction, as there are probably people that would pay you to do that. Since people pay to listen to music, and not to make it, AI has much more utility there. The great thing is, if you are someone who gets joy in making music, AI in no way limits this.
Something interesting... the first 10 seconds or so of the "Death Growl" example[1] is basically copied verbatim from "Ov Fire And The Void" by Behemoth.
More specifically, I think the part that seems copied is at 2:13 of the original[2], as it leads into a solo-ish bit which in the AI version sounds similar still, but goes on to do its own thing:
[1] https://map-yue.github.io/music/moon.death_metal.mp3
[2] https://youtu.be/vAmnsKKrt9w?t=133
> Additionally, our memorization-effect experiments in Section 11 demonstrate that our design maintains creativity without plagiarizing, even under strong training set conditioning.
https://arxiv.org/html/2503.08638v1#S11
Funny because since the Blurred Lines lawsuit you can be infringing for using the same chord progression.
That decision was ridiculous. It's pretty obvious that the Robin Thicke song is a $1.50 Great Value version of "Got To Give It Up" because of the aesthetic similarities but they have nothing to do with each other melodically or harmonically... "Blurred Lines" sounds like I V with a walk at the end whereas "Got To Give It Up" is more like a I IV V. The vocal melodies aren't the same nor is the bass. They have different arrangements. The percussion isn't the same.
The only things they have in common are vibes (in the contemporary sense, not vibraphones). Two dudes singing about sex in falsetto at 120bpm over prototypical R&B/funk elements isn't special. If that's the bar for copyright infringement then 99% of the popular music canon is illegally-derivative. Marvin Gaye was a singular talent but that doesn't mean that his heirs should be able to collect money every time somebody plays an electric piano bassline and sings about making whoopie in alto II.
It was ridiculous but I think it’s important to note it was a jury trial:
1) They had a musician come in and deconstruct the songs, and she showed that many of Blurred Lines elements copied the “rhythm” and “feel” of the sheet music, not the master recording.
2) Robin Thicke said in an interview he told Pharrell they should make something with the groove of Got To Give It Up.
As a non-musical person in a jury, those points are convincing enough that there was intent to copy the song even if the final song is clearly different. Though, it should never have been down to a jury and judge to decide.
Even more hilarious, a couple of years ago the Gayes tried to sue Pharrell because of an interview where he mentioned again Got To Give It Up was inspiration for Blurred Lines. Luckily that failed, but they definitely have it in for Pharrell it seems.
That same jury would have a very hard time processing the fact that all music is based upon the music that came before. Before you know it the JS Bach estate (if there is such a thing) ends up owning all of the music made after the 16th century and we should all be very lucky that Hildegard von Bingen was a nun from a very early age.
Ah but that would be perfect because then all music would be freely available via public domain, recordings exempted of course.
The youtube link is suddenly not available any more (at least in the UK)
Does Shazam think it is the same?
Does it matter? If the AI "comes up" with "Let it Be" melody on kazoo, it wont match "Let it Be" the Beatles single either, but it will still be plagiarized.
I get the incentives for full-song generation models, but it looks like it's the only thing that pops up. Where are the audio models I can use with a positive effect while working on music? Style transfer for instruments, restoration (but not just trained on over-compressed mp3s, talking about bad recording conditions), audio-to-audio editing?
You'd think those would be easier to achieve than something that tries to just replace me completely.
Very nice. Anyone know of projects that aren't tackling the full-song problem but rather instrument parts/loops/stems/acapellas? I'd like something that's more like "infinite AI Loopcloud/Splice" most of these full-song models don't do well to be asked for individual parts in my experience (though I will have to try it with this one).
This gets discussed a lot but unfortunately there's just not much out there around this.
The closest thing I've seen is virtual drummers in Logic X which will follow along with the structure of your song and generate a percussive accompaniment. It's no substitute for a real drummer but it's serviceable.
https://suno.com/studio-waitlist Just a waitlist so far, but looks like this is the direction suno is going
Yeah... I hope this is what their plan is with that, but I'm not entirely certain.
Try https://magenta.withgoogle.com/infinite-crate
Honestly surprised no one's made a true "AI-Splice" yet
Also live AI dueting would be interesting, like having a virtual guitarist you could jam/duet with.
Just saw this today: https://x.com/jesseengel/status/1953496623696556478
Yeah. Or like, a loop that plays continuously and has style parameters exposed you can tweak with a controller like a Midi Fighter Twister and get feedback from in real-time. Then you could do something akin to DJ/live production by having two of these going in sync with each other into a mixer. (Tweak params of the cue track until you like it, transition at a phrase point, repeat).
Yesssss. Something I'd really like is some kind of virtual drummer pedal with an expression pedal (pedals?) so that it can be tweaked in real time.
One of the most fun things about jamming with other musicians is that things change with the ebb and flow of the band's energy. i always miss it when I'm just fooling around at home with a looper.
Like this? https://aistudio.google.com/apps/bundled/promptdj?showPrevie...
Oh! The idea that each parameter could be associated with a text style prompt is a refinement of the idea for sure.
What is the use case for music generation models? I see usecases for alot of the other foundation models like text, image, tts, sst, but why do I want AI generated music?
> What is the use case for music generation models?
New types of electronic instruments.
We’ve been able to use analog circuits, digital logic, and then computers to generate sounds for decades… aka synthesizers.
I would love to see synthesizers which use music generation models under the hood to create new sounds. And / or new interfaces to create different types of instruments.
There’s a lot to explore here, in the same way there was (is) lots of exploration of electronic music starting I suppose with the Theremin in the 1930s.
A major subset is people pretending to be musicians but not care to make music themselves. They just want the title for free. That's the demographic that used loops off of Splice (not even crate digging) + note generators and such and called it a day.
Another more valid subset would be something like a music bed for a video or podcast etc.
A third use is for spamming streaming platforms and making money off undiscerning suckers.
The use case: you are a talent-less idiot who wants to call yourself a musician whit these generated shit.
Now you don’t need to know how to make music! You’re finally free of all those pesky, elitist musicians gate-keeping music!!!!1!
this isnt reddit
And yet your comment is 100% reddit level, from a throw-away account nonetheless.
Content creators often need background music without worrying about licensing
Flood Spotify with trillions of AI-generated songs. Even if only 0.00001% of them earns a penny, they will make bank.
I’ve mostly used them for laughs with my friends. Sometimes generating “custom” songs with funny lyrics, but most fun so far is editing lyrics of existing songs to say ridiculous things for fun.
No real clue how someone would use them for a more serious endeavor, only thing I could imagine would be to quickly iterate/prototype with song structures on a fixed seed to generate ideas for a real composition. Consider the case of an indie game developer or film maker getting some placeholder music to test the experience during early throwaway iterations.
streamers and youtubers are constantly looking for royalty free music options.
i could see something like this baked into an editing tool that allowers video editors to specify a tone or style of music in plain language to serve as background music.
It's the only way for Spotify to turn a profit without all the human work of having to scout, sign and promote flesh-and-blood artists, the things that real music labels do.
Generating crappy background music for reality TV?
Yeah, this seems the most likely. Just loads of royalty free background music for industries that want that.
People love to create custom birthday songs etc for you. It's sickening how soulless this feels.
All the warmth of your friends getting Olive Garden waiters sing "Happy Birthday" to you!
Put more slop on Spotify hoping it will generate some amount of revenue.
Games?
An actual serious answer is to help musicians brainstorm while writing. It's so good at helping me come up with ideas, or converting an idea to another genre.
That's called 'jamming' and it's been possible for musicians since the time cavemen figured out they could hit sticks on things to create a rhythm.
If the computer is fast enough to keep up with real time music, one could imagine a model that can jam with you with other instruments, reacting to and complementing what you're doing.
thats basically existed at least since 1990 with band in a box
So to copy AI slop parts?
Not using streaming services help somewhat, but I'm not looking forward to having to vet artists I come across for whether they did any substantial original thinking or work for themselves. Tired of snake oil foolishness.
Between AI-generated content and people chasing trends or algorithms, it can feel like everyone's just remixing the same stuff endlessly
did a fun experiment around this in 2015 based in karpathy’s unreasonable effectiveness of RNN’s post.
i kept plucking away at it until i got it to a point where it could generate sheet music and guitar tabs in the style of various artists.
would be fun to revisit that project with fresh eyes.
yeah, but have yall made any progress in a model that can have sex with my partner for me?
I'd imagine that would not get much traction, as there are probably people that would pay you to do that. Since people pay to listen to music, and not to make it, AI has much more utility there. The great thing is, if you are someone who gets joy in making music, AI in no way limits this.
>The great thing is, if you are someone who gets joy in making music, AI in no way limits this.
Sure it does. It commodifies your passion.