I understand that nowadays when we think of AI we imagine deep learning and neural networks, but nowhere on the website does it claim that that is the case.
It's a pretty hard problem. I'm about to run a Mechanical Turk survey comparing 10 melodies I've created together with my AI assistant to hit melodies:
Something I’ve noticed when programmers try to break down melodic composition to an algorithm is that they don’t grasp how important the beat or the implied beat is to creating catchy melodic phrases. Good melodies have a question/answer or a phrase/reply relationship with this beat. It’s a conversation. All memorable music has this feature. The kick asks a question and the snare answers. The kick/snare ask a question together and the harmony answers. The harmony asks a question and the melody answers. This is what differentiates sound from music.
AI generated music has promise, but their creators have got to get this. If you’re generating a beat, harmony and melody in isolation linked only by tempo and key and then sandwich them together, you’re doing it completely wrong. They are part of one another and can’t be created in isolation or with weak interaction.
Yes. All algorithmic music I've heard (with few exceptions) feels aimless and meandering because their is no intention. It's not trying to say anything. It lacks purpose and a sense of how it fits into the greater composition. It's more like it's stringing together melodies that, while passable, don't have a feeling or target to aim for short of "stay in scale".
Melodies that I hear from programs rarely have coherency from one measure to the next, and definitely not from one section to the next. Let alone coherency with it's supporting or complimentary elements.
I've used soundraw.io and aiva.ai
The first to quickly get some background music the latter to generate more original music (although harder to configure you can also edit the resulting midi directly)
It still will not replace an original composition but I can totally see a Dall-E, StableDiffusion, Midjourney moment for music on the horizon in the not too distant future
How do Soundraw know for sure the auto-generated music is original and therefore royalty-free? So not close enough to an already released, copyrighted, piece of music.
Given that not even an artist who handcrafts the whole song does know for sure I'd say they don't. They just use generic enough loops that the result is also too generic to run afoul of someone
All those AI songs sound like they are straight from sample packs previews. I am sure that in a few years we will get a real AI music generator, where you can enter a prompt like "a rock song with Freddie Mercury's vocals and lyrics about space".
No idea how it technically works, but from listening to a few samples it sounds like they've setup an environment with different instruments and generate notes to play those instruments.
The environment in all of the samples I've listened to so far sound exactly the same, as if they all share the same mixing configuration and mastering.
OpenAI's jukebox for example (check youtube for samples) is actually generating waveforms and sounds completely different.
The image equivalent to this is that I feel soundraw.io has a database of transparent stock photo images that it uses to compose a collage image that adheres to some strict rules depending on the genre.
Whereas OpenAI Jukebox is more like Dalle2 and Stable Diffusion.
Lol. Each Time I see such software, it's for paid.
Where are open source and open data geek guys?
Imagine this software or Spotify for free and open source and distributed whereas centralised.
> Be kind. Don't be snarky. Have curious conversation; don't cross-examine. Please don't fulminate. Please don't sneer, including at the rest of the community. Edit out swipes.
I would like to see an actual AI music generator with an interface based on text prompts. Soundraw seems to be algorithmic - nothing wrong with that, but they could make it clearer on their site.
By the way, this is a limitation of the paid plan:
> Sorry, you can only save 30 songs.
Edit: after trying this out for a bit, it seems like all the music is just a roughly randomized mix of the exact same loops and themes. Is anyone working on a Midjourney or Stable Diffusion for music?
This reminds me of Magix Music Maker "automatic song compositor" I forgot the exact name, but It was a function inside their software that would automatically compose a song following some settings you could choose (like which genre and sample asset you could wanted). It was not AI though.
Not a music producer but I have been using this service since launch for some of my projects. Fairly user friendly I think. Although new genres have been added, older ones haven't changed much in terms of variety of melodies.
I wonder if you'd have had this impression if you hadn't known that it was algorithmically created. Could be an interesting test to make, a turing test of sorts.
It’s very much like comparing AI generated human faces with faces from stock photos, you’d have to pay attention to weird artefacts and quirks. The use depends on what it needs to be used for, and for AI stuff I think it’s mostly for generic background that nobody cares too much to find those quirks.
This is anecdata, but the SoundRaw-generated audio I use on about 50% of my YouTube videos generates significantly more adverse comments than the stock audio on the remainder.
https://napolitano.de/2021/11/12/soundraw-ai-music-creation-...