listened to a few. That was painful.
I have no doubt that AI will eventually compose some amazing music. This is not it.
Good question. I was wondering the same thing. Perhaps the AI used for visual art is more sophisticated than that used for music… I’ll stop speculating though before I start to ramble.I find it extremely interesting that AI-generated visual art is vastly more palatable than AI-generated music. And I think that's true for most people, not just musicians.
Why is that?
I guess it's because more people are in need of pictures (for any kind of products, websites, novels, games and so on) and therefore there are more projects for AI imagery in development.I find it extremely interesting that AI-generated visual art is vastly more palatable than AI-generated music. And I think that's true for most people, not just musicians.
Why is that?
My hypothesis... it's just a matter of interest.I find it extremely interesting that AI-generated visual art is vastly more palatable than AI-generated music. And I think that's true for most people, not just musicians.
Why is that?
AI art tends to be static. Music is dynamic and constantly changing. AI animation may be a better comparison. I bet we could have AI do a very convincing single chord.I find it extremely interesting that AI-generated visual art is vastly more palatable than AI-generated music. And I think that's true for most people, not just musicians.
Why is that?
I agree with the reasons mentioned above. But also, because it's much simpler for a computer to analyze and reassemble digital pixels than audio files. Look at how complex a program like Melodyne is. And even Melodyne produces crude results when you heavily manipulate an audio file.I find it extremely interesting that AI-generated visual art is vastly more palatable than AI-generated music. And I think that's true for most people, not just musicians.
Why is that?
Movement in music is also a pattern that can be analyzed and reproduced by a human so it will be reproducible by AI at some point too.AI art tends to be static. Music is dynamic and constantly changing. AI animation may be a better comparison. I bet we could have AI do a very convincing single chord.
That's really just superficial. You can feed midi files to an AI. Even scanned music sheets.But also, because it's much simpler for a computer to analyze and reassemble digital pixels than audio files.
I've worked a bit with computer vision (face recognition, etc) and I don't think that's even remotely true. Both are very hard.But as it stands, it's currently much easier to do with pixels.
Might be that programmers behind it are aiming too high too fast.I find it extremely interesting that AI-generated visual art is vastly more palatable than AI-generated music. And I think that's true for most people, not just musicians.
Why is that?
I think we need to be careful to not conflate algorithmically generated art/music with AI generated art/music.Might be that programmers behind it are aiming too high too fast.
IF instead of trying to do styles/genres and commercial music, they focused on royalty free social media stuff theyd be much better. Simple kids music. Simple Ambient. Simple corporate music.
How many times have you grabbed a kontakt/synth patch where it generates a sequence from a chord you press that you felt, hey, just add a drum loops and done.
The AI guys are imo, going the wrong way about this. They are trying to use the visual AI and replicate into music, trying to capture millions of music and replicate it.
when it could be a different affair all together by just having common chord progression be randomized in key and rhythm, different drum loop beds and melodies that follow the chords and bass. YEs, much of it will sound similar but if you look at all of those short social media videos with text, the music is all very similar.
Trrying to do jazz is already difficult by a human sequencing it. Or film score where theres just too many variations.
Rock music where samples dont sound well. Classcial music is also hard even for humans.
If they focused on the kids, corporate, trap, and simpler genres with much less variations theyll get to where they want faster: keeping the rights to whatever you create and they get the royalties of your video. Plus the sub fee.
I think we need to be careful to not conflate algorithmically generated art/music with AI generated art/music.
AGORITHMS VS. AI
Algorithmic art and music have existed for quite some time. On the music side, programs and plugins that can algorithmically generate music have existed for decades. They use pre-programmed patterns and rules, augmented with some randomization, to generate a limited musical result. We can check this box as done!
But AI music is different. With AI, the engine "learns" about music by ingesting and analyzing millions of real world examples, categorizing all the relevant elements that make up a musical work, effictively building a vast body of knowledge about music.
The AI doesn't care how simple or complex a piece of music is. A one-chord ambient background and a symphony both require the same effort to ingest and analyzie. The trick is to have a large enough body of examples, and the tools to effectively break down the music into it's essential components to create meaningful classifications.
What has muddied the distinction between algorithmic and AI, is that many algorithmic engines inaccurately label themselves as "AI", when in fact, they are not AI.
EXAMPLES ARE KING
The ultimate goal with AI music is for us to be able to issue a prompt like: "Write a dirge in the style of Eminem with Brazillian percussion" and it should come up with several reasonable choices of music based on the knowledge it has accumulated about the different elements in the prompt. The AI does this by connecting and combining all the relevant bits of categorized data it has accumulated up to this point.
But if the AI was never fed any examples of Eminem music, it won't be able to satsify the request properly. Complexity is not the issue. The AI simply needs to know what is meant by "dirge", "Eminem", and "Brazillian percussion", and the only way it can do that is through exhaustive analysis of lots and lots and lots of existing examples.
BUT WE'RE EXPECTING TOO MUCH
Beyond the challenge of effectively analyzing and classifying massive volumes of musical works, I believe one of the reasons AI music has not succeeded yet is because our expectations go beyond composition. With AI art, what the computer produces is effectively a finished product: a visual image. Yes, you can tweak it further in Photoshop, but in many cases, what the AI generates is usable as a finished piece of art.
But with AI music, achieving a finished product involves far more than just composition. We're also expecting the AI to arrange and produce a final recording. That's a much bigger ask. Composing, arranging, and production are each worthy of their own AI applications.
In the video I posted above, Huawei's smartphone AI was able to successfully complete Schubert's Unfinished Symphony, but it took a human to orchestrate and arrange it properly. But in terms of pure composing, the AI actually succeeded and did it's job quite well.
I think if we only expected AI to "compose" music, so the final output was either a lead sheet or a MIDI file, then we'd probably be quite impressed with the results so far, and we might even say it's on par with AI art. But the bar has been set much higher for AI music - everyone is expecting a finished piece of recorded music.
With AI art, we simply expect the AI to replace the painter. But with AI music, we expect the AI to replace the composer, the arranger, and the recording engineer/producer. And that requires much greater knowledge and expertise than just pure composing.
All fair and valid points - I appreciate the discourse.Some of the courses i took in colege was about AI and its philosophy, so its always a very intresting and deep subject. ANd its always fasinated me.
AGORITHMS VS. AI
This deals with the notion in AI philosophy of what exactly is intelligence and Artifical Inteligence.
This is basically ye old Turing Test vs Chinese room arguments. The main notion of someone interacting with "it" does it really show signs of intelligence or is it an algorithm performing a mechanical function.
Here is a summary:
The Turing test and Searle's Chinese Room argument represent two alternative definitions of intelligence.
The Turing test is based on the assumption that intelligence is difficult to formally define but can be easily recognised by behavior. So if a computer program behaves and interacts in a way that is practically indistinguishable from the behavior of human being (who we assume is "intelligent") then, based on this assumption, we should say that the computer program is also intelligent. In the Turing test, the particular behavior that is tested is holding a conversation in natural language.
Searle's Chinese Room argument is a challenge to the validity of the Turing test. It posits a complex system that behaves as if it were intelligent (in this case, holding a conversation in Chinese), but in which each component of the system follows an algorithm, and so no part of the system can be said to "understand" Chinese. The Chinese Room argument is based on the assumption that the system as a whole cannot be said to be intelligent unless some of its individual components are intelligent - it is essentially a reductionist argument.
In summary, the Chinese Room argument says that its is possible for a system to simulate intelligence without actually being intelligent. Whereas the Turing test is says that if a system can simulate intelligence then it actually is intelligent.
We can illustrate the difference between the two points of view with an analogy as follows. The Turing test would say that an aeroplane flies because it travels through the air from one place to another - it exhibits "flying" behavior. The Chinese room argument would say that an aeroplane only simulates flight because it does not flap its wings.
In the music AI side, then what exactly do we call intelligence and what dont we call intelligence when creating music.
The idea of creating a composition of a classical style using many amalgamations of melody/counterpoint etc from the classic composers. But not rendered into record form. That could indeed be a tough argument since humans do about the same, by some standard and rules. And there was an article a while back where an algorithm would be able to detect if a song will be succesfull just by matching 80% of these rules.
But now if we come to genre specific recorded music, then what exactly is AI music and what exactly is "good" and what is the difference between algorithmic based music making.
If its like asking a prodcuer to create music and the end result is what we judge it then it might be very dependent on the actual style and producers used as a target. Trap producers and corporate music/kids music etc will churn out a lot of very similar stuff, all usable, all very human. And all can be based on an "algorithm" where they just keep changing several patterns, chord progresions etc, but so the same could be said about artificial intelligence if the intelligence the algorithm is trying to emulate just stays in this realm.
T
he key word is intelligence. Intelligence will come from an algorithm no matter what. And if that algorithm fools humans into creating the sense of intelligence when it seems more AI and "algorithm" based music. But if the standard is trap music, corporate and kids productions where they all follow similar styles, design and output than a human, then that would in itself be intelligent.
And if expand a bit and say mix corporatem music with trap, then the algorithm is just mixing the way humans do; find a pattern they like, mix it with another chord progresion they like, chnage a few things and done, you have something that resembles intelligence if its done with AI.
Mostly, the idea of intelligence is based on what the definition of intillgence is. BEcause by most accounts the thermostat in your house is intelligent. it interacts with the outside world, it adjusts itself ala self awareness, and performs a job. And theres countless exmaples of varying degress of whats considered intelligent.
Does an AI have to do something that sounds like john williams to finally say oh wow thats intelligent or can it be simple trap music or kids music where somehow it matches the intelligence of a human producer of those genres regardless of how it got it there.
so thats my argument. that the level of what we percieve of intelligence for music creation is complexity. And within this complexity we can still call artifical intelligence an algorithm that didnt have to take into account millions of classical, pop, etc style of music, analyze it and output something original. That if we lower the bar of what our definition of "intelligence" is then by all accounts AI music would very well indeed do good music as good as human producers, still be called orginal, and still can be called intelligent regardless of how it got there.