But music doesn't have a form that ties the audio portion so cleanly to a textual description. And when it does, those labels tend to be overly simplified and not really helpful. Music is, in fact, hard to describe.
One project that might interest you is Every Noise at Once[0], which does an amazing job of grouping known artists and songs by their sonic similarly, which results in similar style and listener appeal.
I thought about that quite a bit as I was researching. Image Generation Models were given an image and a caption. This allows CLIPT to work.
But then I thought, music actually has an official notation system of the Standard notation on 5-line musical staves.
Do you think this described the music? or merely informs how it is to be performed?
I'd perceive that as too many. Leave out the one 4 days before and it'd be fine IMO.
Large banks and institutions can afford to pay past regulatory & infra barriers but in our time building APIs for brokerage like TradeStation and interviewing developers/entrepreneurs we found that not only does Fintech have one of the highest startup failure rates but most of those failures can boil down to regulation, data fees, and infra hurdles that fintech entrepreneurs need to jump over. We’re designed to do all of the leg work so that builders can focus on building and getting to market from day one.
To show off the power of our engine product we built two, soon to be open source, flagship products people can use to trade US equities and options. The first is Aries Infinite which is an infinite canvas product for trading and the second is a low latency mobile trading app for US capital markets. You can see them both at Aries.com
How did you come up with the idea for Aries?