I agree that modality specific processing is very shallow at this point, but it still seems not to respect the physicality of the data. Today's modalities are not actually akin to human senses because they should be processed by a different assortment of "sense" organs, e.g. one for things visual, one for things audible, etc.
Now I only wish for an Product Manager model that can render the code and provide feedback on the UI issues. Using Cursor and Gemini, we were able to get a impressively polished UI, but it needed a lot of guidance.
> I haven’t yet come across an agent that can write beautiful code.
Yes, the AI don't mind hundreds of lines of if statements, as long as it works it's happy. It's another thing that needs several rounds of feedback and adjustments to make it human-friendly. I guess you could argue that human-friendly code is soon a thing of the past, so maybe there's no point fixing that part.
I think improving the feedback loops and reducing the frequency of "obvious" issues would do a lot to increase the one-shot quality and raise the productivity gains even further.
A visual queue of reaching within the success range would do a lot too. Was a bit confused whether I was right or not after submitting the answer.
It's a simplification, since 100% of an individual's subscription doesn't go to directly to royalties, it's after expenses.
Smaller artists are frustrated with this model as it favours the big artists.
[1] https://support.spotify.com/us/artists/article/royalties/
Edit: added the link.