Readit News logoReadit News
platers commented on TREAD: Token Routing for Efficient Architecture-Agnostic Diffusion Training   arxiv.org/abs/2501.04765... · Posted by u/fzliu
platers · 13 days ago
I'm struggling to understand where the gains are coming from. What is the intuition for why DiT training was so inefficient?
platers commented on Open music foundation models for full-song generation   map-yue.github.io/... · Posted by u/selvan
lotyrin · 24 days ago
Very nice. Anyone know of projects that aren't tackling the full-song problem but rather instrument parts/loops/stems/acapellas? I'd like something that's more like "infinite AI Loopcloud/Splice" most of these full-song models don't do well to be asked for individual parts in my experience (though I will have to try it with this one).
platers · 24 days ago
https://suno.com/studio-waitlist Just a waitlist so far, but looks like this is the direction suno is going
platers commented on ACE-Step: A step towards music generation foundation model   github.com/ace-step/ACE-S... · Posted by u/wertyk
briga · 4 months ago
Interesting how there is no mention of how the training data for this was collected. This does sound quite a bit better than Meta's MusicGen, but then again that model was also trained on a small licensed dataset.
platers · 4 months ago
It sounds very similar to suno v3.5 (including the audio quality) Likely they trained on suno generations.
platers commented on Jagged Flash Attention Optimization   shaped.ai/blog/jagged-fla... · Posted by u/tullie
platers · 5 months ago
Flash attention natively supports packing multiple variable length sequences into a single call, what is the advantage of jagged flash attention?

u/platers

KarmaCake day247February 2, 2022View Original