Probably the parts that have been around for decades.
As someone who's created open source software in the past, I'm put off mostly the explosion in software complexity over the last two decades, the constant churn in libraries and tooling which places an unreasonable maintenance burden on a spare-time hobbyist, and the bad taste left in the mouth by Microsoft using copilot to license-launder the github corpus.
I still produce open source software and gateware on a recreational basis, but in quiet little backwaters of the noosphere which aren't going to change the world (or be tainted by AI) any time soon!
Isn’t the game with licenses that you must specify what can and cannot be done with the code? It’s a big issue that no one prior to now had the foresight to forbid ML training on code without attribution. If the licenses going back 30 years had that stipulation then it would be easy to take down Copilot and ChatGPT. But the licenses simply don’t cover the use case of training a neural net, so it’s probably going to slip through the cracks just like SaaS slipped through the cracks of the GPL by not distributing code, hence the need for AGPL. So I’m sure we’ll see these kinds of clauses added to licenses going forward, but they can’t be applied retroactively.
The irony in all this is that from the start, open source licensing has been a game of wits where software creators try to cleverly use copyright as a lever of control. Well, they weren’t clever enough. They missed ML training and didn’t forbid it. As a result they’ve basically lost the whole game.
No, not at all. Microsoft's argument is that training an LLM on code is fair use, and thus doesn't trigger copyright-based licensing at all. That's why they include unlicensed code in Copilot, which under a training-triggers-copyright theory they have no right to at all.