Gosh I wish I could hire their marketing company.
Deleted Comment
The original "worst" quote is implying SOTA either stays the same (we keep using the same model) or gets better.
People have been predicting that progress will halt for many years now, just like the many years of Moore's law. By all indications AI labs are not running short of ideas yet (even judging purely by externally-visible papers being published and model releases this week).
We're not even throwing all of what is possible on current hardware technology at the issue (see the recent demonstration chips fabbed specifically for LLMs, rather than general purpose, doing 14k tokens/s). It's true that we may hit a fundamental limit with current architectures, but there's no indication that current architectures are at a limit yet.
If Windows XP were fully supported today I still wouldn't use it, personally, despite having respect for it in its era. The core technology of how, eg OS sandboxing, security, memory, driver etc stacks are implemented have vastly improved in newer OSes.
Deleted Comment
If google has no obligation to provide the service tier, then they should stop providing it instead of providing it under false terms.
This is like if everyone in a city decided to take baths instead of showers, so the municpal water supply decided to ban baths instead of properly segmenting their service based on usage.
Service providers don't have the right to discriminate what their service is used for.
Google's API does let you use any client.
The gemini/antigravity clients are a different (subscription) service. When you reverse engineer the clients and use their internal auth/apis you will typically have very different access patterns to other clients (eg: not using prompt caching), and this is likely showing up in their metrics.
This isn't unusual. A bottomless drink at a restaurant has restrictions: it's for you to drink, not to pass around to others at the table (unless they buy one too). You can't pour it into bottles to take large quantities home, etc. And it's priced accordingly: if sharing/bottling was allowed the price would have to increase.
What a wonderful way to stop people from using your LLM.
All these AI companies trying to get everyone to be locked into their toolchains is just hilariously short sighted. Particularly for dev tools. It's the sure path to get devs to hate your product.
And for what? The devs are already paying a pretty penny to use your LLM. Why do you also need to force them to using your toolkit?
This isn't a sudden change, either: they were always up-front that subscriptions are for their own clients/apps, and API is for external clients. They don't document the internal client API/auth (people extracted it).
I think a more valid complaint might be "The API costs too much" if you prefer alternative clients. But all providers are quite short on compute at the moment from what I hear, and they're likely prioritising what they subsidise.
Everything is happening at the same time in every country. It’s clearly being coordinated.
Building architectural styles used to be per city and now buildings look roughly the same worldwide. Style is dependent on the year built not the location.
Because every architect is "reading the same magazine" worldwide now that the internet exists, rather than debating in their own city.
Similar monoculture of global thought is happening in all fields.