From our side we designed these models to be strong for their size out of the box, and with the goal you'll all finetune it for your use case. With the small size it'll fit on a wide range of hardware and cost much less to finetune. You can try finetuning them yourself in a free colab in under 5 minutes
For picking a Gemma size this is a video I recorded for the 1b to 27b sizes earlier this year, 270m being the newest addition
https://www.youtube.com/watch?v=qcjrduz_YS8
Hacker News Disclaimer I really like working at Google so with that; All my opinions here are my own, I'm a researcher so I'll largely focus on technical questions, and I'll share what I can.
Also the ANE models are limited to 512 tokens of context, so unlikely yet to use these in production.
99% of people wouldn’t be able to find the API keys you need to feed into most MCP servers.
Authentication, session management, etc, should be handled outside of the standard, and outside of the LLM flow entirely.
I recently mused on these here; https://github.com/sunpazed/agent-mcp/blob/master/mcp-what-i...
Cannot recommend it enough - absolutely fantastic library, great documentation, zero issues of any impact to us in five years.
My only wish is for the keyboard accessibility ticket[2] to get some love!
I have to say, the endorsements at the end somehow made this seem worse…
I really is just json-rpc 2.0 under the hood, either piped to stdio or POSTed over http.