We're on a mission to disrupt the corrupt two-party system by building tools that change the rules—and we need your help. GoodParty.org is not a political party; we're a fully remote, US-based team united around making democracy more accessible, transparent, and fair. If creatively disrupting politics for good sounds like a challenge you're up for, check out the roles we're looking to fill right now:
- Chief Technical Officer • Engineering • Full-time
- Content and Communications Director • Growth • Full-time
- Data Engineer • Engineering • Full-time
- Director of Engineering • Engineering • Full-time
- Growth Marketer • Growth • Full-time
- Political Associate • Politics • Contract
- Product Design Manager • Design • Full-time
- Product Marketing Manager • Growth • Full-time
- Senior Full Stack Engineer • Engineering • Full-time
- Senior Product Manager • Product • Full-time
- Social Media Content Manager • Growth • Full-time
- Staff Full Stack Engineer • Engineering • Full-time
- User Success Manager • Operations • Full-time
Work with us: https://goodparty.org/work-with-us/
I thought so too, but sometimes I had better results with one sentence prompt (+README.md) where it delivered the exact thing I wanted. I also had a very detailed prompt with multiple subtasks, all were very detailed +README.md +AGENTS.md and results were very poor.
> This is the single most impressive code-gen project I’ve seen so far. I did not think this was possible yet.
To get that sort of acclaim, a human had to build an embedded programming language from scratch to get to that point. And even with all that effort, the agent itself took $631 and 119 hours to complete the task. I actually don't think this is a knock on the idea at all, this is the direction I think most engineers should be thinking about.
That agent-built HTTP/2 server they're referencing is apparently the only example of this sort of output they've seen to date. But if you're active in this particular space, especially on the open source side of the fence, this kind of work is everywhere. But since they don't manifest themselves as super generic tooling that you can apply to broad task domains as a turnkey solution, they don't get much attention.
I've continually held the line that if any given LLM agent platform works well for your use case and you haven't built said agent platform yourself, the underlying problem likely isn't that hard or complex. For the hard problems, you gotta do some first-principles engineering to make these tools work for you.
Augment does not have a model picker. It uses Claude 3.7 right now. The context engine is the magic sauce. It’s miles ahead of all the other tools, almost always gets it right where others fail.
Their UX is absolutely the best as well. Excluding maybe Claude Code (completely different tho). You can argue what's better, what deserves to stand as a standalone product, but Cursor is easily the most mature 'vibe coding' tool imo.
Hopefully LLMs won't vanish away... I'd be at a net negative.