Very much disagree with that. Getting productive and competent with LLM tooling takes months. I've been deeply invested in this world for a couple of years now and I still feel like I'm only scraping the surface of what's possible with these tools.
Very much disagree with that. Getting productive and competent with LLM tooling takes months. I've been deeply invested in this world for a couple of years now and I still feel like I'm only scraping the surface of what's possible with these tools.
Say I have a type returned by the server that might have more sophisticated types than the server API can represent. For instance, api/:postid/author returns a User, but it could either be a normal User or an anonymous User, in which case fields like `username`, `location`, etc come back null. So in this case I might want to use a discriminated union to represent my User object. And other objects coming back from other endpoints might also need some type alterations done to them as well. For instance, a User might sometimes have Post[] on them, and if the Post is from a moderator, it might have special attributes, etc - another discriminated union.
In the past, I've written functions like normalizeUser() and normalizePost() to solve this, but this quickly becomes really messy. Since different endpoints return different subsets of the User/Post model, I would end up writing like 5 different versions of normalizePost for each endpoint, which seems like a mess.
How do people solve this problem?
E.g.
if (user.user_type === 'authenticated') {
// do something with user.name because the type system knows we have that now
}But they aren't moving nearly as fast as OpenAI. And it remains to be seen if first mover will mean anything.
Businesses are definitely rearranging themselves structurally around AI - at least to try and get the AI valuation multiplier and Executives have levels of FOMO I've never seen before. I report to a CTO and the combination of 100,000 foot hype combined with down in the weeds focus on the "protocol de jour" (with nothing in between that looks like a strategy) is astounding. I just find it exhausting.
It is still simply too early to tell exactly what the new steady state is, but I can tell you that where we're at _today_ is already a massive paradigm shift from what my day-to-day looked like 3 years ago, at least as a SWE.
There will be lots of things thrown at the wall and the things that stick will have a big impact.
Gemini 2.5 is the first model I tested that was able to solve it and it one-shotted it. I think it's not an exaggeration to say LLMs are now better than 95+% of the population at mathematical reasoning.
For those curious the riddle is: There's three people in a circle. Each person has a positive integer floating above their heads, such that each person can see the other two numbers but not his own. The sum of two of the numbers is equal to the third. The first person is asked for his number, and he says that he doesn't know. The second person is asked for his number, and he says that he doesn't know. The third person is asked for his number, and he says that he doesn't know. Then, the first person is asked for his number again, and he says: 65. What is the product of the three numbers?
(All state is stored in localStorage so you can come back to it :) ).
...but now it'll be exciting to let them bake. We need some time to really explore what we can do with them. We're still mostly operating in back-and-forth chats, I think there's going to be lots of experimentation with different modalities of interaction here.
It's like we've just gotten past the `Pets.com` era of GenAI and are getting ready to transition to the app era.
I always imagined if you could have some super mind build an entire complex system, it would find better solutions that got around limitations introduced by the need to make engineering accessible to humans.
This is missing the most interesting changes in generative AI space over the last 18 months:
- Multi-modal: LLMs can consume images, audio and (to an extent) video now. This is a huge improvement on the text-only models of 2023 - it opens up so many new applications for this tech. I use both image and audio models (ChatGPT Advanced Voice) on a daily basis.
- Context lengths. GPT-4 could handle 8,000 tokens. Today's leading models are almost all 100,000+ and the largest handle 1 or 2 million tokens. Again, this makes them far more useful.
- Cost. The good models today are 100x cheaper than the GPT-3 era models and massively more capable.
If nothing else, my workflows as a software developer have changed significantly in these past two years with just what's available today, and there is so much work going into making that workflow far more productive.
> We’re hiring engineers.
Careers page:
> Sorry, no job openings at the moment.
https://www.anthropic.com/jobs?team=4050633008