Yet, the most unexpected thing happened this year on my team of 4 senior/staff-level developers:
Instead of "splintering/pairing off with AI" individually even further, we wound up quadrupling (mobbing) full-time on our biggest project to date. That meant four developers, synchronously, plus Claude Code typing for us, working on one task at a time.
That was one of the most fun, laser-focused and weirdly effective way of combining our XP practice with people and AI.
Please shoot me an email at tanya@tinfoil.sh, would love to work through your use cases.
I just posted the results of another basic interview analysis (4o vs. Llama4) here: https://x.com/SpringStreetNYC/status/1923774145633849780
To your point: Do I understand correctly that, for example, by running the default model of Llama4 via ollama, the context window is very short even when the model's context is, like 10M. In order to "unlock" the full context version, I need to get the unquantized version.
For reference, here's what `ollama show llama4` returns: - parameters 108.6B # llama4:scount - context length 10485760 # 10M - embedding length 5120 - quantization Q4_K_M
Java syntax isn't perfect, but it is consistent, and predictable. And hey, if you're using an Idea or Eclipse (and not notepad, atom, etc), it's just pressing control-space all day and you're fine.
Java memory management seems weird from a Unix Philosophy POV, till you understand whats happening. Again, not perfect, but a good tradeoff.
What do you get for all of these tradeoffs? Speed, memory safety. But with that you still still have dynamic invocation capabilities (making things like interception possible) and hotswap/live redefinition (things that C/CPP cannot do).
Perfect? No, but very practical for the real world use case.
Edit: 1.4, not 1.7
For context:
My wife does leadership coaching and recently used vanilla GPT-4o via ChatGPT to summarize a transcript of an hour-long conversation.
Then, last weekend we thought... "Hey, let's test local LLMs for more privacy control. The open source models must be pretty good in 2025."
So I installed Ollama + Open WebUI plus the models on a 128GB MacBook Pro.
I am genuinely dumbfounded about the actual results we got today of comparing ChatGPT/GPT-4o vs. Llama4, Llama3.3, Llama3.2, DeepSeekR1 and Gemma.
In short: Compared to our reference GPT-4o output, none (as in NONE, zero, zilch, nil) of the above-mentioned open source models were able to create even a basic summary based on the exact same prompt + text.
The open source summaries were offensively bad. It felt like reading the most bland, generic and idiotic SEO slop I've read since I last used Google. None of the obvious topics were part of the summary. Just blah. I tested this with 5 models to boot!
I'm not an OpenAI fan per se, but if this is truly OS/SOTA then, we shouldn't even mention Llama4 or the others in the same breath as the newer OpenAI models.
What do you think?
Feedback: First off, I really like your app's style. I love bold colors. The screenshots and text are clear and understandable - maybe except on how the data gets in there. Even if that's by hand, I still think this is a great first version and a solid product.
While I'm not in your workout target group - nor on iOS - it still resonates with me because I use Oura (the ring) specifically for their detailed heart-rate tracking and stress tracking. My most-used feature in their app is my stress-tracking throughout the day.
Feature request: Only to explain how data gets inserted.
> You're an artist.
> A good one.
> Nope, a great one.
> But you have a sh*tty site.
> You wanna make it better.
> You call the guy.
> Never replies.
> ...
IF this is true (I can't say as I'm not an artist on Spotify), then this alone can sell your product.