Readit News logoReadit News
epolanski commented on My AI Adoption Journey   mitchellh.com/writing/my-... · Posted by u/anurag
libraryofbabel · 7 days ago
This is such a lovely balanced thoughtful refreshingly hype-free post to read. 2025 really was the year when things shifted and many first-rate developers (often previously AI skeptics, as Mitchell was) found the tools had actually got good enough that they could incorporate AI agents into their workflows.

It's a shame that AI coding tools have become such a polarizing issue among developers. I understand the reasons, but I wish there had been a smoother path to this future. The early LLMs like GPT-3 could sort of code enough for it to look like there was a lot of potential, and so there was a lot of hype to drum up investment and a lot of promises made that weren't really viable with the tech as it was then. This created a large number of AI skeptics (of whom I was one, for a while) and a whole bunch of cynicism and suspicion and resistance amongst a large swathe of developers. But could it have been different? It seems a lot of transformative new tech is fated to evolve this way. Early aircraft were extremely unreliable and dangerous and not yet worthy of the promises being made about them, but eventually with enough evolution and lessons learned we got the Douglas DC-3, and then in the end the 747.

If you're a developer who still doesn't believe that AI tools are useful, I would recommend you go read Mitchell's post, and give Claude Code a trial run like he did. Try and forget about the annoying hype and the vibe-coding influencers and the noise and just treat it like any new tool you might put through its paces. There are many important conversations about AI to be had, it has plenty of downsides, but a proper discussion begins with close engagement with the tools.

epolanski · 6 days ago
> It's a shame that AI coding tools have become such a polarizing issue among developers.

Frankly I'm so tired of the usual "I don't find myself more productive", "It writes soup". Especially when some of the best software developers (and engineers) find many utility in those tools, there should be some doubt growing in that crowd.

I have come to the conclusion that software developers, those only focusing on the craft of writing code are the naysayers.

Software engineers immediately recognize the many automation/exploration/etc boosts, recognize the tools limits and work on improving them.

Hell, AI is an insane boost to productivity, even if you don't have it write a single line of code ever.

But people that focus on the craft (the kind of crowd that doesn't even process the concept of throwaway code or budgets or money) will keep laying in their "I don't see the benefits because X" forever, nonsensically confusing any tool use with vibe coding.

I'm also convinced that since this crowd never had any notion of what engineering is (there is very little of it in our industry sadly, technology and code is the focus and rarely the business, budget and problems to solve) and confused it with architectural, technological or best practices they are genuinely insecure about their jobs because once their very valued craft and skills are diminished they pay the price of never having invested in understanding the business, the domain, processes or soft skills.

epolanski commented on GPT-5.3-Codex   openai.com/index/introduc... · Posted by u/meetpateltech
nananana9 · 7 days ago
I've been listening to the insane 100x productivity gains you all are getting with AI and "this new crazy model is a real game changer" for a few years now, I think it's about time I asked:

Can you guys point me ton a single useful, majority LLM-written, preferably reliable, program that solves a non-trivial problem that hasn't been solved before a bunch of times in publicly available code?

epolanski · 6 days ago
I know for a fact I deliver more and at higher quality and while being less tired. Mental energy is also a huge factor, because after digging in code for half a day i'd be exhausted.

People should stop focusing on vibecoding and realize how many things LLMs can do such as investigating messy codebases that took me ages of writing paper notes to connect the dots, finding information about dependencies just by giving them access to replacing painful googling and GitHub issues or outdated documentation digging, etc.

Hell I can jump in projects I know nothing about, copy paste a Jira ticket, investigate, have it write notes, ask questions and in two hours I'm ready to implement with very clear ideas about what's going on. That was multi day work till few years ago.

I can also have it investigate the task at hand and automatically find the many unknowns unknowns that as usual work tasks have, which means cutting deliveries and higher quality software. Getting feedback early is important.

LLMs are super useful even if you don't make them author a single line of code.

And yes, they are increasingly good at writing boilerplate if you have a nice and well documented codebase thus sparing you time. And in my career I've written tons of mostly boilerplate code, that was another api, another form, another table.

And no, this is not vibe coding. I review every single line, I use all of its failures to write better architectural and coding practices docs which further improves the output at each iteration.

Honestly I just don't get how people can miss the huge productivity bonus you get, even if you don't have it edit a singl line of code.

epolanski commented on GPT-5.3-Codex   openai.com/index/introduc... · Posted by u/meetpateltech
RivieraKid · 7 days ago
Do software engineers here feel threatened by this? I certainly am. I'm surprised that this topic is almost entirely missing in these threads.
epolanski · 6 days ago
Software developers should. Software engineers shouldn't.

My experience is that most developers have little to no understanding about engineering at all: meaning weighting pros and cons, understanding the requirements thoroughly, having a business oriented mindset.

Instead they think engineering is about coding practices and technologies to write better code.

That's because they focus on the code, the craft, not money.

epolanski commented on We tasked Opus 4.6 using agent teams to build a C Compiler   anthropic.com/engineering... · Posted by u/modeless
epolanski · 7 days ago
However it was achieved, building a such a complex project like a C compiler on a 20k $ budget in full autonomy is quite impressive.

Imho some commenters focus way too much on the (many, and honestly also shared by the blog post too) cons, that they forget to be genuinely impressed by the steps forward.

epolanski commented on Claude Opus 4.6   anthropic.com/news/claude... · Posted by u/HellsMaddy
EcommerceFlow · 7 days ago
Anecdotal, but it 1 shot fixed a UI bug that neither Opus 4.5/Codex 5.2-high could fix.
epolanski · 7 days ago
+1, same experience, switched model as I've read the news thinking "let's try".

But it spent lots and lots of time thinking more than 4.5, did you had the same impression.

epolanski commented on Claude Opus 4.6   anthropic.com/news/claude... · Posted by u/HellsMaddy
apetresc · 7 days ago
Impressive that they publish and acknowledge the (tiny, but existent) drop in performance on SWE-Bench Verified between Opus 4.5 to 4.6. Obviously such a small drop in a single benchmark is not that meaningful, especially if it doesn't test the specific focus areas of this release (which seem to be focused around managing larger context).

But considering how SWE-Bench Verified seems to be the tech press' favourite benchmark to cite, it's surprising that they didn't try to confound the inevitable "Opus 4.6 Releases With Disappointing 0.1% DROP on SWE-Bench Verified" headlines.

epolanski · 7 days ago
From my limited testing 4.6 is able to do more profound analysis on codebases and catches bugs and oddities better.

I had two different PRs with some odd edge case (thankfully catched by tests), 4.5 kept running in circles, kept creating test files and running `node -e` or `python 3` scripts all over and couldn't progress.

4.6 thought and thought in both cases around 10 minutes and found a 2 line fix for a very complex and hard to catch regression in the data flow without having to test, just thinking.

epolanski commented on The Missing Layer   yagmin.com/blog/the-missi... · Posted by u/lubujackson
epolanski · 7 days ago
> Documentation is hard to maintain because it has no connection to the code. Having an LLM tweak the documentation after every merge is "vibe documenting."

I'm not sure I agree, you don't need to vibe document at all.

What I do in general is: - write two separate business requirements and later implementation markdown files - keep refining the first and second one as the work progresses, stakeholders provide feedback

Before merging I have /docs updated based on requirements and implementation files. New business logic gets included in business docs (what and why), new rules/patterns get merged in architectural/code docs.

Works great and better at every new pr and iteration.

epolanski commented on Don't rent the cloud, own instead   blog.comma.ai/datacenter/... · Posted by u/Torq_boi
speedgoose · 7 days ago
I would suggest to use both on-premise hardware and cloud computing. Which is probably what comma is doing.

For critical infrastructure, I would rather pay a competent cloud provider than being responsible for reliability issues. Maintaining one server room in the headquarters is something, but two servers rooms in different locations, with resilient power and network is a bit too much effort IMHO.

For running many slurm jobs on good servers, cloud computing is very expensive and you sometimes save money in a matter of months. And who cares if the server room is a total loss after a while, worst case you write some more YAML and Terraform and deploy a temporary replacement in the cloud.

Another thing between is colocation, where you put hardware you own in a managed data center. It’s a bit old fashioned, but it may make sense in some cases.

I can also mention that research HPCs may be worth considering. In research, we have some of the world fastest computers at a fraction of the cost of cloud computing. It’s great as long as you don’t mind not being root and having to use slurm.

I don’t know in USA, but in Norway you can run your private company slurm AI workloads on research HPCs, though you will pay quite a bit more than universities and research institutions. But you can also have research projects together with universities or research institutions, and everyone will be happy if your business benefits a lot from the collaboration.

epolanski · 7 days ago
> but two servers rooms in different locations, with resilient power and network is a bit too much effort IMHO

I worked in a company with two server farms (a main and a a backup one essentially) in Italy located in two different regions and we had a total of 5 employees taking care of them.

We didn't hear about them, we didn't know their names, but we had almost 100% uptime and terrific performance.

There was one single person out of 40 developers who's main responsibility were deploys, and that's it.

It costed my company 800k euros per year to run both the server farms (hardware, salaries, energy), and it spared the company around 7-8M in cloud costs.

Now I work for clients that spend multiple millions in cloud for a fraction of the output and traffic, and I think employ around 15+ dev ops engineers.

epolanski commented on AI is killing B2B SaaS   nmn.gl/blog/ai-killing-b2... · Posted by u/namanyayg
chasd00 · 8 days ago
That situation makes sense but then i have to ask why hasn't it been done already? Software developers are not rare and if the use case is so isolated and discreet then surely it would have been tried by now. Even without genAI, CRUD, RDBMS record management, SSO, row level security... none of those things are new or out of reach until now. I think what you'll find is when you sit down with the users and start asking about the parts of the exiting system they actually need you'll never get agreement nor a clear answer. When/if you finally get a set of requirements and after UAT sign-off and then after go-live the users will say "this isn't what i meant" and you're back to square one. Rinse/repeat for years and then one day an exec will say "why are we wasting all this time, let's just subscribe to an OTS saas and make them configure it to meet our needs".
epolanski · 8 days ago
Complacency, bad management, revenue growing faster than costs for a decade hiding the problem, politics.

There's multiple people highly involved into maintaining the status quo which do everything to take any responsibility out of them.

epolanski commented on Women rejecting the hijab have doomed Iran's brutal regime   telegraph.co.uk/news/2026... · Posted by u/binning
almosthere · 8 days ago
In many interviews people keep saying that in Islamic countries the Hijab is totally optional, but that the woman want to wear it.

Now we're finding out the woman that don't want to wear it get shot.

epolanski · 8 days ago
In most muslim-majority secular countries it's more up to tradition than religion. You won't find wearing hijab that common in Albania, Kosovo, Indonesia or many of the central asian stans.

In Turkey or Jordan you will often see hijabs, but most women don't wear it.

Hell, most muslim women in western countries do not wear it.

Thus, I don't know what kind of data backs such comments.

Of course many like wearing it. As a muslim female friend of mine says "it's very convenient when you don't have time/energy to do your hair to just wear an hijab". My female, christian, SO liked wearing the hijab too in muslim countries (even in those were it's not common) because she finds them fashionable.

But I don't think women want to have to wear it and given the choice it seems clear most don't.

u/epolanski

KarmaCake day11888March 14, 2020
About
Freelancer/consultant. Product engineer.

Mostly interested in functional programming and effect systems.

contact me at

enricopolanski at gmail dot com

View Original