Readit News logoReadit News
fgonzag commented on GOG's new owner can't stand Windows either: 'It's such poor-quality software '   pcgamer.com/software/wind... · Posted by u/truxs
dahrkael · a month ago
GOG's business relies on Windows going above and beyond to remain retrocompatible with older versions of itself so trashing on it feels like a shitty move
fgonzag · a month ago
wine is probably better for backwards compatibility than modern windows / microsoft.
fgonzag commented on Bubblewrap: A nimble way to prevent agents from accessing your .env files   patrickmccanna.net/a-bett... · Posted by u/0o_MrPatrick_o0
grewil2 · a month ago
It won’t save you from prompt injektions that attack your network.
fgonzag · a month ago
In theory the docker container should only have the projects directory mounted, open access to the internet, and thats it. No access to anything else on the host or the local network.

Internet to connect with the provider, install packages, and search.

It's not perfect but it's a start.

fgonzag commented on Anthropic: Developing a Claude Code competitor using Claude Code is banned   twitter.com/SIGKITTEN/sta... · Posted by u/behnamoh
oblio · a month ago
> but I could have gone with Nvidia and had much less issues (for double the cost, dual Blackwell's vs quad Radeon W7900s for 192GB of VRAM).

> If you spend twice what I did and go Nvidia you should have nearly no issues running any models.

I goodled what a Radeon W7900 costs and the result on Amazon was €2800 a piece. You say "quad" so that's €11200 (and that's just the GPUs).

You also say "spend twice what I did", which would put the total hardware costs at ~€25000 total.

Excuse me, but this is peak HN detachment from the experience of most people. You propose spending the cost of a car on hardware.

The average person will just pay Anthropic €20 or €100 per month and call it a day, for now.

fgonzag · a month ago
I see a ton of my peers driving around in 80k cars. I drive a 20k used one.

I'm planning a writing a ROCM inference engine anyways, or at least contributing to the rocm vllm or sglang implementations for my cards since I'm interested in the field. Funnily enough, I wouldn't consider myself bullish on AI, I just want to really learn the field so I can evaluate where it's heading.

I spent about 10k on the cards, though the upgrades were piece meal as I found them cheap. I still have to get custom water blocks for them since the original W7900s (which are cheap) are triple slot, so you can't fit 4 of them in any sort of workstation setup (I even looked at rack mount options).

Bought a used thread ripper pro wrx80 motherboard ($600), I bought the cheapest TR Pro CPU for the MB (3945wx, $150), I bought 3 128Gb DDR4-3200 sticks at 230 each before the craze, was planning on populating all 8 channels if prices went down a bit. Each stick is now 900, more than I paid for all 3 combined (730 with S&H and taxes). So the system is staying as is until prices come down a bit.

For AI assisted programming, the best value prop by far is Gemini (free) as the orchestrator + open code using either free models or grok / minimax / glm through their very cheap plans (for minimax or glm) or open router which is very cheap. You can also find some interest providers like Cerebras, who get silly fast token generation, which enables interesting cases.

fgonzag commented on Anthropic: Developing a Claude Code competitor using Claude Code is banned   twitter.com/SIGKITTEN/sta... · Posted by u/behnamoh
oblio · a month ago
> I went from max 20x and chatgpt pro to Claude pro and chat gpt plus + open router providers, and I have now cancelled Claude pro and gpt plus, keeping only Gemini pro (super cheap) and using open router models + a local ai workstation I built using minimax m2.1 and glam 4.7. I use Gemini as the planner and my local models as the churners. Works great, the local models might not be as good as opus 4.5 or sonnet 4.7, but they are consistent which is something I had been missing with all commercial providers.

You went from a 5 minute signup (and 20-200 bucks per month) to probably weeks of research (or prior experience setting up workstations) and probably days of setup. Also probably a few thousand bucks in hardware.

I mean, that's great, but tech companies are a thing because convenience is a thing.

fgonzag · a month ago
My first switch was to open code + open router. I used it to try mixing models for different tasks and to try open weights models before committing to the hardware.

Even paying API pricing it was significantly cheaper than the nearly $500 I was paying monthly (I was spending about $100 month combined between Claude pro, chat gpt plus, and open router credits).

Only when I knew exactly the setup I wanted locally did I start looking at hardware. That part has been a PITA since I went with AMD for budget reasons and it looks like I'll be writing my own inference engine soon, but I could have gone with Nvidia and had much less issues (for double the cost, dual Blackwell's vs quad Radeon W7900s for 192GB of VRAM).

If you spend twice what I did and go Nvidia you should have nearly no issues running any models. But using open router is super easy, there are always free models (grok famously was free for a while), and there are very cheap and decent models.

All of this doesn't matter if you aren't paying for your AI usage out of pocket. I was so Anthropics and OpenAIs value proposition vs basically free Gemini + open router or local models is just not there for me.

fgonzag commented on Anthropic: Developing a Claude Code competitor using Claude Code is banned   twitter.com/SIGKITTEN/sta... · Posted by u/behnamoh
falloutx · a month ago
Opencode is much better anyway and it doesnt change its workflow every couple weeks.
fgonzag · a month ago
Yeah, honestly this is a bad move on anthropic's part. I don't think their moat is as big as they think it is. They are competing against opencode + ACP + every other model out there, and there are quite a few good ones (even open weight ones).

Opus might be currently the best model out there, and CC might be the best tool out of the commercial alternatives, but once someone switches to open code + multiple model providers depending on the task, they are going to have difficulty winning them back considering pricing and their locked down ecosystem.

I went from max 20x and chatgpt pro to Claude pro and chat gpt plus + open router providers, and I have now cancelled Claude pro and gpt plus, keeping only Gemini pro (super cheap) and using open router models + a local ai workstation I built using minimax m2.1 and glam 4.7. I use Gemini as the planner and my local models as the churners. Works great, the local models might not be as good as opus 4.5 or sonnet 4.7, but they are consistent which is something I had been missing with all commercial providers.

fgonzag commented on OLED, Not for Me   nuxx.net/blog/2026/01/09/... · Posted by u/c0nsumer
aappleby · a month ago
This guy is complaining about fringing...on 9- and 10-pixel high fonts. That works out to 1.6mm or 1.8mm high characters on a 140 dpi screen, or about 1/16 of an inch.

He's also got Cleartype on and set to RGB stripe even though the OLED is not RGB stripe (though to be fair, Windows doesn't really make it clear what each page of the ClearType tuner does).

But yeah, if you use a _tiny_ font and sit _really_ close to the screen, you see fringing. In practice for me, it's been unnoticeable.

fgonzag · a month ago
I have a 49" QD-OLED panel. I have never been one to find visual artifacts distracting, but fonts were awfully jaggy in Linux to the point I spent a week tinkering with font config and almost switched panels to a larger miniled since code looked horrible. And I'm someone who was fine with horrible VA low res low quality screens back in the day.

The sub pixel geometry on samsung's qd-oled needs very specific font configuration to be correctly displayed, and even then it just stops looking bad.

Deleted Comment

fgonzag commented on Maybe the default settings are too high   raptitude.com/2025/12/may... · Posted by u/htk
chrisfosterelli · 2 months ago
I often do this in meetings and have gotten into the habit of saying "I'm thinking". It's not much but it gives both of us time to think and explicitly makes it clear I don't expect the person to say something. I think that helps.
fgonzag · 2 months ago
I just blurt out "processing" when they start looking at me weird. People tend to take it well.
fgonzag commented on GLM-4.7: Advancing the Coding Capability   z.ai/blog/glm-4.7... · Posted by u/pretext
anonzzzies · 2 months ago
I have been using 4.6 on Cerebras (or Groq with other models) since it dropped and it is a glimpse of the future. If AGI never happens but we manage to optimise things so I can run that on my handheld/tablet/laptop device, I am beyond happy. And I guess that might happen. Maybe with custom inference hardware like Cerebras. But seeing this generate at that speed is just jaw dropping.
fgonzag · 2 months ago
Apple's M5 Max will probably be able to run it decently (as it will fix the biggest issue with the current lineup, prompt processing, in addition to a bandwidth bump).

That should easily run an 8 bit (~360GB) quant of the model. It's probably going to be the first actually portable machine that can run it. Strix Halo does not come with enough memory (or bandwidth) to run it (would need almost 180GB for weights + context even at 4 bits), and they don't have any laptops available with the top end (max 395+) chips, only mini PCs and a tablet.

Right now you only get the performance you want out of a multi GPU setup.

fgonzag commented on Mistral releases Devstral2 and Mistral Vibe CLI   mistral.ai/news/devstral-... · Posted by u/pember
chrsw · 2 months ago
> run locally for agentic coding. Nowadays I mostly use GPT-OSS-120b for this

What kind of hardware do you have to be able to run a performant GPT-OSS-120b locally?

fgonzag · 2 months ago
The model is 64GB (int4 native), add 20GB or so for context.

There are many platforms out there that can run it decently.

AMD strix halo, Mac platforms. Two (or three without extra ram) of the new AMD AI Pro R9700 (32GB of RAM, $1200), multi consumer gpu setups, etc.

u/fgonzag

KarmaCake day2010March 3, 2016View Original