Readit News logoReadit News
sourcecodeplz commented on Ask HN: What Are You Working On? (Nov 2025)    · Posted by u/david927
sourcecodeplz · a month ago
A Chrome extension to replace the default new tab page.
sourcecodeplz commented on GPT-5 vs. Sonnet: Complex Agentic Coding   elite-ai-assisted-coding.... · Posted by u/intellectronica
bn-l · 5 months ago
Github copilot is utter garbage. The diffing crawls along at a snail’s pace. I think it’s coming up on two years and this must criticised aspect of it still isn’t fixed—-even with all the reverse engineering of how cursor did it. I wish I could find an alternative to cursor (which has other issues). Honestly, that company just threw away a golden opportunity as the first mover.
sourcecodeplz · 5 months ago
Why did they throw it away? Because of the new opaque pricing?
sourcecodeplz commented on Open models by OpenAI   openai.com/open-models/... · Posted by u/lackoftactics
modeless · 5 months ago
Nah, these are much smaller models than Qwen3 and GLM 4.5 with similar performance. Fewer parameters and fewer bits per parameter. They are much more impressive and will run on garden variety gaming PCs at more than usable speed. I can't wait to try on my 4090 at home.

There's basically no reason to run other open source models now that these are available, at least for non-multimodal tasks.

sourcecodeplz · 5 months ago
From my initial web developer test on https://www.gpt-oss.com/ the 120b is kind of meh. Even qwen3-coder 30b-a3b is better. have to test more.
sourcecodeplz commented on Dumb Pipe   dumbpipe.dev/... · Posted by u/udev4096
sourcecodeplz · 5 months ago
hamachi anyone?
sourcecodeplz commented on Qwen3-Coder: Agentic coding in the world   qwenlm.github.io/blog/qwe... · Posted by u/danielhanchen
zelphirkalt · 5 months ago
So far none of these models can write even a slightly complicated function well for me. I tried Mistral, ChatGPT, Qwen Coder 2, Claude, ... they apparently all fail when the solution requires to make use of continuations and such. Probably, because they don't have enough examples in their training data or something.

Example: Partition a linked list in linear time. None of these models seems to be able to get, that `reverse` or converting the whole list to a vector are in themselves linear operations and therefore forbid themselves. When you tell them to not use those, they still continue to do so and blatantly claim, that they are not using them. Á la:

"You are right, ... . The following code avoids using `reverse`, ... :

[code that still uses reverse]"

And in languages like Python they will cheat, because Python's list is more like an array, where random access is O(1).

This means they only work well, when you are doing something quite mainstream, where the amount of training data is a significantly strong signal in the noise. But even there they often struggle. For example I found them somewhat useful for doing Django things, but just as often they gave bullshit code, or it took a lot of back and forth to get something useful out of them.

I think it is embarrassing, that with sooo much training data, they are still unable to do much more than going by frequency in training data when suggesting "solutions". They are "learning" differently than a human being. When a human being sees a new concept, they can often apply that new concept, even if that concept does not happen to be needed that often, as long as they remember the concept. But in these LLMs it seems they deem everything that isn't mainstream irrelevant.

sourcecodeplz · 5 months ago
I use it more like documentation, I know it can't really invent things for me.
sourcecodeplz commented on Qwen3-Coder: Agentic coding in the world   qwenlm.github.io/blog/qwe... · Posted by u/danielhanchen
oblio · 5 months ago
No, it's unhealthy. Folks already have day jobs, families, other responsibilities.

Having to tack on top of that 2-4h of work per day is not normal, and again, it's probably unhealthy.

sourcecodeplz · 5 months ago
Not if you see it as a hobby.
sourcecodeplz commented on Qwen3-Coder: Agentic coding in the world   qwenlm.github.io/blog/qwe... · Posted by u/danielhanchen
stuartjohnson12 · 5 months ago
Qwen has previously engaged in deceptive benchmark hacking. They previously claimed SOTA coding performance back in January and there's a good reason that no software engineer you know was writing code with Qwen 2.5.

https://winbuzzer.com/2025/01/29/alibabas-new-qwen-2-5-max-m...

Alibaba is not a company whose culture is conducive to earnest acknowledgement that they are behind SOTA.

sourcecodeplz · 5 months ago
Benchmarks are one thing but the people really using these models, do it for a reason. Qwen team is top in open models, esp. for coding.
sourcecodeplz commented on Qwen3-Coder: Agentic coding in the world   qwenlm.github.io/blog/qwe... · Posted by u/danielhanchen
Cheer2171 · 5 months ago
You can get a used 5 year old Xeon Dell or Lenovo Workstation and 8x64GB of ECC DDR4 RAM for about $1500-$2000.

Or you can rent a newer one for $300/mo on the cloud

sourcecodeplz · 5 months ago
Everyone keeps saying this but it is not really useful. Without a dedicated GPU & VRAM, you are waiting overnight for a response... The MoE models are great but they need dedicated GPU & VRAM to work fast.
sourcecodeplz commented on Qwen3-Coder: Agentic coding in the world   qwenlm.github.io/blog/qwe... · Posted by u/danielhanchen
zettabomb · 5 months ago
Any significant benefits at 3 or 4 bit? I have access to twice that much VRAM and system RAM but of course that could potentially be better used for KV cache.
sourcecodeplz · 5 months ago
For coding you want more precision so the higher the quant the better. But there is discussion if a smaller model in higher quant is better than a larger one in lower quant. Need to test for yourself with your use cases I'm afraid.

e: They did announce smaller variants will be released.

sourcecodeplz commented on Qwen3-Coder: Agentic coding in the world   qwenlm.github.io/blog/qwe... · Posted by u/danielhanchen
danielhanchen · 5 months ago
Ye the model looks extremely powerful! I think they're also maybe making a small variant as well, but unsure yet!
sourcecodeplz · 5 months ago
Yes they are:

"Today, we're announcing Qwen3-Coder, our most agentic code model to date. Qwen3-Coder is available in multiple sizes, but we're excited to introduce its most powerful variant first: Qwen3-Coder-480B-A35B-Instruct."

https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct

u/sourcecodeplz

KarmaCake day481February 19, 2021
About
my email: metricsalliance at gmail
View Original