Hacker Newsnew | past | comments | ask | show | jobs | submit | sourcecodeplz's commentslogin

A Chrome extension to replace the default new tab page.

Why did they throw it away? Because of the new opaque pricing?


They let their moat dry right up.


From my initial web developer test on https://www.gpt-oss.com/ the 120b is kind of meh. Even qwen3-coder 30b-a3b is better. have to test more.


hamachi anyone?


I use it more like documentation, I know it can't really invent things for me.


Not if you see it as a hobby.


Ergo my point about work and personal obligations (family, especially small kids). 2-4 hours per day for a solitary hobby is a surefire way to a divorce and estranged kids.


I'm married, kids, got an elderly parent at end of life that I'm caring for, and so on and so forth. How do I do it? Balance, right now, the kids are packing their bags to go to camp, so I have about 10 mins. I just replied to my prompt from last night, and will head out to drop them off, when I come in, I'll have a reply and enter my next prompt before I sign in for work. When the kids come in from school, they stay in my office and do their workbooks or watch TV while I sink in some work. You don't have to stay there for 4 straight hours, I get on the computer for 5 minutes, do a few and step out, then from that time till I get back on, I keep thinking about whatever problem I'm trying to solve.


not everyone has those personal obligations.


Benchmarks are one thing but the people really using these models, do it for a reason. Qwen team is top in open models, esp. for coding.


Everyone keeps saying this but it is not really useful. Without a dedicated GPU & VRAM, you are waiting overnight for a response... The MoE models are great but they need dedicated GPU & VRAM to work fast.


Well, yeah, you're supposed to put in a GPU. It's a MoE model, the common tensors should be on the GPU, which also does prompt processing.

The RAM is for the 400gb of experts.


For coding you want more precision so the higher the quant the better. But there is discussion if a smaller model in higher quant is better than a larger one in lower quant. Need to test for yourself with your use cases I'm afraid.

e: They did announce smaller variants will be released.


Yes the higher the quant, the better! The other approach is dynamically choosing to upcast some layers!


I can say that this really works great, I'm a heavy user of the unsloth dyanmic quants. I run DeepSeek v3/r1 in Q3, and ernie-300b and KimiK2 in Q3 too. Amazing performance. I run Qwen3-235b in both Q4 and Q8 and can barely tell the difference so much so that I just keep Q4 since it's twice as fast.


What hardware do you use, out of curiosity?


In the current era of MoE models, the system RAM memory bandwidth determines your speed more than the GPU does.


Thanks for using them! :)


Yes they are:

"Today, we're announcing Qwen3-Coder, our most agentic code model to date. Qwen3-Coder is available in multiple sizes, but we're excited to introduce its most powerful variant first: Qwen3-Coder-480B-A35B-Instruct."

https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct


Oh yes fantastic! Excited for them!


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: