Hacker Newsnew | past | comments | ask | show | jobs | submit | more film42's commentslogin

Very cool post. If Jeff Geerling is reading this, I wouldn't mind watching a video on each of these ;)


Check out saveitforparts on youtube, he does lots of this stuff.


Gabe just quit his dayjob to go full-time on SaveItForParts, so hopefully we'll be seeing even more cool stuff in the near future. Me personally, I'm hoping for a collab between him and Jeff. They've had some interaction already (Jeff donated Gabe a spare computer he had lying around) so maybe... That would be epic if it did happen.


They already did a quick collab with Jeff's PIs in space video. But I'm definitely hoping for more.


Aaah. I haven't watched that one yet, so I didn't even know. I'll definitely check that out later today.


I think I update Vundle like once every 3 years.


The 1M token context was Gemini's headlining feature. Now, the only thing I'd like Claude to work on is tokens counted towards document processing. Gemini will often bill 1/10th the tokens Anthropic does for the same document.


Agree but pricing wise, Gemini 2.5 pro wins. Gemini input tokens are half the cost of Claude 4. Output is $5/million cheaper than Claude. But, document processing is significantly cheaper. A 5MB PDF (customer invoice) with Gemini is like 5k tokens vs 56k with Claude.

The only downside with Gemini (and it's a big one) is availability. We get rate limited by their dynamic QoS all the time even if we haven't reached our quota. Our GCP sales rep keeps recommending "provisioned throughput," but it's both expensive, and doesn't fit our workload type. Plus, the VertexAI SDK is kind of a PITA compared to Anthropic.


Google products are such a pain to work with from an API perspective that I actively avoid them where possible.


Is there a crowd-sourced sentiment score for models? I know all these scores are juiced like crazy. I stopped taking them at face value months ago. What I want to know is if other folks out there actually use them or if they are unreliable.


Besides the LM Arena Leaderboard mentioned by a sibling comment, if go to the r/LocalLlama/ subreddit, you can very unscientifically get a rough sentiment of the performance of the models by reading the comments (and maybe even check the upvotes). I think the crowd's knee-jerk reaction is unreliable though, but that's what you asked for.


Not anymore tho. It used to be the place to vibe-check a model ~1 year ago, but lately it's filled with toxic my team vs. your team, memes about CEOs (wtf) and general poor takes on a lot of things.

For a while it was china vs. world, but lately it's even more divided, with heavy camping on specific models. You can still get some signal, but you have to either ban a lot of accounts, or read new during different tzs so you can get some of that "i'm just here for the tech stack" vibe from posters.


Yeah, some people just can't stop acting as if tech companies were sport teams, and it gets annoying fast.


I don't really go there much anymore but, when I was, there seemed to be an innordinate amount of Chinese nationalism from young accounts speaking odd English.


This has been around for a while https://lmarena.ai/leaderboard/text/coding


openrouter usage stats


https://openrouter.ai/rankings

The new qwen3 model is not out yet.


Since the ranking is based on token usage, wouldn't this ranking be skewed by the fact that small models' APIs are often used for consumer products, especially free ones? Meanwhile reasoning models skew it in the opposite direction, but to what extent I don't know.

It's an interesting proxy, but idk how reliable it'd be.


Also, these small models are meant to be run local so not going to appear on openrouter...


Thanks for your comment! I have a few PDFs that I need to generate for groups of users every so often and since wkhtmltopdf is considered EOL, I've been forced to use chrome (which sucks to manage). I just rewrote that code to use Typst (via the typst gem) and it's so so so much better.


Looks great! I'm definitely in the market for something like this; and building on top of helm charts makes me want to try it out.

Can Canine automatically upgrade my helm charts? That would be killer. I usually stay on cloud-hosted paid plans because remembering to upgrade is not fun. The next reason is that I often need to recall the ops knowledge just after I've forgotten it.


It can apply upgrades but I don't think it solves your core problem, which is how to perform upgrades safely. Most of the time its totally fine, but sometimes a config key changes across versions.

Upgrading helm charts without some manual monitoring seems like it might still be an unsolved problem :(


Congrats to the Crunchy Data team! Thanks for making containerized postgres so easy for years and years. Wish you all the best!


Congrats to the teams! Like others have said, your pricing ends up killing adoption for my company. We ended up self-hosting Airbyte. It ain't perfect but at least we're not paying $10/GB to replicate data within our own VPC.


I'm guessing any useful use of AI has already been adopted by some volunteers. Wikipedia might be able to build tools around the parts that work well, but the volunteers will end up footing the bill for the AI spend. Wikipedia will probably pivot to building an AI research product which they can sell to universities/ b2c.


> Wikipedia will probably pivot to building an AI research product which they can sell to universities/ b2c.

Why would they do this? All of wikipedia is publicly available for any use. They literally do not have a competitive advantage (and don't seem interested in it, either).


Exactly. But using AI to summarize articles, stitch them together, etc. under the Wikipedia brand as a product is something they could easily sell. I can totally see a university buying WikiResearch™ for every student.


I don't anticipate them selling anything, ever.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: