Hacker Newsnew | past | comments | ask | show | jobs | submit | IshKebab's commentslogin

I don't. Google has at least a few advantages:

1. Google books, which they legally scanned. No dubious training sets for them. They also regularly scrape the entire internet. And they have YouTube. Easy access to the best training data, all legally.

2. Direct access to the biggest search index. When you ask ChatGPT to search for something it is basically just doing what we do but a bit faster. Google can be much smarter, and because it has direct access it's also faster. Search is a huge use case of these services.

3. They have existing services like Android, Gmail, Google Maps, Photos, Assistant/Home etc. that they can integrate into their AI.

The difference in model capability seems to be marginal at best, or even in Google's favour.

OpenAI has "it's not Google" going for it, and also AI brand recognition (everyone knows what ChatGPT is). Tbh I doubt that will be enough.


And they have hardware as well, and their own cloud platform.

In my view Google is uniquely well positioned because, contrary to the others, it controls most of the raw materials for Ai.


Google's most significant advantage in this space is its organizational experience in providing services at this scale, as well as its mature infrastructure to support them. When the bubble pops, it's not lights-out or permanently degraded performance.

Yeah kind of conspicuously absent! They said

> The previous server was 12 year old hardware

which is pretty mad. You can buy a second hand system with tons of ram and a 16-core Ryzen for like $400. 12-year old hardware is only marginally faster than a RPi 5.


> 12-year old hardware is only marginally faster than a RPi 5

My 14yo laptop-used-as-server disagrees. Also keep in mind that CPU speeds barely improved between about 2012 and 2017, and 2025 is again a lull https://www.cpubenchmark.net/year-on-year.html

I'm also factoring in the ability to use battery bypass in phones I buy now because they are so powerful, I might want to use them as free noiseless server in the future. You can do a heck of a lot on phone hardware nowadays, paying next to nothing for power and no additional cost on your existing internet connection. A RPi 5 is that same ballpark


> 12-year old hardware is only marginally faster than a RPi 5.

A Dell R620 is over 12 years old and WAY faster than a RPi 5 though...

Sure, it'll be way less power efficient, but I'd definitely trust it to serve more concurrent users than a RPi.


Plus the fact that it's been running for 5 years. Does that mean they bought 7 year old hardware back then? Or is that just when it was last restarted?

Unfortunately you can’t even get the RAM for $400 anymore.

I was able to find 2 x 16GB DDR4 for $150...

Building a budget AM4 system for roughly $500 would be within the realm of reason. ($150 mobo, $100 cpu, $150 RAM, that leaves $100 for storage, still likely need power and case.)

https://www.amazon.com/Timetec-Premium-PC4-19200-Unbuffered-...

https://www.amazon.com/MSI-MAG-B550-TOMAHAWK-Motherboard/dp/...

For a server that's replacing a 12 year old system, you don't need DDR5 and other bleeding edge hardware.


I don't think 32GB is going to be enough lol

Also, you would want ECC for something this important.

I seriously wonder if doing the same build twice by two people in two locations wouldn't provide the same benefit and others for less money.

(I might be spoiled by sane reproducible build systems. Maybe F-droid isn't.)


"F-Droid is not hosted in a data centre with proper procedures, access controls, and people whose jobs are on the line. Instead it's in some guy's bedroom."

Not reassuring.


It could just be a colo, there are still plenty of data centres around the globe that will sell you a space in a shared rack with a certain power density per U of space. The list of people who can access that shared locked rack is likely a known quantity with most such organisations and I know in the past we had some details of the people who were responsible for it

In some respects, having your entire reputation on the line matters just as much. And sure, someone might have a server cage in their residence, or maybe they run their own small business and it's there. But the vagueness is troubling, I agree.

A picture of the "living conditions" for the server would go a long way.


Depends on the thread model, which one is worse.

State actor? Gets into data centre, or has to break into a privately owned apartment.

Criminal/3rd party state intelligence service? Could get into both, at a risk or with blackmail, threats, or violence.

Dumb accidents? Well, all buildings can burn or have an power outage.


> State actor? Gets into data centre, or has to break into a privately owned apartment.

I don’t think a state actor would actually break in to either in this case, but if they did then breaking into the private apartment would be a dream come true. Breaking into a data center requires coordination and ensuring a lot of people with access and visibility stay quiet. Breaking into someone’s apartment means waiting until they’re away from the premises for a while and then going in.

Getting a warrant for a private residence also would likely give them access to all electronic devices there as no 3rd party is keeping billing records of which hardware is used for the service.

> Dumb accidents? Well, all buildings can burn or have an power outage.

Data centers are built with redundant network connectivity, backup power, and fire suppression. Accidents can happen at both, but that’s not the question. The question is their relative frequency, which is where the data center is far superior.


>Breaking into a data center requires coordination and ensuring a lot of people with access and visibility stay quiet

Or just a warrant and a phone call to set up remote access? In the UK under RIPA you might not even need a warrant. In USA you can probably bribe someone to get a National Security Letter issued.

Depending on the sympathies of the hosting company's management you might be able to get access with promises.

I dare say F-Droid trust their friends/colleagues more than they trust randos at a hosting company.

As an F-Droid user, I think I might too? It's a tough call.


>I don’t think a state actor would actually break in to either in this case

Read Jabber.ru Hetzner accident: https://notes.valdikss.org.ru/jabber.ru-mitm/


> Data centers are built with redundant network connectivity, backup power, and fire suppression. [...] The question is their relative frequency, which is where the data center is far superior.

Well, I remember one incident were a 'professional' data center burned down including the backups.

https://en.wikipedia.org/wiki/OVHcloud#Incidents

I know no such incident for some basement hosting.

Doesn't mean much. I'm just a bit surprised so many people are worried because of the server location and no one had mentioned yet the quite outstanding OVH incident.


I'm not going to pretend datacenters are magical places immune to damage. I worked at a company where the 630 Third Street datacenter couldn't keep temperatures stable during a San Francisco heatwave and the Okex crypto exchange has experienced downtime because the Alibaba Zone C datacenter their matching engine is on experienced A/C failure. So it's not all magic, but if you didn't encounter home-lab failure it's because you did not sample the population appropriately.

https://www.reddit.com/r/homelab/comments/wvqxs7/my_homelab_...

I don't have a bone to pick here. If F-Droid wants to free-ball it I think that's fine. You can usually run things for max cheap by just sticking them on a residential Google Fiber line in one of the cheap power states and then just making sure your software can quickly be deployed elsewhere in times of outage. It's not a huge deal unless you need always-on.

But the arguments being made here are not correct.


Surely "Juan's home server in basement burns down" would make the headlines. You're totally right.

> The question is their relative frequency, which is where the data center is far superior.

as a year long f-droid user I can't complain


I think there are countless examples of worse failures by organisations that meet your criteria for far more valuable assets than some free apps.

The 'cloud' has come full circle

Eh...

The set of people who can maliciously modify it is the people who run f-droid, instead of the cloud provider and the people who run f-droid.

It'd be nice if we didn't have to trust the people who run f-droid, but given we do I see an argument that it's better for them to run the hardware so we only have to trust them and not someone else as well.


You actually do not have to trust the people who run f-droid for those apps whose maintainers enroll in reproducible builds and multi-party signing, which only f-droid supports unlike any alternatives.

That looks cool, which might just be the point of your comment, but I don't think it actually changes the argument here.

You still have to trust the app store to some extent. On first use, you're trusting f-droid to give you the copy of the app with appropriate signatures. Running in someone else's data-center still means you need to trust that data-center plus the people setting up the app store, instead of just the app store. It's just a breach of trust is less consequential since the attacker needs to catch the first install (of apps that even use that technology).


F-droid makes the most sense when shipped as the system appstore, along with pinned CA keychains as Calyxos did. Ideally f-droid was compiled from source and validated by the rom devs.

The F-droid app itself can then verify signatures from both third party developers and first party builds on an f-droid machine.

For all its faults (of which there are many) it is still a leaps and bounds better trust story than say Google Play. Developers can only publish code, and optional signatures, but not binaries.

Combine that with distributed reproducible builds with signed evidence validated by the app and you end up not having to trust anything but the f-droid app itself on your device.


None of this mitigates the fact that apriori you don't know if you're being served the same package manifest/packages as everyone else - and as such you don't know how many signatures any given package you are installing should have.

Yes, theoretically you can personally rebuild every package and check hashes or whatever, but that's preventative steps that no reasonable threat model assumes you are doing.


Why have we normalized "app stores" that build software whose authors likely already provide packages of?

I've been using Obtainium more recently, and the idea is simple: a friendly UI that pulls packages directly from the original source. If I already trust the authors with the source code, then I'm inclined to trust them to provide safe binaries for me to use. Involving a middleman is just asking for trouble.

App stores should only be distributors of binaries uploaded and signed by the original authors. When they're also maintainers, it not only significantly increases their operational burden, but requires an additional layer of trust from users.


The cloud isn't the only other option, they could still own and run their own hardware but do it in a proper colocation datacenter.

Can you use Gleam for ad-hoc scripting? In my mind that means two requirements that most languages fail at.

1. You can import by relative file path. (Python can't.)

2. You can specify third party dependencies in a single file script and have that work properly with IDEs.

Deno is the best option I've found that has both of those and is statically typed.

I'm hoping Rust will eventually too but it's going to be at least a year or two.


Typescript is a lot nicer than Python in many ways. Especially via Deno, and especially for scripting (imports work like people want!).

There are some things that aren't as good, e.g. Python's arbitrary precision integers are definitely nicer for scripting. And I'd say Python's list comprehension syntax is often quite nice even if it is weirdly irregular.

But overall Deno is a much better choice for ad-hoc scripting than Python.



I am aware, but Python has that by default. In Javascript it's opt-in and less ergonomic. E.g. try loading a 64-bit integer from JSON.

I agree, but bigints are missing from json because the json spec defines all numbers as 64 bit floats. Any other kind of number in JSON is nonstandard.

JavaScript itself supports bigint literals just fine. Just put an ‘n’ after your number literal. Eg 0xffffffffffffffn.

There’s a whole bunch of features I wish we could go in and add to json. Like comments, binary blobs, dates and integers / bigints. It would be so much nicer to work with if it has that stuff.


> the json spec defines all numbers as 64 bit floats

It absolutely doesn't. It doesn't impose any limits on number precision or magnitude.


That's not what was happening in this example though. It would be UB even if it was a POD.

Yeah that's my view too. It's definitely fine to wait a couple of years (at least), and see what emerged as most effective and then just learn that, instead of dumping a ton of time now into keeping up with the hamster wheel.

Unless you're in web dev because it seems like that's one of the few domains where AI actually works pretty well today.


Or if you like learning new stuff. Personally that has been best part of being programmer.

I love learning new stuff, but for whatever reason the AI stuff doesn’t interest me. So I learn other stuff, only so much time in the day.

I like learning new stuff, but not if it's going to be completely obsolete in 6 months.

Everyone has an obligation not to be a dick.

I meant "be a dick" only in the sense of providing a no-holds-barred critique of the product. Generalizing it is strawmanning.

Sorry what? Who did he murder?

This is the sort of lunacy that immediately gives your argument zero weight.


Search terms that will help you on your journey include “DOGE”, “Kenya”, and “cholera”.

I’m sorry that the death of hundreds of thousands of people did not register in your world, but I feel like charges of “lunacy” are a bit rich coming from someone living in a soft bubble of ignorance.


> different developers on HN have very different jobs and skill levels.

Definitely this. When I use AIs for web development they do an ok job most of the time. Definitely on par with a junior dev.

For anything outside of that they're still pretty bad. Not useless by any stretch, but it's still a fantasy to think you could replace even a good junior dev with AI in most domains.

I am slightly worried for my job... but only because AI will keep improving and there is a chance it will be as good as me one day. Today it's not a threat at all.


Yea, LLMs produce results on par with what I would expect out of a solid junior developer. They take direction, their models act as the “do the research” part, and they output lots of code: code that has to be carefully scrutinized and refined. They are like very ambitious interns who never get tired and want to please, but often just produce crap that has to be totally redone or refactored heavily in order to go into production.

If you think LLMs are “better programmers than you,” well, I have some disappointing news for you that might take you a while to accept.


> LLMs produce results on par with what I would expect out of a solid junior developer

This is a common take but it hasn't been my experience. LLMs produce results that vary from expert all the way to slightly better than markov chains. The average result might be equal to a junior developer, and the worst case doesn't happen that often, but the fact that it happens from time to time makes it completely unreliable for a lot of tasks.

Junior developers are much more consistent. Sure, you will find the occasional developer that would delete the test file rather than fixing the tests, but either they will learn their lesson after seeing your wth face or you can fire them. Can't do that with llms.


I think any further discussion about quality just needs to have the following metadata:

- Language

- Total LOC

- Subject matter expertise required

- Total dependency chain

- Subjective score (audited randomly)

And we can start doing some analysis. Otherwise we're pissing into ten kinds of winds.

My own subjective experience is earth shattering at webapps in html and css (because I'm terrible and slow at it), and annoyingly good but a bit wrong usually in planning and optimization in rust and horribly lost at systems design or debugging a reasonably large rust system.


I agree in that these discussions (this whole hn thread tbh) are seriously lacking in concrete examples to be more than holy wars 3.0.

Besides one point: junior developers can learn from their egregious mistakes, llms can't no matter how strongly worded you are in their system prompt.

In a functional work environment, you will build trust with your coworkers little by little. The pale equivalent in LLMs is improving system prompts and writing more and more ai directives that might or might not be followed.


> Besides one point: junior developers can learn from their egregious mistakes, llms can't no matter how strongly worded you are in their system prompt.

I think if you set off an LLM to do something, and it does a "egregious mistake" in the implementation, and then you adjust the system prompt to explicitly guard against that or go towards a different implementation and you restart from scratch again yet it does the exact same "egregious mistake", then you need to try a different model/tool than the one you've tried that with.

It's common with smaller models, or bigger models that are heavily quanitized that they aren't great at following system/developer prompts, but that really shouldn't happen with the available SOTA models, I haven't had something ignored like that in years by now.


And honestly this is precisely why I don't fear unemployment, but I do fear less employment overall. I can learn and get better and use LLMs as a tool. So there's still a "me" there steering. Eventually this might not be the case. But if automating things has taught me anything, it's that removing the person is usually such a long tail cost that it's cheaper to keep someone in the loop.

But is this like steel production or piloting (few highly trained experts are in the loop) or more like warehouse work (lots of automation removed any skills like driving or inventory work etc).


This seems to be one of the huge weaknesses of current LLMs: Despite the words "intelligence" and "machine learning" we throw around, they aren't really able to learn and improve their skills without someone changing the model. So, they repeat the same mistakes and invent new mistakes by random chance.

If I was tutoring a junior developer, and he accidentally deleted the whole source tree or something egregious, that would be a milestone learning point in his career, and he would never ever do it again. But if the LLM does it accidentally, it will be apologetic, but after the next context window clear, it has the same chances of doing it again.


I can in fact fire an LLM. It's even easier than firing a junior developer.

Or rather, it's more like a contractor. If I don't like the job they did, I don't give them the next job.


you say this as if web development isnt 90% of software.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: