Well, this particular moron reads hacker news, too, and actually has a formal diagnosis for both autism and severe ADHD. But thank you for calling me a moron because someone took the article I published for free to be of help to similar folks. Your day is obviously off to a better start than mine. Le sigh indeed.
Yeah but if you look to the present... there aren't really any jobs where someone is blissfully wandering the earth delegating tasks. Most of the time I can't even take a walk on calls because someone wants to screen share something with me...
I'd like you to be right, but I live in society where joy at work is often considered antithetical to productivity. No matter how much more productive I get, that space is used to fill in more productivity. We'll need more than tooling to stop this.
In aggregate it was good for society, but it was a disaster for a lot of people and a lot of areas. This is the theme of e.g. The Grapes of Wrath.
We should welcome automation and efficiency, but also address the situation of the "losers" of the development and not just expect the invisible hand will sort everything out.
Can you elaborate why having a less diverse farming economy is good for every individual? Automation didn’t invent commodities so it’s unrelated to the advent of a food surplus. It might make obtaining surplus easier but it didn’t give more purpose to people by forcing them to sell their land to a bigger corporation. Even if 50% of farmers don’t want to be farmers anymore doesn’t mean they’d gladly give up their job for a recliner and ubi.
I feel it will get there in short order..but for the time being I feel that we'll be doing some combination of scattershot smaller & maintenance tasks across Codex while continuing to build and do serious refactoring in an IDE...
What works nice also is the text to speech. I find it easier and faster to give more context by talking rather than typing, and the extra content helps the AI to do its job.
And even though the speech recognition fails a lot on some of the technical terms or weirdly named packages, software, etc, it still does a good job overall (if I don’t feel like correcting the wrong stuff).
It’s great and has become somewhat of a party trick at work. Some people don’t even use AI to code that often, and when I show them “hey have you tried this?” And just tell the computer what I want? Most folks are blown away.
Not for me. I first ask Advanced Voice to read me some code and have Siri listen and email it to an API I wrote which uses Claude to estimate the best cloud provider to run that code based on its requirements and then a n8n script deploys it and send me the results via twilio.
That sounds exhausting. Wouldn't it be faster to include you package.json in the context?
I sometimes do this (using Cline), plus create a .cline file at project root which I refine over time and which describes both the high level project overview, details of the stack I'm using, and technical details I want each prompt to follow.
Then each actual prompt can be quite short: read files x, y, and z, and make the following changes... where I keep the changes concise and logically connected - basically what I might do for a single pull request.
My point was that a prompt that simple could be held and executed very well by sonnet, but all other models (especially reasoning models) crash and burn.
It's a 15 line tsx file so context shouldn't be an issue.
Makes me wonder if reasoning models are really proper models for coding in existing codebases
Your last point matches what I’ve seen some people (simonw?) say they’re doing currently: using aider to work with two models—one reasoning model as an architect, and one standard LLM as the actual coder. Surprisingly, the results seem pretty good vs. putting everything on
one model.
This is probably the right way to think about it. O1-pro is an absolute monster when it comes to architecture. It is staggering the breadth and depth that it sees. Ask it to actually implement though, and it trips over its shoelaces almost immediately.
The biggest delta over regular o1 that I've seen is asking it to make a PRD of an app that I define as a stream-of-consciousness with bullet points.
It's fantastic at finding needles in the haystack, so the contradictions are nonexistent. In other words, it seems to identify which objects would interrelate and builds around those nodes, where o1 seems to think more in "columns."
To sum it up, where o1 feels like "5 human minute thinking," o1-pro feels like "1 human hour thinking"
Had the same experience. Cursor ux and also Zed are designed to make copy pasta'ing back and forth obsolete. It's certainly more comfortable to iterate within the IDE, for me.