“If it takes longer to explain to the system all the things you want to do and all the details of what you want to do, then all you have is just programming by another name,”
I think this is going to make the difference between junior and senior engineers even more drastic than it is today. It's really hard to know what/how to even describe real problems to these tools, and the people who invest the most in their tooling now, are going to be most successful. It's hard for someone who hasn't designed a large codebase already to do this in an ai native way where they don't have the experience of abstracting at the right level and things like that.
Today's equivalent, I've often found some of the best engineers I know have insane setups with nvim or emacs. They invest in their tool chain, and are now bringing AI into.
That quote really perfectly encapsulates the challenge with these tools. There is an assumption that inherently code is hard to write and so if you could code in natural language it would save time. But code isn’t actually that hard to write. Sure some people are genuinely bad at it just like I’m genuinely bad at drawing but a bit of practice and most people can be perfectly competent at it.
The hard part is the engineering. Understanding and breaking down the problem, and then actually solving it. If all we gain out of these tools is that we don’t have to write code by hand anymore they are moderately useful but they won’t really be a step change in software development speed.
It's not too different in my opinion from the skills need to build complicated machinery like Boeing 747s despite how much Wallstreet and PHBs want to believe it's fungible. Having competent experienced engineers on the ground level watching these processes and constantly revising and adapting to everything from personnel, material, or vendor changes is so far irreplaceable.
Maybe if we get super AGI one day. Even then I suspect that from a thermodynamics perspective that might not be cost effective as you often need localized on site intelligence.
It's an interesting question but I bet humans combined with AI tooling will remain cost competitive for a long time barring leaps in say quantum compute. After all organic brains operate at the atomic level already and were honed in an extremely competitive environment for billions of years. The calories and resources required to create highly efficient massively powerful neural compute had incredibly thin resource "margins" with huge advantages for species to utilize.
So the author is providing some personal annotations and opinions on a summary of a “new paper” which was actually published five months ago, which itself was a summary of research with the author’s personal annotations and opinions added? These are exactly the kind of jobs that I want AI to automate.
Both humans and coding agents have their strengths and weaknesses, but I've been appreciating help from coding agents, especially with languages or frameworks where I have less expertise, and the agent has more "knowledge", either in its weights or in its ability to more quickly ingest documentation.
One weakness of coding agents is that sometimes all it sees are the codes, and not the outputs. That's why I've been working on agent instructions/tools/MCP servers that empower it with all the same access that I have.
For example, this is a custom chat mode for GitHub Copilot in VS Code:
https://raw.githubusercontent.com/Azure-Samples/azure-search...
I give it access to run code, run tests and see the output, run the local server and see the output, and use the Playwright MCP tools on that local server. That gives the agent almost every ability that I have - the only tool that it lacks is the breakpoint debugger, as that is not yet exposed to Copilot. I'm hoping it will be in the future, as it would be very interesting to see how an agent would step through and inspect variables.
I've had a lot more success when I actively customize the agent's environment, and then I can collaborate more easily with it.
For me it's simple: even the best models are "lazy" and will confidently declare they're finished when they're obviously not, and the immensely increased amount of training effort to get ChatGPT 5's mild improvements on benchmarks suggests that that quality won't go away anytime soon.
Sounds like it's partially about a nuanced trade-off. It can just as well be too eager and add changes I didn't ask for. Being lazy is better than continuing on a bad path.
There's a long distance between "nuanced behavior" and what it actually does now, which is "complete 6 items of an explicit 10-item task list and then ask the user again if they want to continue".
> what makes you think they couldn't produce a larger, smarter, more expensive model?
Because they already did try making a much larger, more expensive model, it was called GPT-4.5. It failed, it wasn't actually that much smarter despite being insanely expensive, and they retired it after a few months.
Calculators wont give you completely wrong results, not even once, where "AI" does that way too often. If calculators did too, mathemeticians simply would not use them.
Nope, that's not the reason. It's because it's just a query that probabilistically creates a garden path of tokens out of a compressed form of the training data, requiring a real coder to evaluate which parts of are useful.
Amazing how someone writing for an IEEE website can't keep their eyes on the fundamentals.
I'm fatigued by these articles that just broadly claim AI can't code because its painting a broad stroke against a widely diverse use of AI for different stacks.
It's horribly outdated way of thinking that an singular AI entity would be able to handle all stacks all problems directed at it because no developer is using it that way.
AI is a great tool for both coders and artists and these outlandish titles that grab attention really seem to be echo chambers aimed at people who are convinced that AI isn't going to replace them which is true but the opposite is also true.
A lot of comments here seem to be similar. I see people claiming that AI has all but taken over doing their work for them, and others claiming that it's almost useless. But usually, nobody even briefly mentions what the work is (other than, presumably, something related to programming).
I imagine there's a big difference in using AI for building, say, an online forum vs. building a flight control system, both in terms of what the AI theoretically can do, and in terms of what we maybe should or should not be letting the AI do.
Yeah. I use it for analytics/dataviz stuff (which involves a lot of python to run spark jobs, glue different APis to get some extra column of data, making png or svg pictures,, making D3 based web sites in html and JavaScript). That all works pretty well.
I also write high performance Go server code, where it works a lot less well. It doesn't follow my rules for ptr APIs or using sync mutexes or atomic operations across a code base. It (probably slightly older version than SOTA) didn't read deep call chains accurately for refactoring. It's still worth trying but if that was my sole work it would probably not be worth it.
On the other hand for personal productivity, emacs functions and config, getting a good .mcp.json, it is also very good and generates code that partakes in the exponential growth of good code. (Unlike data viz where there is a tendency to build something and then the utility declines over time).
I can confidently state that for CRUD web apps, its truly over as in those jobs are never going to command the same wages it once used to.
With the recent models its now encroaching similarly on all fronts, I think the next few iterations we'll see LLM solidify itself as a meta-compiler that will be deployed locally for more FCS type systems.
At the end of the day the hazards are still same with or without AI, you need checks and bounds, you need proper vetting of code and quality but overall it probably doesn't make sense to charge an hourly rate because an AI would drastically cut down such billing schemes.
For me "replacement" is largely a 70~80% reduction in either hourly wages, job positions or both and from the job market data I see it can get there.
"crud web apps" sounds like WordPress or Django anyways
I mean, it's already kind of valueless ? The true value lies in what this crud app is about, marketing, and the extra bits you can add to make it special.
Well, AI really can't code any more than a compiler can. They all require a human to write the original code, even the machine does translate it into other code.
And until the day that humans are no longer driving the bus that will remain the case.
You can say generate a c program that uses gcc 128 bit floats and systematically generates all quadratic roots in order of the max size of their minimal polynomial coefficients, and then sort them and calculate the distribution of the intervals between adjacent numbers, and it just does it. That's qualitatively different from the compilers I have used. Now I was careful to use properly technical words to pull in the world of numeric computation and c programming. But still saved me a lot of time. It was even able to bolt on multithreaded parallelism to speed it up using c stuff I never heard of.
> That's qualitatively different from the compilers I have used.
Is it? I can, in most traditional programming languages commonly used to today, using decades old compiler technology, say something like "x = [1,2,3]" and it will, for example, systematically generate all the code necessary to allocate memory without any need for me to be any more explicit about it. It would be fair to say AI offers an even higher level abstraction, like how most programming languages used today are a higher level abstraction over assembly, but fundamentally different it is not.
"generate a c program that uses gcc 128 bit floats and systematically generates all quadratic roots in order of the max size of their minimal polynomial coefficients, and then sort them and calculate the distribution of the intervals between adjacent numbers" is just code. You still have to write the code get AI to translate it into a lower-level abstraction. It doesn't magically go off and do its own autonomous thing.
There is so much more than coding, it’s the business rules, the constraints held in other parts of the company, the undocumented code that holds a company with no written context … you start a project from scratch ? AI can help you but you better be highly specific because hallucinations and errors will be there along the way.
Now for the hundreds or billion of loc of private source code softwares that holds thousands of companies with not tightly formatted context and most of it just in some people’s memory … how would AI replace that ?
Try let an AI rewrite a full ERP responsible of business logic with costly consequences on business and money …
It will be just a tool, where sometimes we loose time and most time save a bit of time if we use it well
I'm tired of the anthropomorphization marketing behind AI driving this kind of discussion. In a few years, all this talk will sound as dumb as stating "MS Word spell checker will replace writers" or "Photoshop will replace designers".
We'll reap the productivity benefits from this new tool, create more work for ourselves, output will stabilize at a new level and salaries will stagnate again, as it always happens.
Generative AI is replacing writers, designers, actors, ... it is nothing like just a spell checker or Phtoshop.
Everyday, I see ads on YouTube with smooth-talking, real-looking AI-generated actors. Each one represents one less person that would have been paid.
There is no exact measure of correctness in design; one bad bit does not stop the show. The clients don't even want real art. Artists sometimes refer to commercial work as "selling out", referring to hanging their artistic integrity on the hook to make a living. Now "selling out" competes with AI which has no artistic integrity to hang on the hook, works 24 hours a day for peanuts and is astonishingly prolific.
> Everyday, I see ads on YouTube with smooth-talking, real-looking AI-generated actors. Each one represents one less person that would have been paid.
Were AI-generated actors chosen over real actors, or was the alternative using some other low-cost method for an advertisement like just colorful words moving around on a screen? Or the ad not being made at all?
The existence of ads using generative AI "actors" doesn't prove that an actor wasn't paid. This is the same logical fallacy as claiming that one pirated copy of software represents a lost sale.
Yes, this. I recently used Midjourney to generate a super-convincing drone shot video for a presentation. The alternative would have been a stock photo.
Probably took me the same amount of time to generate a pleasing video as I would have spent browsing Shutterstock. Only difference is my money goes to one corporation instead of the other.
As far as the video is concerned, it adds a bit of a wow factor to get people interested, but ultimately it's the same old graphs and bullet points with words and numbers that matter. And those could just as well have been done on an overhead transparency in 1987.
A) J. Crew and others are using AI models instead of real models. Retail modeling was steady pay-the-bills work for models and actors and is actively being replaced by AI image generators— sometimes trained on images of a particular model they’re no longer paying. Writers and stock photographers are in much worse shape.
B) Even in cases where AI actors are used where there wouldn’t have been actors before, the skillset is still devalued, and even that modest insulation for higher-end work is almost certainly temporary. Someone doing a worse version of what you do for 1% of your pay affects the market, and saving 99% is great incentive for companies to change their strategy until the worse version is good enough.
It means that being a live actor is less of a differentiator. Of course great movie stars will remain, playing live, or animating computer characters, no matter. But simpler works like ads featuring a human now become more accessible.
Among other things, this will remove most entry-level jobs, making senior-level actors more rare and expensive.
I think this means that personal branding is going to get even more important than it already is (for example, people watching movies specifically because of Ryan Reynolds, or avoiding them because of Jared Leto)
It's really both effects happening at once. AI is just like the invention of the assembly line, or the explosion of mass produced consumer packaged goods starting from the first cotton gin. Automation allows a massive increase in quantity of goods, and even when quantity comes with tradeoffs to quality vs artisanally produced goods, they still come to dominate. Processed cheese or instant coffee is pretty objectively worse that the 'real' thing, but that didn't stop cheap mass production still made those products compelling for many million/billion of consumers.
You can still find a tailor who will hand make you a bespoke clothing or sew your own clothes yourself (as even the boomer generation often did growing up), but tailored clothing is a tiny fraction of the amount of clothing in circulation. Do tailors and artisanal cheese makers still exist? Yep, they are not extinct. But they are hugely marginalized compared to machine-made alternatives.
the crazy thing is, I can get locally-roasted beans that are single-origin microlots from all over the world, in part because of the coffee boom that was a result of instant coffee and the desire for better.
I agree with your sentiment. But where I struggle is: to what degree do each of those ads “represent one less person who would have been paid” versus those that represent one additional person who would not be able to afford to advertise in that medium.
Of course that line of reasoning reduces similar to other automation / minimum wage / etc discussions
YouTube has the lowest quality ads of any online platform I use by several orders of magnitude. AI being used for belly fat and erectile dysfunction ads is not exactly good for its creative reputation
Local governments in BR have already made ads using generative AI that were shown during prime time TV hours[1].
You can argue that is a bad thing (local designers/content producers/actors/etc lost revenue, while the money was sent to $BigTech) or that this was a good thing (lower cost to make ad means taxpayer money saved, paying $BigTech has lower chance of corruption vs hiring local marketing firm - which is very common here).
I have no doubt there will be AI advertising. I bet it’s the primary untapped revenue stream. My argument is that it will be associated with cheap, untrustworthy products over time, even if it’s possible to spend more money and get better AI ads. Same thing as social/search ads.
There's a difference between taking one thing and putting something else in it's spot, and truly REPLACING something. Yes, some ads have AI generated actors. You know because you can tell because they're "not quite right", rather than focusing on the message of the ad. Noticing AI in ads turns more people off than on, so AI ads are treated by a lot of people as an easy "avoid this company" signal. So those AI ads are in lieu of real actors, but not actually replacing them because people don't want to watch AI actors in an ad. The ad ceases to be effective. The "replacement" failed.
Animated films have competed for box office dollars since basically the dawn of cinema. Animated characters have fan followings.
Just wait; the stuff is coming. Ultra-realistic full-length feature films with compelling AI characters that are not consistent from beginning to end, but appear in multiple features.
“ Everyday, I see ads on YouTube with smooth-talking, real-looking AI-generated actors. Each one represents one less person that would have been paid.”
The thing is that they would not have paid for the actor anyway. It’s that having an “actor” and special effects for your ads cost nothing, so why not?
The quality of their ads went up, the money changing hands did not change.
Anecdata: I know writers, editors, and white collar non-tech workers of all kinds who use AI daily and like it.
When GPT3.5 first landed a lifelong writer/editor saw a steep decrease in jobs. A year later the jobs changed to "can you edit this AI generated text to sound human", and now they continue to work doing normal editing for human or human-ish writing while declining the slop-correction deluge because it is terrible work.
I can't help but see the software analogy for this.
I'm not a "real coder" either, but it sounds like the "No True Scotsman" trap when people say, “AI can’t be a real coder,” and then redefine “real coder” to mean something AI can’t currently do (like full autonomy or deep architectural reasoning). This makes the claim unfalsifiable and ignores the fact that AI already performs several coding tasks effectively. Yeah, I get it, context handling, long-horizon planning, and intent inference all stink, but the tools are all 'real' to me.
That's based on the assumption models would not soon cross that treshold of autonomy and self-reflection that suddenly makes an escalating number of jobs (with cheap humanoids, even physical) automatable for ridiculous pricing.
Even if this isn't certain, likelihood could be considered quite high and thus we urgently need a public debate / design process for the peaceful, post-commercial, post-competitive, open-access post-scarcity economy some (RBE / commoning community) have been sketching for years and years. Seems this development defies most people's sense of imagination - and that's precisely why we need to raise public awareness for the freedom and fun OPEN SOURCE EVERYTHING & Universal Basic Services could bring to our tormented world. 2 billion without access to clean water? we can do much better if we break free from our collective fixation on money as the only means and way to deal with things ever.
People are stupid, always have been - took thousands of years to accept brain as the seat of thought because “heart beat faster when excited, means heart is source of excitement”.
Heck, people literally used to think eyes are the source of light since everything is dark when you close them.
People are immensely, incredibly, unimaginably stupid. It has taken a lot of miracles put together to get us where we are now…but the fundamentals of what we are haven’t changed.
> took thousands of years to accept brain as the seat of thought because “heart beat faster when excited, means heart is source of excitement”
So what you are saying is that beings without a central nervous system cannot experience "excitement"?
or perhaps the meaning of too many words has changed, and their context. When Hippocrates claimed that the brain was an organ to cool the blood, perhaps he meant that we use our thought to temper our emotions, i.e. what he said agrees with our modern understanding.
However, many people read Hippocrates and laugh at him, because they think he meant the brain was some kind of radiator.
Maybe because we stopped talking about "excitable" people as being "hot-blooded"
>or perhaps the meaning of too many words has changed, and their context. When Hippocrates claimed that the brain was an organ to cool the blood, perhaps he meant that we use our thought to temper our emotions, i.e. what he said agrees with our modern understanding.
The belief that the heart was the seat of thought and emotion was shared by numerous cultures[0], and was based on their naive interpretation of physiology and biology and cannot be dismissed as a modern misinterpretation of a single vague aphorism by a single person due to the preponderance of documentary evidence to the contrary from contemporary sources. Also, you're probably talking about Aristotle, not Hippocrates.
>Maybe because we stopped talking about "excitable" people as being "hot-blooded"
You're confusing ignorance with stupidity. People at the time were coming to the best conclusions they could based on the evidence they had. That isn't stupid. If humans were truly "incredibly, unimaginably stupid" we wouldn't have even gotten to the point of creating agriculture, much less splitting the atom. We didn't get here through "miracles," we got here through hard work and intelligence.
Stupid is people in 2025 believing the world is flat and germ theory is a hoax. Ignorance becomes stupidity when our species stands on the shoulders of giants but some people simply refuse to open their eyes.
Proofreaders still exist, despite spell checker. Art assistants still exist, despite Photoshop. There's always more work to do, you just incorporate the new tools and bump the productivity, until it gets so commoditized it stops being a competitive advantage.
Saying AI "replaces" anyone is just a matter of rhetoric to justify lower salaries, as always.
> all this talk will sound as dumb as stating "MS Word spell checker will replace writers" or "Photoshop will replace designers".
You cannot use just a spell checker to write a book (no matter how bad) or photoshop (non-AI) plugins to automatically create meaningful artwork, replacing human intervention.
Business people "up the ladder" are already threatening with reducing the workforce and firing people because they can (allegedly) be replaced by AI. No writer was ever threatened by a spellchecker.
Hollywood studio execs are putting pressure on writers, and now they can leverage AI as yet another tool against them.
I'm tired of all the "yet another tool" reductionism. It reeks of cope.
It took under a decade to get AI to this stage - where it can build small scripts and tiny services entirely on its own. I see no fundamental limitations that would prevent further improvements. I see no reason why it would stop at human level of performance either.
There’s this saying that humans are terrible at predicting exponential growth. I believe we need another saying, those who expect exponential growth have a tough time not expecting it.
It’s not under a decade for ai to get to this stage but multiple decades of work, with algorithms finally able to take advantage of gpu hardware to massively excel.
There’s already feeling that growth has slowed, I’m not seeing the rise in performance at coding tasks that I saw over the past few years. I see no fundamental improvements that would suggest exponential growth or human level of performance.
I'm not sure if there will be exponential growth, but I also don't believe that it's entirely necessary. Some automation-relevant performance metrics, like "task-completion time horizon", appear to increase exponentially - but do they have to?
All you really need is for performance to keep increasing steadily at a good rate.
If the exponential growth tops out, and AI only gains a linear two days per year of "task-completion time horizon" once it does? It'll be able to complete a small scrum sprint autonomously by year 2035. Edging more and more into the "seasoned professional developer" territory with each passing year, little by little.
ok, ok! just like you can find for much less computation power involved using a search engine, forums/websites having if not your question, something similar or a snippet [0] helping you solve your doubt... all of that free of tokens and companies profiting over what the internet have build! even FOSS generative AI can give billions USD to GPU manufacturers
there are more of all these bottlenecks among the proprietary or open source project worlds, which have yet to collaborate amongst themselves to unify the patterns in their disparate codebases and algorithms into a monolith designed to compress representations of repeated structures edited for free by a growing userbase of millions and the maturing market of programmers who grew up with cheap GPUs and reliable optimization libraries
the article's subtitle is currently false, people collaborate more with the works of others through these systems and it would be extremely difficult to incentivize any equally signifciant number of the enterprise software shops, numerics labs, etc to share code: even joint ventures like Accenture do not scrape all their own private repos and report their patterns back to Microsoft every time they re-implement the same .NET systems over and over
How about the fact that AI is only trained to complete text and literally has no "mind" within which to conceive or reason about concepts? Fundamentally, it is only trained to sound like a human.
The simplest system that acts entirely like a human is a human.
An LLM base model isn't trained for abstract thinking, but it still ends up developing abstract thinking internally - because that's the easiest way for it to mimic the breadth and depth of the training data. All LLMs operate in abstracts, using the same manner of informal reasoning as humans do. Even the mistakes they make are amusingly humanlike.
There's no part of an LLM that's called a "mind", but it has a "forward pass", which is quite similar in function. An LLM reasons in small slices - elevating its input text to a highly abstract representation, and then reducing it back down to a token prediction logit, one token at a time.
They merely act sort of like a human. Which is entirely expected - given that the datasets they're trained on only capture some facets of human behavior.
Don't expect them to show mastery of spatial reasoning or agentic behavior or physical dexterity out of the box.
They still capture enough humanlike behavior to yield the most general AI systems ever built.
It doesn’t develop any thinking, it’s just predicting tokens based on a statistical model.
This has been demonstrated so many times.
They don’t make mistakes. It doesn’t make any sense to claim they do because their goal is simply to produce a statistically likely output. Whether or not that output is correct outside of their universe is not relevant.
What you’re doing is anthropomorphizing them and then trying to explain your observations in that context. The problem is that doesn’t make any sense.
When you reach into a "statistical model" and find that it has generalized abstracts like "deceptive behavior", or "code error"? Abstracts that you can intentionally activate or deactivate - making an AI act as if 3+5 would return a code error, or as if dividing by zero wouldn't? That's abstract thinking.
Those are real examples of the kind of thing that can be found in modern production grade AIs. Not "anthropomorphizing" means not understanding how modern AI operates at all.
I suspect once you have studied how we actually got to where we are today, you might see why your lack of seeing any limitations may not be the flex you think it is.
We see massive initial growth followed by a slowdown constantly.
There is zero reason to think AI is some exception that will continue to exponentially improve without limit. We already seem to be at the point of diminishing returns. Sinking absurd amounts of money and resources to train models that show incremental improvements.
To get this far they have had to spend hundreds of billions and have used up the majority of the data they have access to. We are at the point of trying to train AI on generated data and hoping that it doesn’t just cause the entire thing degrade.
> I see no fundamental limitations that would prevent further improvements
How can you say this when progress has so clearly stagnated already? The past year has been nothing but marginal improvements at best, culminating in GPT-5 which can barely be considered an upgrade over 4o in terms of pure intelligence despite the significant connotation attached to the number.
Marginal improvements? Were you living under a rock for the past year?
Even o1 was a major, groundbreaking upgrade over 4o. RLVR with CoT reasoning opened up an entire new dimension of performance scaling. And o1 is, in turn, already obsoleted - first by o3, and then by GPT-5.
I’ve seen comments here claiming that this site is either a bunch of coders coping about the limitations of AI and how it can’t take their job, or a bunch of startup dummies totally on the AI hype train.
Now, there’s a little room between the two—maybe the site is full of coders on a cope train, hoping that we’ll be empowered by nice little tools rather than totally replaced. And, ya know, multiple posters with multiple opinions, some contradictions are expected.
But I do find it pretty funny to see the multiple posters here describe the site they are using as suffering from multiple, contradictory, glaringly obvious blindspots.
I think everyone is looking for back and white switches. Either coding agents are writing your code or they aren't. Humans will always be in the mix in some form, but the amount and skills they use is going to be radically different as time goes on.
I personally haven't written any significant code by hand since claude code landed. I also have a high tolerance for prompting and re-prompting. Some of my colleagues would get upset if it wasn't mostly one shotting issues and had a really low tolerance for it going off the rails.
Since gpt-5-high came out, I rarely have to re-prompt. Strong CI pipeline and well defined AGENTS.md goes an incredibly long way.
I can't help thinking that a big factor is how bad our tools were before. There are many leaking abstractions, boilerplate code and cruft. Dealing with that fills a sizeable percent of a programmer's job.
These tools seem great because they are less sensitive than humans to the mess and lift us over the tedious work. But at the same time, they're giving us an excuse to not fix what needed to be fixed and, in doing so, they're adding more crap to the heap.
Maybe what we need is forcing the tools to build on a simpler base, so we can keep an understanding of the results.
True, but it's catching up fast. A year ago, I used AI for small OS scripts. It worked fine and saved me the time of looking up switches for commands. Now, I can ask it to create a simple game of about 200 lines, and it does a pretty good job of writing bug-free code within a few seconds. It's only going to get better. Even if the tech doesn't improve further, I can see a future where all apps are endlessly configurable.
A big part of my career has been the modification of enterprise software to fit a company's needs. Rarely was any one addition more than a few hundred lines of code. I can see a future where there will be simple options for a non-coder to add to an app.
True, it's not a coder, but that doesn't mean it won't fundamentally change how apps are made and it will reduce the number of master programmers needed. It won't replace all programmers, but it will greatly reduce the number that are needed, which country they work in and the language they use to program apps.
Programming has mainly been a career that requires the individual to understand English. That is changing. I can see a future where code can be created in multiple human languages. Programming was well-paid because relatively few people had the expertise to do it. That won't be the case, and the pay will adjust downward as needed. AI might not be a coder, but it will let many more people become coders. In the future, coding will be in the same pay range as clerical work. Companies will be hiring Programming Clerks rather than Programming Engineers.
I think you're right that LLMs are democratizing access to coding, but unless and until AI models reach a point where they can say 'no' to their users, the scenario you're imagining ('endlessly configurable apps') will probably lead to software that collapses under its own complexity.
Years ago, I supported a team of finance professionals who were largely quite competent at coding but knew nothing about software engineering. They had thousands of scripts and spreadsheets: they used version control, but kept separate long-lived branches for client-specific variations of different models. There were no tests for anything; half the tools would break when the clocks changed.
They weren't dumb, but their incentives weren't about building anything we might recognize as an engineered application. I suspect something similar will happen turning end users loose with AI.
There's been a strong theme recently here on HN of confusing programming (the act of writing code to meet specifications) and Engineering(the writing of specifications, and the oversight of said process, along with overview of testing).
AI is definitely not ready for an Engineering role. My recent experience with ChatGPT5(Preview) via Visual Studio Code tells me that it might perform acceptably as a junior programmer. However, because I'm an old retired self taught programmer who only ever managed 1 other programmer, I lack the experience to know what's acceptable as a Junior Programmer at FAANG and elsewhere.
> There's been a strong theme recently here on HN of confusing programming (the act of writing code to meet specifications) and Engineering(the writing of specifications, and the oversight of said process, along with overview of testing).
You're making a distinction that might be interesting in some esoteric sense, but that doesn't exist in the industry. When it comes to software, the architects are also the construction crew. There's nothing to confuse if the terms are effectively synonymous.
I think this is generally true, and there are SO MANY blog posts and articles. Using something like Claude Code to build an entire SaaS from nothing can seem like magic, but there is still a point where the code is too big and any LLM will lose context.
But there is a "sweet spot" where it's amazing, specifically highly targeted tasks with a specific context. I wanted a simple media converter app that tied into ffmpeg, and I didn't want to install any of the spammy or bloated options I found... so I got Claude to build one. It took about 30 minutes and works great.
I also asked it to update some legacy project, and it fell down a testing a loop where it failed to understand the testing database was missing. Had I not had years of knowledge, I would've looked at the output and suggestions Claude was giving and spent hours on it... but it was a simple command that fixed it. As with all new tech, your milage will vary.
Anecdotes are unreliable, for one, your described use case and the tools you are using suggest you are at a very basic level and unable to extract the full capabilities of the tooling and models which many of use to solve if not complete complex software.
Just be aware that AI is a tool not a replacement but a human apt at AI as a tool will replace the former.
I didn't want to get into the details, because I've already talked about BitGrid here endlessly, and was trying to stay on the topic of AI usefulness, but since you asked.
I'm trying to build a software stack that can eventually take something like a PyTorch model, and unwind everything, resulting in a directed acyclic graph of individual bit-level operations (OR, AND, XOR). That graph will then be compiled into a bitstream suitable for an FPGA-like substrate that eliminates the memory/compute divide, the BitGrid[1].
FPGA routing is a non-trivial problem, I'm hoping to get it down to seconds. I'm currently trying to build the software stack to make it usable.
The goal is to answer questions about BitGrid:
How efficiently can I pack a program into the hardware?
Is the model I've chosen for a cell optimal?
How many femtojoules per operation would a cell actually take?
If the answers are favorable, then in the deep (and improbable) future, it's possible that there could be a set of racks with an array of thousands resulting in a system that could stream ChatGPT at aggregate rate of a gigatoken per second, for far less than the Trillion dollars Meta plans to spend.
This isn't just some CRUD application with a web front end. There are a number of layers of abstraction at play, and the LLMs seem to handle it well if you limit the depth under consideration.
[1] BitGrid eliminates the traditional memory/compute divide that causes most of the energy consumption of CPUs, GPUs, and other accelerators. Even FPGA systems tend to focus on emulation of these models, and routing fabric for minimum latency, instead of maximum performance. Because all the active lines only reach nearest neighbors, power consumption for a given operation can be far lower than the traditional approach.
PS: I pay $10/month for GitHub CoPilot, which apparently now includes ChatGPT5
“If it takes longer to explain to the system all the things you want to do and all the details of what you want to do, then all you have is just programming by another name”
If it's taking you that long to direct the AI, then either you're throwing too small a problem at it, or too big a problem at it, or you're not directing its attention properly.
In a way, your prompts should feel like writing user documentation:
Refactor all of the decode functions in decoder.rs to return the number of bytes decoded
in addition to the decoded values they already return. While refactoring, follow these principles:
* Use established best practices and choose the most idiomatic approaches.
* Avoid using clone or unwrap.
* Update the unit tests to account for the extra return values, and make sure the tests check the additional return values.
When you're finished, run clippy and fix any issues it finds. Then run rustfmt on the crate.
And then for some bizarre reason sometimes it doesn't really work, AI added a bunch of random shit, then you can feel your rage bubbling up in real-time as you have to re-prompt it.
I've yet to have that happen. But then again so far I've only used it for Rust, and it's hard for the AI to maintain hallucinations with such a strict compiler.
“Feels primitive,” is holding up a lot of assumptions.
What feels primitive to me is how we approach programming in industry as a process of trial and error rather than one of rigour.
These are tools that automate copy-pasting from Stack Overflow and GitHub, running tools, and generating a ton of noise to sift through. They hallucinate code, documentation, and various other artifacts that are sometimes useful and are occasionally complete BS.
Some people find that they can make useful tools out of these things. Great.
A real programmer is still a human.
Update: nothing wrong with trial and error as a process. I use it a lot. But there are lots of places where we use this method that seem inappropriate and sometimes even dangerous. Yet it’s the most common tool we have and everything starts to look like a nail.
But what is meant by "real coder"? Does it mean an experienced senior coder? In that case, no, AI isn't ready to replace one. How about a junior coder who is fresh out of university and starting his first job? Surely that counts as a "real coder"? I'd argue that AI can and does already replace such coders, depending on the type of work they have been assigned.
Could you elaborate your argument? I'd say AI is far from replacing any human contributor at this point.
Context doesn't work the same way as memory + experience in humans. While humans have an impression and a flexible mental model of any single domain, AI needs hard data, which is hard to manage with context and can't really be worked around by fine-tuning in practice, lest you have to retrain the model on each and every code merge.
Sure. My point is that it's not unusual for the quality of work of a junior coder to be fairly low, and that work typically needs to be double checked by someone more senior before it can be used in production. Often, an LLM will produce a higher quality output than that junior developer in which case, given that the work has to be checked by a senior coder no matter what, the LLM can replace the junior developer.
Also, not every project is highly complex or involves huge codebases and, for those that are, a junior developer might be assigned to a reasonably self contained and small module within the project (i.e. something that a LLM can do better at).
Finally, although this wasn't part of my original argument: before modern LLMs, let's say that a senior developer could produce N units of work per Y units of time. Now, with the help of a LLM, they are able to produce XN units of work per Y unit of time. I don't think it's too controversial to claim that some portion of developers are now more productive than they were before. And if developers in general are more productive, then you can expect some reduction in the number of developers that you need for a given project. So, in an indirect sense, LLMs are already capable of replacing coders.
The pattern I see time & time again in software development: a small set of few and usually very talented senior developers build a software project and they build examples of how everything is done. Junior developers then extend those examples, eg: "add another column to this table." Juniors are just extending the existing patterns, seniors are the ones that create them.
This pattern I see play out time and time again. Senior builds the integration to various team services, then juniors tweak & mimic those to do more.
> let's say that a senior developer could produce N units of work per Y units of time.
So, it's more the case that senior developers are building full production assembly lines, that can then be replicated to do similar things relatively easily. The cost of the units of work are not at all equal.
How does this apply to AI? Good luck getting AI to do something novel, by definition it can't. OTOH, extending an existing pattern is trivial for a senior developer to do after they have built that pattern; juniors spend a lot of time at it cause they have to learn everything.
So, who is the AI going to replace? If the junior, then you have a senior developer using an AI to do something that is already trivial for them to do. No way AI is replacing the senior.. So who? Who is getting replaced by the AI and who is using it?
My #1 issue with AI for programming is that while I do trust it to produce OK boilerplate code, I can also just download that boilerplate code, if I haven't already written it myself. Free software is the answer.
This is profound, and I think you have dove to the core of why LLMs are taking off for code.
They fix a problem: the mutual incompatibility between software. "Free software" is not the answer, as it's used today: for example, the canonical Free Software OS, Linux, is fragmented into many mutually incompatible ABIs.
We need to tear down barriers between software. Barriers between languages, frameworks and OS ABIs. I don't have a good plan , but think it is the true solution. LLMs paper over these incompatibilities, and in doing so, provide value. Imperfect, but good enough until we destroy the barriers properly.
Example: Why am I using an LLM for my rust code? Because it is taking code that solves my problem, but in Python, C, or papers, and translating in a way compatible with the rust dialect I use. The problem is solved multiple times, but not in a way I can use without a translation layer. LLMs provide that.
"According to the study, AI still struggles with several crucial facets of coding: sweeping scopes involving huge codebases, the extended context lengths of millions of lines of code, higher levels of logical complexity, and long-horizon or long-term planning about the structure and design of code to maintain code quality."
uhhh, not sure even the best people or teams are very good at this either. Condemning AI for not being capable of something we're not capable of, ok...
“If it takes longer to explain to the system all the things you want to do and all the details of what you want to do, then all you have is just programming by another name.”
This is called the specification process, which hopefully is already occurring today.
There's so much self-serving bias in articles like this, as well as the comments on HN, Reddit, etc. It's good to critique AI, but that self-serving line is frequently crossed by many people.
> These kinds of complex tasks can be difficult for AI development tools to navigate, resulting in hallucinations about where the bug is or its root cause, as well as irrelevant suggestions or code fixes with subtle problems.
How is this any different than the way I program?
I've contributed genuinely useful features to FLOSS projects "as well as irrelevant suggestions or code fixes with subtle problems", mostly the latter as there was always a few stages of improvement and/or finishing by the core devs of the program I used to haunt. Honestly, I was less than half as useful as the current crop of robots and they still tolerated (in fact, encouraged) my involvement.
> Honestly, I was less than half as useful as the current crop of robots and they still tolerated (in fact, encouraged) my involvement.
Don’t sell yourself short. External contributors are extremely valuable as they are often users and provide real world validation of a need for whatever they are contributing. They also retain any knowledge for any feedback they receive that they can apply to future contributions. And they also become advocates for that software, helping it grow its user base.
AI can't replace humans, but it can be an exoskeleton for human pilots.
AI code tab-complete is fantastic. It's at least an order of magnitude more powerful than IDE-assisted auto refactors.
AI graphics design tools are probably the single best thing in the field. Editing photos, creating new graphics, making marketing materials, shooting and editing videos is now extremely easy. It's a 100x speed up and a 1000x cost reduction. You still have to re-roll the generations repeatedly, but with a competent editing tool you can speed run any design work. This is one area where non-experts can also use the tools.
Can AI categorically not be a "real coder" or is the limitation in more trivial things such as scope and breadth? My experiences suggest to me that it is technically quite close.
“If it takes longer to explain to the system all the things you want to do and all the details of what you want to do, then all you have is just programming by another name,”
I think this is going to make the difference between junior and senior engineers even more drastic than it is today. It's really hard to know what/how to even describe real problems to these tools, and the people who invest the most in their tooling now, are going to be most successful. It's hard for someone who hasn't designed a large codebase already to do this in an ai native way where they don't have the experience of abstracting at the right level and things like that.
Today's equivalent, I've often found some of the best engineers I know have insane setups with nvim or emacs. They invest in their tool chain, and are now bringing AI into.
That quote really perfectly encapsulates the challenge with these tools. There is an assumption that inherently code is hard to write and so if you could code in natural language it would save time. But code isn’t actually that hard to write. Sure some people are genuinely bad at it just like I’m genuinely bad at drawing but a bit of practice and most people can be perfectly competent at it.
The hard part is the engineering. Understanding and breaking down the problem, and then actually solving it. If all we gain out of these tools is that we don’t have to write code by hand anymore they are moderately useful but they won’t really be a step change in software development speed.
It's not too different in my opinion from the skills need to build complicated machinery like Boeing 747s despite how much Wallstreet and PHBs want to believe it's fungible. Having competent experienced engineers on the ground level watching these processes and constantly revising and adapting to everything from personnel, material, or vendor changes is so far irreplaceable.
Maybe if we get super AGI one day. Even then I suspect that from a thermodynamics perspective that might not be cost effective as you often need localized on site intelligence.
It's an interesting question but I bet humans combined with AI tooling will remain cost competitive for a long time barring leaps in say quantum compute. After all organic brains operate at the atomic level already and were honed in an extremely competitive environment for billions of years. The calories and resources required to create highly efficient massively powerful neural compute had incredibly thin resource "margins" with huge advantages for species to utilize.
So the author is providing some personal annotations and opinions on a summary of a “new paper” which was actually published five months ago, which itself was a summary of research with the author’s personal annotations and opinions added? These are exactly the kind of jobs that I want AI to automate.
Both humans and coding agents have their strengths and weaknesses, but I've been appreciating help from coding agents, especially with languages or frameworks where I have less expertise, and the agent has more "knowledge", either in its weights or in its ability to more quickly ingest documentation.
One weakness of coding agents is that sometimes all it sees are the codes, and not the outputs. That's why I've been working on agent instructions/tools/MCP servers that empower it with all the same access that I have. For example, this is a custom chat mode for GitHub Copilot in VS Code: https://raw.githubusercontent.com/Azure-Samples/azure-search...
I give it access to run code, run tests and see the output, run the local server and see the output, and use the Playwright MCP tools on that local server. That gives the agent almost every ability that I have - the only tool that it lacks is the breakpoint debugger, as that is not yet exposed to Copilot. I'm hoping it will be in the future, as it would be very interesting to see how an agent would step through and inspect variables.
I've had a lot more success when I actively customize the agent's environment, and then I can collaborate more easily with it.
For me it's simple: even the best models are "lazy" and will confidently declare they're finished when they're obviously not, and the immensely increased amount of training effort to get ChatGPT 5's mild improvements on benchmarks suggests that that quality won't go away anytime soon.
Sounds like it's partially about a nuanced trade-off. It can just as well be too eager and add changes I didn't ask for. Being lazy is better than continuing on a bad path.
There's a long distance between "nuanced behavior" and what it actually does now, which is "complete 6 items of an explicit 10-item task list and then ask the user again if they want to continue".
gpt-5 is extremely cheap, what makes you think they couldn't produce a larger, smarter, more expensive model?
gpt-5 was created to be able to service 200m daily active users.
> what makes you think they couldn't produce a larger, smarter, more expensive model?
Because they already did try making a much larger, more expensive model, it was called GPT-4.5. It failed, it wasn't actually that much smarter despite being insanely expensive, and they retired it after a few months.
that was not a reasoning model.
None of them are reasoning models. Some of them have a loop of word-outputting.
Using a calculator won't make you a mathematician, but a mathematicians with a calculator can show you amazing things.
Calculators wont give you completely wrong results, not even once, where "AI" does that way too often. If calculators did too, mathemeticians simply would not use them.
Most specifically, random wrong results. Some calculators have issues with rounding, but if you understand those issues, it's consistent.
Imagine driving your car, you turn right, but today turning right slams on the brakes, and 10 people rear end you! That's current AI.
Nope, that's not the reason. It's because it's just a query that probabilistically creates a garden path of tokens out of a compressed form of the training data, requiring a real coder to evaluate which parts of are useful.
Amazing how someone writing for an IEEE website can't keep their eyes on the fundamentals.
I'm fatigued by these articles that just broadly claim AI can't code because its painting a broad stroke against a widely diverse use of AI for different stacks.
It's horribly outdated way of thinking that an singular AI entity would be able to handle all stacks all problems directed at it because no developer is using it that way.
AI is a great tool for both coders and artists and these outlandish titles that grab attention really seem to be echo chambers aimed at people who are convinced that AI isn't going to replace them which is true but the opposite is also true.
A lot of comments here seem to be similar. I see people claiming that AI has all but taken over doing their work for them, and others claiming that it's almost useless. But usually, nobody even briefly mentions what the work is (other than, presumably, something related to programming).
I imagine there's a big difference in using AI for building, say, an online forum vs. building a flight control system, both in terms of what the AI theoretically can do, and in terms of what we maybe should or should not be letting the AI do.
Yeah. I use it for analytics/dataviz stuff (which involves a lot of python to run spark jobs, glue different APis to get some extra column of data, making png or svg pictures,, making D3 based web sites in html and JavaScript). That all works pretty well.
I also write high performance Go server code, where it works a lot less well. It doesn't follow my rules for ptr APIs or using sync mutexes or atomic operations across a code base. It (probably slightly older version than SOTA) didn't read deep call chains accurately for refactoring. It's still worth trying but if that was my sole work it would probably not be worth it.
On the other hand for personal productivity, emacs functions and config, getting a good .mcp.json, it is also very good and generates code that partakes in the exponential growth of good code. (Unlike data viz where there is a tendency to build something and then the utility declines over time).
I can confidently state that for CRUD web apps, its truly over as in those jobs are never going to command the same wages it once used to.
With the recent models its now encroaching similarly on all fronts, I think the next few iterations we'll see LLM solidify itself as a meta-compiler that will be deployed locally for more FCS type systems.
At the end of the day the hazards are still same with or without AI, you need checks and bounds, you need proper vetting of code and quality but overall it probably doesn't make sense to charge an hourly rate because an AI would drastically cut down such billing schemes.
For me "replacement" is largely a 70~80% reduction in either hourly wages, job positions or both and from the job market data I see it can get there.
"crud web apps" sounds like WordPress or Django anyways I mean, it's already kind of valueless ? The true value lies in what this crud app is about, marketing, and the extra bits you can add to make it special.
Well, AI really can't code any more than a compiler can. They all require a human to write the original code, even the machine does translate it into other code.
And until the day that humans are no longer driving the bus that will remain the case.
You can say generate a c program that uses gcc 128 bit floats and systematically generates all quadratic roots in order of the max size of their minimal polynomial coefficients, and then sort them and calculate the distribution of the intervals between adjacent numbers, and it just does it. That's qualitatively different from the compilers I have used. Now I was careful to use properly technical words to pull in the world of numeric computation and c programming. But still saved me a lot of time. It was even able to bolt on multithreaded parallelism to speed it up using c stuff I never heard of.
> That's qualitatively different from the compilers I have used.
Is it? I can, in most traditional programming languages commonly used to today, using decades old compiler technology, say something like "x = [1,2,3]" and it will, for example, systematically generate all the code necessary to allocate memory without any need for me to be any more explicit about it. It would be fair to say AI offers an even higher level abstraction, like how most programming languages used today are a higher level abstraction over assembly, but fundamentally different it is not.
"generate a c program that uses gcc 128 bit floats and systematically generates all quadratic roots in order of the max size of their minimal polynomial coefficients, and then sort them and calculate the distribution of the intervals between adjacent numbers" is just code. You still have to write the code get AI to translate it into a lower-level abstraction. It doesn't magically go off and do its own autonomous thing.
There is so much more than coding, it’s the business rules, the constraints held in other parts of the company, the undocumented code that holds a company with no written context … you start a project from scratch ? AI can help you but you better be highly specific because hallucinations and errors will be there along the way. Now for the hundreds or billion of loc of private source code softwares that holds thousands of companies with not tightly formatted context and most of it just in some people’s memory … how would AI replace that ? Try let an AI rewrite a full ERP responsible of business logic with costly consequences on business and money … It will be just a tool, where sometimes we loose time and most time save a bit of time if we use it well
I'm tired of the anthropomorphization marketing behind AI driving this kind of discussion. In a few years, all this talk will sound as dumb as stating "MS Word spell checker will replace writers" or "Photoshop will replace designers".
We'll reap the productivity benefits from this new tool, create more work for ourselves, output will stabilize at a new level and salaries will stagnate again, as it always happens.
Generative AI is replacing writers, designers, actors, ... it is nothing like just a spell checker or Phtoshop.
Everyday, I see ads on YouTube with smooth-talking, real-looking AI-generated actors. Each one represents one less person that would have been paid.
There is no exact measure of correctness in design; one bad bit does not stop the show. The clients don't even want real art. Artists sometimes refer to commercial work as "selling out", referring to hanging their artistic integrity on the hook to make a living. Now "selling out" competes with AI which has no artistic integrity to hang on the hook, works 24 hours a day for peanuts and is astonishingly prolific.
> Everyday, I see ads on YouTube with smooth-talking, real-looking AI-generated actors. Each one represents one less person that would have been paid.
Were AI-generated actors chosen over real actors, or was the alternative using some other low-cost method for an advertisement like just colorful words moving around on a screen? Or the ad not being made at all?
The existence of ads using generative AI "actors" doesn't prove that an actor wasn't paid. This is the same logical fallacy as claiming that one pirated copy of software represents a lost sale.
Yes, this. I recently used Midjourney to generate a super-convincing drone shot video for a presentation. The alternative would have been a stock photo.
Probably took me the same amount of time to generate a pleasing video as I would have spent browsing Shutterstock. Only difference is my money goes to one corporation instead of the other.
As far as the video is concerned, it adds a bit of a wow factor to get people interested, but ultimately it's the same old graphs and bullet points with words and numbers that matter. And those could just as well have been done on an overhead transparency in 1987.
A) J. Crew and others are using AI models instead of real models. Retail modeling was steady pay-the-bills work for models and actors and is actively being replaced by AI image generators— sometimes trained on images of a particular model they’re no longer paying. Writers and stock photographers are in much worse shape.
B) Even in cases where AI actors are used where there wouldn’t have been actors before, the skillset is still devalued, and even that modest insulation for higher-end work is almost certainly temporary. Someone doing a worse version of what you do for 1% of your pay affects the market, and saving 99% is great incentive for companies to change their strategy until the worse version is good enough.
It is likewise a fallacy that no pirated copy of software represents a lost sale.
Use of AI is exerting a downward pressure on artists and designers to get paid.
It's not true that AI is only servicing the pent-up demand for that kind of work from clients who would never have paid people to do it.
It means that being a live actor is less of a differentiator. Of course great movie stars will remain, playing live, or animating computer characters, no matter. But simpler works like ads featuring a human now become more accessible.
Among other things, this will remove most entry-level jobs, making senior-level actors more rare and expensive.
I think this means that personal branding is going to get even more important than it already is (for example, people watching movies specifically because of Ryan Reynolds, or avoiding them because of Jared Leto)
It's really both effects happening at once. AI is just like the invention of the assembly line, or the explosion of mass produced consumer packaged goods starting from the first cotton gin. Automation allows a massive increase in quantity of goods, and even when quantity comes with tradeoffs to quality vs artisanally produced goods, they still come to dominate. Processed cheese or instant coffee is pretty objectively worse that the 'real' thing, but that didn't stop cheap mass production still made those products compelling for many million/billion of consumers.
You can still find a tailor who will hand make you a bespoke clothing or sew your own clothes yourself (as even the boomer generation often did growing up), but tailored clothing is a tiny fraction of the amount of clothing in circulation. Do tailors and artisanal cheese makers still exist? Yep, they are not extinct. But they are hugely marginalized compared to machine-made alternatives.
the crazy thing is, I can get locally-roasted beans that are single-origin microlots from all over the world, in part because of the coffee boom that was a result of instant coffee and the desire for better.
I agree with your sentiment. But where I struggle is: to what degree do each of those ads “represent one less person who would have been paid” versus those that represent one additional person who would not be able to afford to advertise in that medium.
Of course that line of reasoning reduces similar to other automation / minimum wage / etc discussions
It reminds me of the piracy lawsuits that claimed damages as if every download would have been a sale
The extreme opposite idea that no unlicensed use of software is a lost sale is likewise a fantasy.
YouTube has the lowest quality ads of any online platform I use by several orders of magnitude. AI being used for belly fat and erectile dysfunction ads is not exactly good for its creative reputation
Local governments in BR have already made ads using generative AI that were shown during prime time TV hours[1].
You can argue that is a bad thing (local designers/content producers/actors/etc lost revenue, while the money was sent to $BigTech) or that this was a good thing (lower cost to make ad means taxpayer money saved, paying $BigTech has lower chance of corruption vs hiring local marketing firm - which is very common here).
[1]https://www.cnnbrasil.com.br/tecnologia/video-feito-com-inte...
I have no doubt there will be AI advertising. I bet it’s the primary untapped revenue stream. My argument is that it will be associated with cheap, untrustworthy products over time, even if it’s possible to spend more money and get better AI ads. Same thing as social/search ads.
There's a difference between taking one thing and putting something else in it's spot, and truly REPLACING something. Yes, some ads have AI generated actors. You know because you can tell because they're "not quite right", rather than focusing on the message of the ad. Noticing AI in ads turns more people off than on, so AI ads are treated by a lot of people as an easy "avoid this company" signal. So those AI ads are in lieu of real actors, but not actually replacing them because people don't want to watch AI actors in an ad. The ad ceases to be effective. The "replacement" failed.
Realistic video generation only became a thing in the last year or so.
How long do you suppose it will be before we can't tell the difference between it and reality anymore? A few years at the most. Then what?
I don't think AI will ever be able to compete with real actors, not in a meaningful way.
Animated films have competed for box office dollars since basically the dawn of cinema. Animated characters have fan followings.
Just wait; the stuff is coming. Ultra-realistic full-length feature films with compelling AI characters that are not consistent from beginning to end, but appear in multiple features.
The public will swallow it up.
“ Everyday, I see ads on YouTube with smooth-talking, real-looking AI-generated actors. Each one represents one less person that would have been paid.”
The thing is that they would not have paid for the actor anyway. It’s that having an “actor” and special effects for your ads cost nothing, so why not?
The quality of their ads went up, the money changing hands did not change.
Anecdata: I know writers, editors, and white collar non-tech workers of all kinds who use AI daily and like it.
When GPT3.5 first landed a lifelong writer/editor saw a steep decrease in jobs. A year later the jobs changed to "can you edit this AI generated text to sound human", and now they continue to work doing normal editing for human or human-ish writing while declining the slop-correction deluge because it is terrible work.
I can't help but see the software analogy for this.
And as people get more used to the patterns of AI it’s getting called out more and more.
> Each one represents one less person that would have been paid
or equally, one more advert which (let's say rightly) wouldn't have been made.
seriously though, automation allows us to do things that would not have been possible or affordable before. some of these are good things.
I'm not a "real coder" either, but it sounds like the "No True Scotsman" trap when people say, “AI can’t be a real coder,” and then redefine “real coder” to mean something AI can’t currently do (like full autonomy or deep architectural reasoning). This makes the claim unfalsifiable and ignores the fact that AI already performs several coding tasks effectively. Yeah, I get it, context handling, long-horizon planning, and intent inference all stink, but the tools are all 'real' to me.
That's based on the assumption models would not soon cross that treshold of autonomy and self-reflection that suddenly makes an escalating number of jobs (with cheap humanoids, even physical) automatable for ridiculous pricing. Even if this isn't certain, likelihood could be considered quite high and thus we urgently need a public debate / design process for the peaceful, post-commercial, post-competitive, open-access post-scarcity economy some (RBE / commoning community) have been sketching for years and years. Seems this development defies most people's sense of imagination - and that's precisely why we need to raise public awareness for the freedom and fun OPEN SOURCE EVERYTHING & Universal Basic Services could bring to our tormented world. 2 billion without access to clean water? we can do much better if we break free from our collective fixation on money as the only means and way to deal with things ever.
People are stupid, always have been - took thousands of years to accept brain as the seat of thought because “heart beat faster when excited, means heart is source of excitement”.
Heck, people literally used to think eyes are the source of light since everything is dark when you close them.
People are immensely, incredibly, unimaginably stupid. It has taken a lot of miracles put together to get us where we are now…but the fundamentals of what we are haven’t changed.
> took thousands of years to accept brain as the seat of thought because “heart beat faster when excited, means heart is source of excitement”
So what you are saying is that beings without a central nervous system cannot experience "excitement"?
or perhaps the meaning of too many words has changed, and their context. When Hippocrates claimed that the brain was an organ to cool the blood, perhaps he meant that we use our thought to temper our emotions, i.e. what he said agrees with our modern understanding.
However, many people read Hippocrates and laugh at him, because they think he meant the brain was some kind of radiator.
Maybe because we stopped talking about "excitable" people as being "hot-blooded"
>or perhaps the meaning of too many words has changed, and their context. When Hippocrates claimed that the brain was an organ to cool the blood, perhaps he meant that we use our thought to temper our emotions, i.e. what he said agrees with our modern understanding.
The belief that the heart was the seat of thought and emotion was shared by numerous cultures[0], and was based on their naive interpretation of physiology and biology and cannot be dismissed as a modern misinterpretation of a single vague aphorism by a single person due to the preponderance of documentary evidence to the contrary from contemporary sources. Also, you're probably talking about Aristotle, not Hippocrates.
>Maybe because we stopped talking about "excitable" people as being "hot-blooded"
Also people still say "hot blooded" all the time.
[0]https://en.wikipedia.org/wiki/Cardiocentric_hypothesis
You're confusing ignorance with stupidity. People at the time were coming to the best conclusions they could based on the evidence they had. That isn't stupid. If humans were truly "incredibly, unimaginably stupid" we wouldn't have even gotten to the point of creating agriculture, much less splitting the atom. We didn't get here through "miracles," we got here through hard work and intelligence.
Stupid is people in 2025 believing the world is flat and germ theory is a hoax. Ignorance becomes stupidity when our species stands on the shoulders of giants but some people simply refuse to open their eyes.
You say it as a joke, but spell check has replaced certain tiers of editors. And Photoshop has replaced certain tiers of designers.
Not a joke.
Proofreaders still exist, despite spell checker. Art assistants still exist, despite Photoshop. There's always more work to do, you just incorporate the new tools and bump the productivity, until it gets so commoditized it stops being a competitive advantage.
Saying AI "replaces" anyone is just a matter of rhetoric to justify lower salaries, as always.
Bad ones
In a few years AI will have progressed a fair bit in a way that MS spell checker didn't.
> tired of anthropomorphization
The thing is trained on heaps and heaps of human output. You better anthropomorphize if you want to stay ahead of the curve.
> all this talk will sound as dumb as stating "MS Word spell checker will replace writers" or "Photoshop will replace designers".
You cannot use just a spell checker to write a book (no matter how bad) or photoshop (non-AI) plugins to automatically create meaningful artwork, replacing human intervention.
Business people "up the ladder" are already threatening with reducing the workforce and firing people because they can (allegedly) be replaced by AI. No writer was ever threatened by a spellchecker.
Hollywood studio execs are putting pressure on writers, and now they can leverage AI as yet another tool against them.
I'm tired of all the "yet another tool" reductionism. It reeks of cope.
It took under a decade to get AI to this stage - where it can build small scripts and tiny services entirely on its own. I see no fundamental limitations that would prevent further improvements. I see no reason why it would stop at human level of performance either.
There’s this saying that humans are terrible at predicting exponential growth. I believe we need another saying, those who expect exponential growth have a tough time not expecting it.
It’s not under a decade for ai to get to this stage but multiple decades of work, with algorithms finally able to take advantage of gpu hardware to massively excel.
There’s already feeling that growth has slowed, I’m not seeing the rise in performance at coding tasks that I saw over the past few years. I see no fundamental improvements that would suggest exponential growth or human level of performance.
I'm not sure if there will be exponential growth, but I also don't believe that it's entirely necessary. Some automation-relevant performance metrics, like "task-completion time horizon", appear to increase exponentially - but do they have to?
All you really need is for performance to keep increasing steadily at a good rate.
If the exponential growth tops out, and AI only gains a linear two days per year of "task-completion time horizon" once it does? It'll be able to complete a small scrum sprint autonomously by year 2035. Edging more and more into the "seasoned professional developer" territory with each passing year, little by little.
> ... entirely on its own
ok, ok! just like you can find for much less computation power involved using a search engine, forums/websites having if not your question, something similar or a snippet [0] helping you solve your doubt... all of that free of tokens and companies profiting over what the internet have build! even FOSS generative AI can give billions USD to GPU manufacturers
[0] just a silly script that can lead a bunch of logic: https://stackoverflow.com/questions/70058132/how-do-i-make-a...
You can’t see any bottlenecks? Energy? Compute power? Model limitations? Data? Money?
there are more of all these bottlenecks among the proprietary or open source project worlds, which have yet to collaborate amongst themselves to unify the patterns in their disparate codebases and algorithms into a monolith designed to compress representations of repeated structures edited for free by a growing userbase of millions and the maturing market of programmers who grew up with cheap GPUs and reliable optimization libraries
the article's subtitle is currently false, people collaborate more with the works of others through these systems and it would be extremely difficult to incentivize any equally signifciant number of the enterprise software shops, numerics labs, etc to share code: even joint ventures like Accenture do not scrape all their own private repos and report their patterns back to Microsoft every time they re-implement the same .NET systems over and over
So maybe the truth is somewhere in between - there is no way AI is not going to have a major societal impact - like social media.
If we don't see some serious fencing, I will not be surprised by some spectacular AI-caused failures in the next 3 years that wipe out companies.
Business typically follows a risk-based approach to things, and in this case entire industries are yolo'ing.
> I see no fundamental limitations
How about the fact that AI is only trained to complete text and literally has no "mind" within which to conceive or reason about concepts? Fundamentally, it is only trained to sound like a human.
The simplest system that acts entirely like a human is a human.
An LLM base model isn't trained for abstract thinking, but it still ends up developing abstract thinking internally - because that's the easiest way for it to mimic the breadth and depth of the training data. All LLMs operate in abstracts, using the same manner of informal reasoning as humans do. Even the mistakes they make are amusingly humanlike.
There's no part of an LLM that's called a "mind", but it has a "forward pass", which is quite similar in function. An LLM reasons in small slices - elevating its input text to a highly abstract representation, and then reducing it back down to a token prediction logit, one token at a time.
> The simplest system that acts entirely like a human is a human.
LLM's do not act entirely like a human. If they did, we'd be celebrating AGI!
They merely act sort of like a human. Which is entirely expected - given that the datasets they're trained on only capture some facets of human behavior.
Don't expect them to show mastery of spatial reasoning or agentic behavior or physical dexterity out of the box.
They still capture enough humanlike behavior to yield the most general AI systems ever built.
It doesn’t develop any thinking, it’s just predicting tokens based on a statistical model.
This has been demonstrated so many times.
They don’t make mistakes. It doesn’t make any sense to claim they do because their goal is simply to produce a statistically likely output. Whether or not that output is correct outside of their universe is not relevant.
What you’re doing is anthropomorphizing them and then trying to explain your observations in that context. The problem is that doesn’t make any sense.
When you reach into a "statistical model" and find that it has generalized abstracts like "deceptive behavior", or "code error"? Abstracts that you can intentionally activate or deactivate - making an AI act as if 3+5 would return a code error, or as if dividing by zero wouldn't? That's abstract thinking.
Those are real examples of the kind of thing that can be found in modern production grade AIs. Not "anthropomorphizing" means not understanding how modern AI operates at all.
I don't think you have any idea what you're talking about at all.
You've clearly read a lot of social media content about AI, but have you ever read any philosophy?
your comment reeks of hype. no evidence whatsoever for your prediction, just an assertion that you personally don't see it not coming true
It took closer to 100 years for AI to get to this stage. Check out: https://en.wikipedia.org/wiki/History_of_artificial_intellig...
I suspect once you have studied how we actually got to where we are today, you might see why your lack of seeing any limitations may not be the flex you think it is.
When are you starting time from? AI has been a topic of research for over 70 years
We see massive initial growth followed by a slowdown constantly.
There is zero reason to think AI is some exception that will continue to exponentially improve without limit. We already seem to be at the point of diminishing returns. Sinking absurd amounts of money and resources to train models that show incremental improvements.
To get this far they have had to spend hundreds of billions and have used up the majority of the data they have access to. We are at the point of trying to train AI on generated data and hoping that it doesn’t just cause the entire thing degrade.
> I see no fundamental limitations that would prevent further improvements
How can you say this when progress has so clearly stagnated already? The past year has been nothing but marginal improvements at best, culminating in GPT-5 which can barely be considered an upgrade over 4o in terms of pure intelligence despite the significant connotation attached to the number.
Marginal improvements? Were you living under a rock for the past year?
Even o1 was a major, groundbreaking upgrade over 4o. RLVR with CoT reasoning opened up an entire new dimension of performance scaling. And o1 is, in turn, already obsoleted - first by o3, and then by GPT-5.
>> It reeks of cope.
haha, well said, I've got to remember that one. HN is a smelly place when it comes to AI coping.
I’ve seen comments here claiming that this site is either a bunch of coders coping about the limitations of AI and how it can’t take their job, or a bunch of startup dummies totally on the AI hype train.
Now, there’s a little room between the two—maybe the site is full of coders on a cope train, hoping that we’ll be empowered by nice little tools rather than totally replaced. And, ya know, multiple posters with multiple opinions, some contradictions are expected.
But I do find it pretty funny to see the multiple posters here describe the site they are using as suffering from multiple, contradictory, glaringly obvious blindspots.
I think everyone is looking for back and white switches. Either coding agents are writing your code or they aren't. Humans will always be in the mix in some form, but the amount and skills they use is going to be radically different as time goes on.
I personally haven't written any significant code by hand since claude code landed. I also have a high tolerance for prompting and re-prompting. Some of my colleagues would get upset if it wasn't mostly one shotting issues and had a really low tolerance for it going off the rails.
Since gpt-5-high came out, I rarely have to re-prompt. Strong CI pipeline and well defined AGENTS.md goes an incredibly long way.
I can't help thinking that a big factor is how bad our tools were before. There are many leaking abstractions, boilerplate code and cruft. Dealing with that fills a sizeable percent of a programmer's job.
These tools seem great because they are less sensitive than humans to the mess and lift us over the tedious work. But at the same time, they're giving us an excuse to not fix what needed to be fixed and, in doing so, they're adding more crap to the heap.
Maybe what we need is forcing the tools to build on a simpler base, so we can keep an understanding of the results.
True, but it's catching up fast. A year ago, I used AI for small OS scripts. It worked fine and saved me the time of looking up switches for commands. Now, I can ask it to create a simple game of about 200 lines, and it does a pretty good job of writing bug-free code within a few seconds. It's only going to get better. Even if the tech doesn't improve further, I can see a future where all apps are endlessly configurable.
A big part of my career has been the modification of enterprise software to fit a company's needs. Rarely was any one addition more than a few hundred lines of code. I can see a future where there will be simple options for a non-coder to add to an app.
True, it's not a coder, but that doesn't mean it won't fundamentally change how apps are made and it will reduce the number of master programmers needed. It won't replace all programmers, but it will greatly reduce the number that are needed, which country they work in and the language they use to program apps.
Programming has mainly been a career that requires the individual to understand English. That is changing. I can see a future where code can be created in multiple human languages. Programming was well-paid because relatively few people had the expertise to do it. That won't be the case, and the pay will adjust downward as needed. AI might not be a coder, but it will let many more people become coders. In the future, coding will be in the same pay range as clerical work. Companies will be hiring Programming Clerks rather than Programming Engineers.
> Programming has mainly been a career that requires the individual to understand English.
Disagree, programming is a career where in order to be good you can
1. Break down a big problem into smaller ones, creating abstractions
2. Implement those abstractions one by one to end up with a full program
3. Refactor those abstractions if requirements change or (better) reimplement an abstraction a different way
You do all of this to make complex software digestible by a human, so that they don't have to have the entire system 'in context'.
This prophesied view of software development will mean you end up with code that's likely only maintainable by the model itself.
I can't imagine the vendor lock in of that.... You have the source, but it is in such a state that no human can maintain it?
> I can't imagine the vendor lock in of that.... You have the source, but it is in such a state that no human can maintain it?
It’s much worse than that.
What happens when the erroneous output caused by model blind spots gets fed back into the model?
Those blind spots get reinforced.
Doesn’t matter how small that error rate is (and it’s not small). The errors will compound.
Vendor lock-in won’t matter because it will simply stop working/become totally unrecoverable.
I think you're right that LLMs are democratizing access to coding, but unless and until AI models reach a point where they can say 'no' to their users, the scenario you're imagining ('endlessly configurable apps') will probably lead to software that collapses under its own complexity.
Years ago, I supported a team of finance professionals who were largely quite competent at coding but knew nothing about software engineering. They had thousands of scripts and spreadsheets: they used version control, but kept separate long-lived branches for client-specific variations of different models. There were no tests for anything; half the tools would break when the clocks changed.
They weren't dumb, but their incentives weren't about building anything we might recognize as an engineered application. I suspect something similar will happen turning end users loose with AI.
My best understanding/explanation of LLM tools is "plausible extension of a context"
That is, you have some context, ie the prompt and any other text, and the LLM produces a plausible continuation or alteration of that prompt and text.
My intuition leads me to a thought like: To progress, the context must compress into a fractal representation.
I feel very confident that someone smarter and MUCH better paid than me is already working on this.
There's been a strong theme recently here on HN of confusing programming (the act of writing code to meet specifications) and Engineering(the writing of specifications, and the oversight of said process, along with overview of testing).
AI is definitely not ready for an Engineering role. My recent experience with ChatGPT5(Preview) via Visual Studio Code tells me that it might perform acceptably as a junior programmer. However, because I'm an old retired self taught programmer who only ever managed 1 other programmer, I lack the experience to know what's acceptable as a Junior Programmer at FAANG and elsewhere.
> There's been a strong theme recently here on HN of confusing programming (the act of writing code to meet specifications) and Engineering(the writing of specifications, and the oversight of said process, along with overview of testing).
You're making a distinction that might be interesting in some esoteric sense, but that doesn't exist in the industry. When it comes to software, the architects are also the construction crew. There's nothing to confuse if the terms are effectively synonymous.
You're making a distinction that might be interesting in some esoteric sense, but that doesn't exist in the industry.
Sure it exists. Not everywhere, not for all software, but it does exist.
I think this is generally true, and there are SO MANY blog posts and articles. Using something like Claude Code to build an entire SaaS from nothing can seem like magic, but there is still a point where the code is too big and any LLM will lose context.
But there is a "sweet spot" where it's amazing, specifically highly targeted tasks with a specific context. I wanted a simple media converter app that tied into ffmpeg, and I didn't want to install any of the spammy or bloated options I found... so I got Claude to build one. It took about 30 minutes and works great.
I also asked it to update some legacy project, and it fell down a testing a loop where it failed to understand the testing database was missing. Had I not had years of knowledge, I would've looked at the output and suggestions Claude was giving and spent hours on it... but it was a simple command that fixed it. As with all new tech, your milage will vary.
Anecdotes are unreliable, for one, your described use case and the tools you are using suggest you are at a very basic level and unable to extract the full capabilities of the tooling and models which many of use to solve if not complete complex software.
Just be aware that AI is a tool not a replacement but a human apt at AI as a tool will replace the former.
No offense but your experience with AI is fairly primitive if that's where you're at.
Please elaborate.
Share what you built and how you prompted and what you are making from it, and how many tokens you paid to rummage through.
I didn't want to get into the details, because I've already talked about BitGrid here endlessly, and was trying to stay on the topic of AI usefulness, but since you asked.
I'm trying to build a software stack that can eventually take something like a PyTorch model, and unwind everything, resulting in a directed acyclic graph of individual bit-level operations (OR, AND, XOR). That graph will then be compiled into a bitstream suitable for an FPGA-like substrate that eliminates the memory/compute divide, the BitGrid[1].
FPGA routing is a non-trivial problem, I'm hoping to get it down to seconds. I'm currently trying to build the software stack to make it usable.
The goal is to answer questions about BitGrid:
If the answers are favorable, then in the deep (and improbable) future, it's possible that there could be a set of racks with an array of thousands resulting in a system that could stream ChatGPT at aggregate rate of a gigatoken per second, for far less than the Trillion dollars Meta plans to spend.This isn't just some CRUD application with a web front end. There are a number of layers of abstraction at play, and the LLMs seem to handle it well if you limit the depth under consideration.
[1] BitGrid eliminates the traditional memory/compute divide that causes most of the energy consumption of CPUs, GPUs, and other accelerators. Even FPGA systems tend to focus on emulation of these models, and routing fabric for minimum latency, instead of maximum performance. Because all the active lines only reach nearest neighbors, power consumption for a given operation can be far lower than the traditional approach.
PS: I pay $10/month for GitHub CoPilot, which apparently now includes ChatGPT5
you can't proclaim any sort of knowledge about AI's currently capabilities by opening up a codebase and typing a couple prompts into the vscode agent.
I agree, using "chatgpt 5 in visual studio code" screams unfamiliarity with what is current.
just stop yeah?
The article is so banal
AI is apparently not ready to take my order at Popeyes either
“If it takes longer to explain to the system all the things you want to do and all the details of what you want to do, then all you have is just programming by another name”
If it's taking you that long to direct the AI, then either you're throwing too small a problem at it, or too big a problem at it, or you're not directing its attention properly.
In a way, your prompts should feel like writing user documentation:
I feel you've done a decent job of disguising "you're not approaching it with the right mindset".
It's more "you need to learn how a tool really works in order to use it most effectively"
Talking to a chat bot like you'd talk to a human is a quick way to be disappointed. They don't work that way, despite sometimes sounding like they do.
But don't forget, "we don't know how it really works".
More like "it is impossible to infer anything useful from what we know how they work".
But the end result is the same.
And then for some bizarre reason sometimes it doesn't really work, AI added a bunch of random shit, then you can feel your rage bubbling up in real-time as you have to re-prompt it.
I've yet to have that happen. But then again so far I've only used it for Rust, and it's hard for the AI to maintain hallucinations with such a strict compiler.
“Feels primitive,” is holding up a lot of assumptions.
What feels primitive to me is how we approach programming in industry as a process of trial and error rather than one of rigour.
These are tools that automate copy-pasting from Stack Overflow and GitHub, running tools, and generating a ton of noise to sift through. They hallucinate code, documentation, and various other artifacts that are sometimes useful and are occasionally complete BS.
Some people find that they can make useful tools out of these things. Great.
A real programmer is still a human.
Update: nothing wrong with trial and error as a process. I use it a lot. But there are lots of places where we use this method that seem inappropriate and sometimes even dangerous. Yet it’s the most common tool we have and everything starts to look like a nail.
But what is meant by "real coder"? Does it mean an experienced senior coder? In that case, no, AI isn't ready to replace one. How about a junior coder who is fresh out of university and starting his first job? Surely that counts as a "real coder"? I'd argue that AI can and does already replace such coders, depending on the type of work they have been assigned.
Could you elaborate your argument? I'd say AI is far from replacing any human contributor at this point.
Context doesn't work the same way as memory + experience in humans. While humans have an impression and a flexible mental model of any single domain, AI needs hard data, which is hard to manage with context and can't really be worked around by fine-tuning in practice, lest you have to retrain the model on each and every code merge.
Sure. My point is that it's not unusual for the quality of work of a junior coder to be fairly low, and that work typically needs to be double checked by someone more senior before it can be used in production. Often, an LLM will produce a higher quality output than that junior developer in which case, given that the work has to be checked by a senior coder no matter what, the LLM can replace the junior developer.
Also, not every project is highly complex or involves huge codebases and, for those that are, a junior developer might be assigned to a reasonably self contained and small module within the project (i.e. something that a LLM can do better at).
Finally, although this wasn't part of my original argument: before modern LLMs, let's say that a senior developer could produce N units of work per Y units of time. Now, with the help of a LLM, they are able to produce XN units of work per Y unit of time. I don't think it's too controversial to claim that some portion of developers are now more productive than they were before. And if developers in general are more productive, then you can expect some reduction in the number of developers that you need for a given project. So, in an indirect sense, LLMs are already capable of replacing coders.
The pattern I see time & time again in software development: a small set of few and usually very talented senior developers build a software project and they build examples of how everything is done. Junior developers then extend those examples, eg: "add another column to this table." Juniors are just extending the existing patterns, seniors are the ones that create them.
This pattern I see play out time and time again. Senior builds the integration to various team services, then juniors tweak & mimic those to do more.
> let's say that a senior developer could produce N units of work per Y units of time.
So, it's more the case that senior developers are building full production assembly lines, that can then be replicated to do similar things relatively easily. The cost of the units of work are not at all equal.
How does this apply to AI? Good luck getting AI to do something novel, by definition it can't. OTOH, extending an existing pattern is trivial for a senior developer to do after they have built that pattern; juniors spend a lot of time at it cause they have to learn everything.
So, who is the AI going to replace? If the junior, then you have a senior developer using an AI to do something that is already trivial for them to do. No way AI is replacing the senior.. So who? Who is getting replaced by the AI and who is using it?
My #1 issue with AI for programming is that while I do trust it to produce OK boilerplate code, I can also just download that boilerplate code, if I haven't already written it myself. Free software is the answer.
This is profound, and I think you have dove to the core of why LLMs are taking off for code.
They fix a problem: the mutual incompatibility between software. "Free software" is not the answer, as it's used today: for example, the canonical Free Software OS, Linux, is fragmented into many mutually incompatible ABIs.
We need to tear down barriers between software. Barriers between languages, frameworks and OS ABIs. I don't have a good plan , but think it is the true solution. LLMs paper over these incompatibilities, and in doing so, provide value. Imperfect, but good enough until we destroy the barriers properly.
Example: Why am I using an LLM for my rust code? Because it is taking code that solves my problem, but in Python, C, or papers, and translating in a way compatible with the rust dialect I use. The problem is solved multiple times, but not in a way I can use without a translation layer. LLMs provide that.
"According to the study, AI still struggles with several crucial facets of coding: sweeping scopes involving huge codebases, the extended context lengths of millions of lines of code, higher levels of logical complexity, and long-horizon or long-term planning about the structure and design of code to maintain code quality."
uhhh, not sure even the best people or teams are very good at this either. Condemning AI for not being capable of something we're not capable of, ok...
“If it takes longer to explain to the system all the things you want to do and all the details of what you want to do, then all you have is just programming by another name.”
This is called the specification process, which hopefully is already occurring today.
There's so much self-serving bias in articles like this, as well as the comments on HN, Reddit, etc. It's good to critique AI, but that self-serving line is frequently crossed by many people.
> These kinds of complex tasks can be difficult for AI development tools to navigate, resulting in hallucinations about where the bug is or its root cause, as well as irrelevant suggestions or code fixes with subtle problems.
How is this any different than the way I program?
I've contributed genuinely useful features to FLOSS projects "as well as irrelevant suggestions or code fixes with subtle problems", mostly the latter as there was always a few stages of improvement and/or finishing by the core devs of the program I used to haunt. Honestly, I was less than half as useful as the current crop of robots and they still tolerated (in fact, encouraged) my involvement.
> Honestly, I was less than half as useful as the current crop of robots and they still tolerated (in fact, encouraged) my involvement.
Don’t sell yourself short. External contributors are extremely valuable as they are often users and provide real world validation of a need for whatever they are contributing. They also retain any knowledge for any feedback they receive that they can apply to future contributions. And they also become advocates for that software, helping it grow its user base.
LLMs, not so much.
AI can't replace humans, but it can be an exoskeleton for human pilots.
AI code tab-complete is fantastic. It's at least an order of magnitude more powerful than IDE-assisted auto refactors.
AI graphics design tools are probably the single best thing in the field. Editing photos, creating new graphics, making marketing materials, shooting and editing videos is now extremely easy. It's a 100x speed up and a 1000x cost reduction. You still have to re-roll the generations repeatedly, but with a competent editing tool you can speed run any design work. This is one area where non-experts can also use the tools.
Can AI categorically not be a "real coder" or is the limitation in more trivial things such as scope and breadth? My experiences suggest to me that it is technically quite close.