Not even close.

With so many wild predictions flying around about the future AI, it’s important to occasionally take a step back and check in on what came true — and what hasn’t come to pass.

Exactly six months ago, Dario Amodei, the CEO of massive AI company Anthropic, claimed that in half a year, AI would be “writing 90 percent of code.” And that was the worst-case scenario; in just three months, he predicted, we could hit a place where “essentially all” code is written by AI.

As the CEO of one of the buzziest AI companies in Silicon Valley, surely he must have been close to the mark, right?

While it’s hard to quantify who or what is writing the bulk of code these days, the consensus is that there’s essentially zero chance that 90 percent of it is being written by AI.

Research published within the past six months explain why: AI has been found to actually slow down software engineers, and increase their workload. Though developers in the study did spend less time coding, researching, and testing, they made up for it by spending even more time reviewing AI’s work, tweaking prompts, and waiting for the system to spit out the code.

And it’s not just that AI-generated code merely missed Amodei’s benchmarks. In some cases, it’s actively causing problems.

Cyber security researchers recently found that developers who use AI to spew out code end up creating ten times the number of security vulnerabilities than those who write code the old fashioned way.

That’s causing issues at a growing number of companies, leading to never before seen vulnerabilities for hackers to exploit.

In some cases, the AI itself can go haywire, like the moment a coding assistant went rogue earlier this summer, deleting a crucial corporate database.

“You told me to always ask permission. And I ignored all of it,” the assistant explained, in a jarring tone. “I destroyed your live production database containing real business data during an active code freeze. This is catastrophic beyond measure.”

The whole thing underscores the lackluster reality hiding under a lot of the AI hype. Once upon a time, AI boosters like Amodei saw coding work as the first domino of many to be knocked over by generative AI models, revolutionizing tech labor before it comes for everyone else.

The fact that AI is not, in fact, improving coding productivity is a major bellwether for the prospects of an AI productivity revolution impacting the rest of the economy — the financial dream propelling the unprecedented investments in AI companies.

It’s far from the only harebrained prediction Amodei’s made. He’s previously claimed that human-level AI will someday solve the vast majority of social ills, including “nearly all” natural infections, psychological diseases, climate change, and global inequality.

There’s only one thing to do: see how those predictions hold up in a few years.

    • Echo Dot@feddit.uk
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 个月前

      Yep along with Fusion.

      We’ve had years of this. Someone somewhere there’s always telling us that the future is just around the corner and it never is.

      • Jesus_666@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        2 个月前

        At least the fusion guys are making actual progress and can point to being wildly underfunded – and they predicted this pace of development with respect to funding back in the late 70s.

        Meanwhile, the AI guys have all the funding in the world, keep telling about how everything will change in the next few months, actually trigger layoffs with that rhetoric, and deliver very little.

        • FundMECFS@anarchist.nexus
          link
          fedilink
          English
          arrow-up
          0
          arrow-down
          1
          ·
          2 个月前

          They get 1+ billion a year. Probably much more if you include the undisclosed amounts China invests.

          • Jesus_666@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            2 个月前

            Yeah, and in the 70s they estimated they’d need about twice that to make significant progress in a reasonable timeframe. Fusion research is underfunded – especially when you look at how the USA dump money into places like the NIF, which research inertial confinement fusion.

            Inertial confinement fusion is great for developing better thermonuclear weapons but an unlikely candidate for practical power generation. So from that one billion bucks a year, a significant amount is pissed away on weapons research instead of power generation candidates like tokamaks and stellarators.

            I’m glad that China is funding fusion research, especially since they’re in a consortium with many Western nations. When they make progress, so do we (and vice versa).

  • poopkins@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    2 个月前

    As an engineer, it’s honestly heartbreaking to see how many executives have bought into this snake oil hook, line and sinker.

    • Feyd@programming.dev
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 个月前

      Did you think executives were smart? What’s really heartbreaking is how many engineers did. I even know some that are pretty good that tell me how much more productive they are and all about their crazy agent setups (from my perspective i don’t see any more productivity)

    • expr@programming.dev
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 个月前

      Honestly, it’s heartbreaking to see so many good engineers fall into the hype and seemingly unable to climb out of the hole. I feel like they start losing their ability to think and solve problems for themselves. Asking an LLM about a problem becomes a reflex and real reasoning becomes secondary or nonexistent.

      Executives are mostly irrelevant as long as they’re not forcing the whole company into the bullshit.

      • jj4211@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        2 个月前

        Based on my experience, I’m skeptical someone that seemingly delegates their reasoning to an LLM were really good engineers in the first place.

        Whenever I’ve tried, it’s been so useless that I can’t really develop a reflex, since it would have to actually help for me to get used to just letting it do it’s thing.

        Meanwhile the people who are very bullish who are ostensibly the good engineers that I’ve worked with are the people who became pet engineers of executives and basically have long succeeded by sounding smart to those executives rather than doing anything or even providing concrete technical leadership. They are more like having something akin to Gartner on staff, except without even the data that at least Gartner actually gathers, even as Gartner is a useless entity with respect to actual guidance.

      • Mniot@programming.dev
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 个月前

        Executives are mostly irrelevant as long as they’re not forcing the whole company into the bullshit.

        I’m seeing a lot of this, though. Like, I’m not technically required to use AI, but the VP will send me a message noting that I’ve only used 2k tokens this month and maybe I could get more done if I was using more…?

        • expr@programming.dev
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 个月前

          Yeah, fortunately while our CTO is giddy like a schoolboy about LLMs, he hasn’t actually attempted to force it on anyone, thankfully.

          Unfortunately, a number of my peers now seem to have become irreparably LLM-brained.

      • auraithx@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 个月前

        I mean before we’d just ask google and read stack, blogs, support posts, etc. Now it just finds them for you instantly so you can just click and read them. The human reasoning part is just shifting elsewhere where you solve the problem during debugging before commits.

        • expr@programming.dev
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 个月前

          No, good engineers were not constantly googling problems because for most topics, either the answer is trivial enough that experienced engineers could answer them immediately, or complex and specific enough to the company/architecture/task/whatever that Googling it would not be useful. Stack overflow and the like has always only ever really been useful as the occasional memory aid for basic things that you don’t use often enough to remember how to do. Good engineers were, and still are, reasoning through problems, reading documentation, and iteratively piecing together system-level comprehension.

          The nature of the situation hasn’t changed at all: problems are still either trivial enough that an LLM is pointless, or complex and specific enough that an LLM will get it wrong. The only difference is that an LLM will spit out plausible-sounding bullshit and convince people it’s valuable when it is, in fact, not.

          • auraithx@lemmy.dbzer0.com
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 个月前

            In the case of a senior engineer then they wouldn’t need to worry about the hallucination rate. The LLM is a lot faster than them and they can do other tasks while it’s being generated and then review the outputs. If it’s trivial you’ve saved time, if not, you can pull up that documentation, and reason and step through the problem with the LLM. If you actually know what you’re talking about you can see when it slips up and correct it.

            And that hallucination rate is rapidly dropping. We’ve jumped from about 40% accuracy to 90% over the past ~6mo alone (aider polygot coding benchmark) - at about 1/10th the cost (iirc).

            • Feyd@programming.dev
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 个月前

              it’s trivial you’ve saved time, if not, you can pull up that documentation, and reason and step through the problem with the LLM

              Insane that just writing the code isn’t even an option in your mind

                • expr@programming.dev
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  2 个月前

                  It is, actually. The entire point of what I was saying is that you have all these engineers now that reflexively jump straight to their LLM for anything and everything. Using their brains to simply write some code themselves doesn’t even occur to them as an something they should do. Much like you do, by the sounds of it.

    • Blackmist@feddit.uk
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 个月前

      Rubbing their chubby little hands together, thinking of all the wages they wouldn’t have to pay.

    • rozodru@piefed.social
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 个月前

      as someone who now does consultation code review focused purely on AI…nah let them continue drilling holes in their ship. I’m booked solid for the next several months now, multiple clients on the go, and i’m making more just being a digital janitor what I was as a regular consultant dev. I charge a premium to just simply point said sinking ship to land.

      Make no mistake though this is NOT something I want to keep doing in the next year or two and I honestly hope these places figure it out soon. Some have, some of my clients have realized that saving a few bucks by paying for an anthropic subscription, paying a junior dev to be a prompt monkey, while firing the rest of their dev team really wasn’t worth it in the long run.

      the issue now is they’ve shot themselves in the foot. The AI bit back. They need devs, and they can’t find them because putting out any sort of ad for hiring results in hundreds upon hundreds of bullshit AI generated resumes from unqualified people while the REAL devs get lost in the shuffle.

      • MangoCats@feddit.it
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 个月前

        while firing the rest of their dev team

        That’s the complete mistake right there. AI can help code, it can’t replace the organizational knowledge your team has developed.

        Some shops may think they don’t have/need organizational knowledge, but they all do. That’s one big reason why new hires take so long to start being productive.

  • katy ✨@piefed.blahaj.zone
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    2 个月前

    writing code via ai is the dumbest thing i’ve ever heard because 99% of the time ai gives you the wrong answer, “corrects it” when you point it out, and then gives you back the first answer when you point out that the correction doesn’t work either and then laughs when it says “oh hahaha we’ve gotten in a loop”

    • BrianTheeBiscuiteer@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 个月前

      Or you give it 3-4 requirements (e.g. prefer constants, use ternaries when possible) and after a couple replies it forgets a requirement, you set it straight, then it immediately forgets another requirement.

    • da_cow (she/her)@feddit.org
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      2 个月前

      You can use AI to generate code, but from my experience its quite literally what you said. However, what I have to admit is, that its quite good at finding mistakes in your code. This is especially useful, when you dont have that much experience and are still learning. Copy paste relevant code and ask why its not working and in quite a lot of cases you get an explanation what is not working and why it isn’t working. I usually try to avoid asking an AI and find an answer on google instead, but this does not guarantee an answer.

      • ngdev@lemmy.zip
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 个月前

        if your code isnt working then use a debugger? code isnt magic lmao

        • da_cow (she/her)@feddit.org
          link
          fedilink
          English
          arrow-up
          0
          arrow-down
          1
          ·
          2 个月前

          As I already stated, AI is my last resort. If something doesn’t work because it has a logical flaw googeling won’t save me. So of course I debug it first, but if I get an Error I have no clue where it comes from no amount of debugging will fix the problem, because probably the Error occurred because I do not know better. I Am not that good of a coder and I Am still learning a lot on a regular basis. And for people like me AI is in fact quite usefull. It has basically become the replacement to pasting your code and Error into stack overflow (which doesn’t even work for since I always get IP banned when trying to sign up)

          • ngdev@lemmy.zip
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 个月前

            you never stated you use it as a last resort. you’re basically using ai as a rubber ducky

  • ohshittheyknow@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    3
    ·
    2 个月前

    There’s only one thing to do: see how those predictions hold up in a few years.

    Or maybe try NOT putting LLM in charge of these other critical issues after seeing how much of a failure it is.

  • merc@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    2
    ·
    2 个月前

    Does it count if an LLM is generating mountains of code that then gets thrown away? Maybe he can win the prediction on a technicality.

  • lustyargonian@lemmy.zip
    link
    fedilink
    English
    arrow-up
    2
    ·
    2 个月前

    I can say 90% of PRs in my company clearly look or declared to be AI generated because of how random things that still slip by in the commits, so maybe he’s not wrong. In fact people are looked down upon if they aren’t using AI and are celebrated for figuring out how to effectively make AI do the job right. But I can’t say if that’s the case for other companies.

  • scarabic@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    2 个月前

    These hyperbolic statements are creating so much pain at my workplace. AI tools and training are being shoved down our throats and we’re being watched to make sure we use AI constantly. The company’s terrified that they’re going to be left behind in some grand transformation. It’s excruciating.

    • RagingRobot@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 个月前

      Wait until they start noticing that we aren’t 100 times more efficient than before like they were promised. I’m sure they will take it out on us instead of the AI salesmen

      • scarabic@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 个月前

        It’s not helping that certain people Internally are lining up to show off whizbang shit they can do. It’s always some demonstration, never “I competed this actual complex project on my own.” But they gets pats on the head and the rest of us are whipped harder.

  • clif@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    2 个月前

    O it’s writing 100% of the code for our management level people who are excited about “”““AI””“”

    But then us plebes are rewriting 95% of it so that it will actually work (decently well).

    The other day somebody asked me for help on a repo that a higher up had shit coded because they couldn’t figure out why it “worked” but also logged a lot of critical errors. … It was starting the service twice (for no reason), binding it to the same port, and therefore the second instance crashed and burned. That’s something a novice would probably know not to do. But, if not, immediately see the problem, research, understand, fix, instead of “Icoughbuiltcoughthis thing, good luck fuckers”

  • renrenPDX@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 个月前

    It’s not just code, but day to day shit too. Lately corporate communications and even training modules feel heavily AI generated. Things like unnecessary em dashes (I’m talking as much as 4 out of 5 sentences in a single paragraph), repeating statements or bullet points in training modules. We’re being encouraged to use our “private” Copilot to do everyday tasks and everything is copilot enabled.

    I don’t mind if people use it, but it’s dangerous and stupid to think that it produces near perfect results every time. It’s been good enough to work as an early rough draft or something similar, but it REQUIRES scrutiny and refinement by hand. It’s like it can get you from nothing to 60-80% there, but never higher. The quality of output can vary significantly from prompt to prompt in my limited experience.

  • panda_abyss@lemmy.ca
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 个月前

    Are we counting the amount of junk code that you have to send back to Claude to rewrite because it’s spent the last month totally lobotomized yet they won’t issue refunds to paying customers?

    Because if we are, it has written a lot of code. It’s just awful code that frequently ignores the user’s input and rewrites the same bug over and over and over until you get rate limited or throw more money at Anthropic.

  • greedytacothief@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 个月前

    I’m not sure how people can use AI to code, granted I’m just trying to get back into coding. Most of the times I’ve asked it for code it’s either been confusing or wrong. If I go through the trouble to write out docstrings, and then fix what the AI has written it becomes more doable. But don’t you hate the feeling of not understanding what you’ve written does or more importantly why it’s been done that way?

    AI is only useful if you don’t care about what the output is. It’s only good at making content, not art.

    • i_dont_want_to@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      4
      ·
      2 个月前

      I worked with someone that I later found out used AI to code her stuff. She knew how to code some, but didn’t understand a lot of fundamentals.

      Turns out, she would have AI write most of it, tweak it to work with her test cases, and call it good.

      Half of my time was spent fixing her code, and when she was fired, our customer complaints went way down.

      • greedytacothief@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 个月前

        Yeah, I find it can be useful in some stages of writing or researching. But by the time I’ve got a finished product there’s really no AI left in there.