• itkovian@lemmy.world
    link
    fedilink
    arrow-up
    24
    ·
    22 days ago

    Haven’t used any coding LLMs. I honestly have no clue about the accuracy of the comic. Can anyone enlighten me?

    • Clearwater@lemmy.world
      link
      fedilink
      arrow-up
      6
      ·
      22 days ago

      They’re okay for tasks which are reasonably a single file. I use them for simple Python scripts since they generally spit out something very similar to what I’d write, just faster. However there is a tipping point where a task becomes too complex and they fall on their face and it becomes faster to write the code yourself.

      I’m never going to pay for AI, so I’m really just burning the AI company’s money as I do it, too.

    • hotdogcharmer@lemmy.world
      link
      fedilink
      arrow-up
      6
      ·
      22 days ago

      I point blank refuse to use them. I’ve seen how they’ve affected my coworker and my boss - these two people now simply cannot read documentation, do not trust their own abilities to write code, and cannot debug anything that they write. My job has become more difficult since this shit started being pushed on us all.

    • Draces@lemmy.world
      link
      fedilink
      arrow-up
      10
      arrow-down
      4
      ·
      22 days ago

      The comic is only accurate if you expect it do everything for you, you’re bad at communicating, and you’re using an old model. Or if you’re just unlucky

      • jj4211@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        21 days ago

        I’d add it depends also on your field. If you spend a lot of time assembling technically bespoke solutions, but they are broadly consistent with a lot of popular projects, then it can cut through a lot in short order. When I come to a segment like that, LLM tends to go a lot further.

        But if you are doing something because you can’t find anything vaguely like what you want to do, it tends to only be able to hit like 3 or so lines of useful material a minority of the time. And the bad suggestions can be annoying. Less outright dangerous after you get used to being skeptical by default, but still annoying as it insists on re emphasizing a bad suggestion.

        So I can see where it can be super useful, and also how it can seem more trouble than it is worth.

        Claude and GPT have been my current experience. The best improvement I’ve seen is for the suggestions getting shorter. Used to have like 3 maybe useful lines bundled with a further dozen lines of not what I wanted. Now the first three lines might be similar, but it’s less likely to suggest a big chunk of code.

        Was helping someone the other day and the comic felt pretty accurate. It did exactly the opposite of what the user prompted for. Even after coaxing it to be in the general ballpark, it has about half the generated code being unrelated to the requested task, with side effects that would have seemed functional unless you paid attention and noticed that throughout would have been about 70% lower than you should expect. Was a significant risk as the user was in over their head and unable to understand the suggestions they needed to review, as they were working in a pretty jargon heavy ecosystem (not the AI fault, they had to invoke standard libraries that had incomprehensible jargon heavy syntax)

    • SpaceCowboy@lemmy.ca
      link
      fedilink
      arrow-up
      5
      ·
      22 days ago

      Yeah kinda. I ask it to do something simple like create a a typescript interface for some JSON and it just gives me what I want… most of the time.

      Other times it will explain to me what JSON is, what Typescript is, what interfaces are and how they’re used, blah blah, and somewhere in there there’s the code I actually wanted. Once it helpfully commented the code… in Korean. Even when it works and comments things in English the comments can be kinda useless since it doesn’t actually know what I’m doing.

      It’s trying to give you what you want but can sometimes get confused about what you’re asking for and give a bunch of stuff you didn’t actually want. So yeah, the comic is accurate… on occasion. But many times LLMs will give good results, and it’s getting better, so it’ll mostly work ok for simple requests. But yeah, sometimes it’ll give you a lot more stuff than what you wanted.

    • jj4211@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      21 days ago

      It’s sometimes useful, often obnoxious, sometimes both.

      It tends to shine on very blatantly obvious boilerplate stuff that is super easy, but tedious. You can be sloppy with your input and it will fix it up to be reasonable. Even then you’ve got to be careful, as sometimes what seems blatantly obvious still gets screwed up in weird ways. Even with mistakes, it’s sometimes easier to edit that going from scratch.

      Using an enabled editor that looks at your activity and suggests little snippets is useful, but can be really annoying when it gets particularly insistent on a bad suggestion and keeps nagging you with “hey look at this, you want to do this right?”

      Overall it’s merely mildly useful to me, as my career has been significantly about minimizing boilerplate with decent success. However for a lot of developers, there’s a ton of stupid boilerplate, owing to language design, obnoxiously verbose things, and inscrutable library documentation. I think that’s why some developers are scratching their heads wondering what the supposed big deal is and why some think it’s an amazing technology that has largely eliminated the need for them to manually code.

    • chuckleslord@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      22 days ago

      The only thing I trust it with is refactoring for readability and writing scripts. But I also despise LLMs, so that’s all I’d give them.

    • lemming741@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      3
      ·
      22 days ago

      Much like Amazon has an incentive to not show you the specific thing it knows you’re searching for, people theorize that these interfaces are designed to burn through your tokens.

    • BigBenis@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      21 days ago

      I find you get much better results if you talk to the LLM like you were instructing a robot, not a human. That means being very specific about everything.

      It’s the difference between, “I would like water” and, “I would like a glass filled with drinking water and a few cubes of ice mixed in”.