You might’ve noticed that ChatGPT — and AI in general — isn’t good at math. There’s a reason, and it has to do with how modern AI is built.

Basically, they’re autocorrect on steroids. Which some of us have been saying for, like, ages.

  • Sloan the Serval@pawb.social
    link
    fedilink
    arrow-up
    7
    ·
    edit-2
    14 hours ago

    LLMs (I refuse to call them AI, as there’s no intelligence to be found) are simply random word sequence generators based on a trained probability model. Of course they’re going to suck at math, because they’re not actually calculating anything, they’re just dumping what their algorithm “thinks” is the most likely response to user input.

    “The ability to speak does not make you intelligent” - Qui-Gon Jin

  • baldingpudenda@lemmy.world
    link
    fedilink
    arrow-up
    17
    arrow-down
    6
    ·
    1 day ago

    Why can’t this person working on his linguist degree not able to do high level math? It’s not their specialty.

    • Tony Bark@pawb.socialOP
      link
      fedilink
      arrow-up
      20
      arrow-down
      1
      ·
      1 day ago

      To be fair, even someone with a linguist degree knows basic math. GPT can’t even get that right. That’s the biggest problem (and red flag).

    • snooggums@lemmy.world
      link
      fedilink
      arrow-up
      7
      arrow-down
      2
      ·
      1 day ago

      Programming languages are structured and have rigid syntax that fits well in a LLM model, so it spitting out working code for simple things is like having a sentence that is structured like a normal person.

      The code might not do what you are actually trying to do, or might work while being inefficient, even if it runs.