• iopq@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    arrow-down
    8
    ·
    22 days ago

    I’m still waiting for the definition of intelligence that won’t have the same moving of goalposts the Turing Test did

    • Barbarian@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      14
      ·
      22 days ago

      I’m happy with the Oxford definition: “the ability to acquire and apply knowledge and skills”.

      LLMs don’t have knowledge as they don’t actually understand anything. They are algorithmic response generators that apply scores to tokens, and spit out the highest scoring token considering all previous tokens.

      If asked to answer 10*5, they can’t reason through the math. They can only recognize 10, * and 5 as tokens in the training data that is usually followed by the 50 token. Thus, 50 is the highest scoring token, and is the answer it will choose. Things get more interesting when you ask questions that aren’t in the training data. If it has nothing more direct to copy from, it will regurgitate a sequence of tokens that sounds as close as possible to something in the training data: thus a hallucination.

      • assassin_aragorn@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        ·
        22 days ago

        This can be intuitively understood if you’ve gone through difficult college classes. There’s two ways to prepare for exams. You either try to understand the material, or you try to memorize it.

        The latter isn’t good for actually applying the information in the future, and it’s most akin to what an LLM does. It regurgitates, but it doesn’t learn. You show it a bunch of difficult engineering problems, and it won’t be able to solve different ones that use the same principle.

      • theherk@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        4
        ·
        22 days ago

        The human could be described in very similar terms. People think we’re magic or something, but we to are just a weighted neural network assembling outputs based strictly on training data built from reinforcement. We are just for the moment much much better with massive models. Of course that is reductive but many seem to forget that brains suffer similarly when outside of training data.