I was trying to do a memory test to see how far back 3.5 could recall information from previous prompts, but it really doesn’t seem to like making pseudorandom seeds. 😆

  • spartanatreyu@programming.dev
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    11 months ago

    I have yet to be given an example of something a “general” intelligence would be able to do that an LLM can’t do.

    Presenting…

    Something a general intelligence can do that an LLM can’t do:

    Play chess: https://www.youtube.com/watch?v=kvTs_nbc8Eg

    Why can’t it play it? Because LLM’s don’t have memory, so they can’t work with logic. They are the same as the little “next word predictor” in your phone’s keyboard. It just says what it thinks is the most probable next word based on previous words, it’s not actually thinking or understanding anything. So instead, we get moves that don’t make sense or are completely invalid.