Avram Piltch is the editor in chief of Tom’s Hardware, and he’s written a thoroughly researched article breaking down the promises and failures of LLM AIs.

  • @nyan@lemmy.cafe
    link
    fedilink
    English
    110 months ago

    Going over it, I think that sentience (not necessarily sapience, but that would be nice) is the secret sauce. In order for me to accept that an AI knows something (as opposed to posessing data which it does not actually understand), it has to demonstrate awareness.

    So how can a text-based AI demonstrate awareness, given the constraints of the interface through which it must operate? Reliably generalizing from data not immediately part of the response to the current prompt might do it. Or demonstrating that it understands the consequences of its actions in the real world. Even just indicating that it knows when it’s making things up would be a good start.

    For instance, take the case of the ChatGPT-generated fake legal citations. An AI which would have been fed masses of information relating to law (I’d expect that to include law school textbooks, from archive.org if nowhere else) demonstrated very clearly that it did not know that making up legal cases in response to a factual query was a Very Bad Idea. It did not generalize from data outside the domain of lists of case names that would have told it not to do that, or provide any indication that it knew its actions could be harmful. That AI had data, but not knowledge.

    So we’re back to connections and conceptual models of the world again.

    • @CanadaPlus
      link
      1
      edit-2
      10 months ago

      An AI which would have been fed masses of information relating to law (I’d expect that to include law school textbooks, from archive.org if nowhere else) demonstrated very clearly that it did not know that making up legal cases in response to a factual query was a Very Bad Idea. It did not generalize from data outside the domain of lists of case names that would have told it not to do that, or provide any indication that it knew its actions could be harmful.

      I mean, was it a bad idea? For the lawyer sure, but ChatGPT was not penalised by it’s own cost function. It may well have known in some way that is was just guessing, and that generally a legal document is serious business, but it doesn’t have any reason to care unless we build one in. Alignment is a whole other dimension to intelligence.

      Reliably generalizing from data not immediately part of the response to the current prompt might do it. Or demonstrating that it understands the consequences of its actions in the real world.

      It sounds like the biggest models do this reasonably well. Commonsense reasoning would count, right?

      • @nyan@lemmy.cafe
        link
        fedilink
        English
        210 months ago

        I think I’m going to bow out of this conversation, on the grounds that I doubt either of us is going to persuade the other, which makes it pointless.

        • @CanadaPlus
          link
          110 months ago

          Alright, that’s fair. We’ll watch what happens next, it was a pleasure, honestly.