• GBU_28@lemm.ee
    link
    fedilink
    English
    arrow-up
    7
    ·
    11 months ago

    A huge value add of.chatgpt is that you can have running, contextual conversation. That requires memory.

    • GamingChairModel@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      11 months ago

      All of these LLMs should have walls between individual users, though, so that the chat history of one user is never accessible to any other user. Applying some kind of restriction to the LLM training and how chats are used is a conversation we can have, but the article and the example given is a much, much simpler problem that a user checking his own chat history was able to see other user’s chats.

    • Farid@startrek.website
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      1
      ·
      edit-2
      11 months ago

      It doesn’t actually have memory in that sense. It can only remember things that are in the training data and within its limited context (4-32k tokens, depending on model). But when you send a message, ChatGPT does a semantic search of everything in the conversation and tries to fit the relevant parts inside the context, if there’s room.

      • GBU_28@lemm.ee
        link
        fedilink
        English
        arrow-up
        6
        ·
        edit-2
        11 months ago

        I’m familiar, it’s just easiest for the layman to consider the model having “memory” as historical search is a lot like it at arm’s length