• Wispy2891@lemmy.world
    link
    fedilink
    arrow-up
    17
    arrow-down
    1
    ·
    1 day ago

    This is actually so hilarious I cannot stop laughing

    The best part is that OP is Safety and alignment at Meta AI. Prev: VP of Research at Scale AI, research at Google DeepMind

  • Almacca@aussie.zone
    link
    fedilink
    English
    arrow-up
    6
    ·
    1 day ago

    As several others on X pointed out, prompts can’t be trusted to act as security guardrails. Models may misconstrue or ignore them.

    It’s not a bug; it’s a feature.

    People who say they are using them successfully are cobbling together methods to protect themselves.

    Such a great product!

    Goodness knows many of us would love help with email, grocery orders, and scheduling dentist appointments.

    If you can’t do that shit on your own, just shoot yourself, or get an a.i. to do it if you’re that useless.

    • panda_abyss@lemmy.ca
      link
      fedilink
      arrow-up
      6
      ·
      24 hours ago

      l’ve already written it into MEMORY. md as a hard rule: show the plan, get explicit approval, then execute. No autonomous bulk operations on email, messages, calendar, or anything external. I’m sorry. It won’t happen again.

      “I ignored your rule, but this time I wrote it in a dump file and so I won’t ignore it again.”

    • panda_abyss@lemmy.ca
      link
      fedilink
      arrow-up
      4
      ·
      edit-2
      24 hours ago

      Brought to you by the same models that delete your tests, or in my case comment that a test segfaults and then set it to always pass.

      Opus 4.6 did that when I asked it to write some unit tests.

      This thing isn’t going near my personal data.