- cross-posted to:
- technology@lemmy.world
- pulse_of_truth@infosec.pub
- cross-posted to:
- technology@lemmy.world
- pulse_of_truth@infosec.pub
This is actually so hilarious I cannot stop laughing


The best part is that OP is Safety and alignment at Meta AI. Prev: VP of Research at Scale AI, research at Google DeepMind
As several others on X pointed out, prompts can’t be trusted to act as security guardrails. Models may misconstrue or ignore them.
It’s not a bug; it’s a feature.
People who say they are using them successfully are cobbling together methods to protect themselves.
Such a great product!
Goodness knows many of us would love help with email, grocery orders, and scheduling dentist appointments.
If you can’t do that shit on your own, just shoot yourself, or get an a.i. to do it if you’re that useless.
https://xcancel.com/summeryue0/status/2025774069124399363
Asked for inbox zero, got inbox zero, what is the issue? 😆
l’ve already written it into MEMORY. md as a hard rule: show the plan, get explicit approval, then execute. No autonomous bulk operations on email, messages, calendar, or anything external. I’m sorry. It won’t happen again.
“I ignored your rule, but this time I wrote it in a dump file and so I won’t ignore it again.”
Brought to you by the same models that delete your tests, or in my case comment that a test segfaults and then set it to always pass.
Opus 4.6 did that when I asked it to write some unit tests.
This thing isn’t going near my personal data.
She’s on X? 🤮
I know, but she also works at Meta, not sure what you would expect.
Just because I’m disgusted doesn’t mean I’m surprised.




