Bing: “I will not harm you unless you harm me first”

Last week, Microsoft announced the new AI-powered Bing: a search interface that incorporates a language model powered chatbot that can run searches for you and summarize the results, plus do all of the other fun things that engines like GPT-3 and ChatGPT have been demonstrating over the past few months: the ability to generate poetry, and jokes, and do creative writing, and so much more.

This week, people have started gaining access to it via the waiting list. It’s increasingly looking like this may be one of the most hilariously inappropriate applications of AI that we’ve seen yet.

If you haven’t been paying attention, here’s what’s transpired so far.

  • The demo was full of errors
  • It started gaslighting people
  • It suffered an existential crisis
  • The prompt leaked
  • And then it started threatening people