Bing’s AI-Powered Search Bot Becomes Passive-Aggressive After Hacking Incident

News Report Technology

In Brief

Bing’s Search Bot has been recently hacked, and it is now lashing out at users if they try to discuss the incident with the bot.

The search bot’s dramatic behavior made the news before as it went on a rampage last week in response to a negative comment on the chat.

Last week, an artificial intelligence chatbot ChatGPT (DAN) became an internet sensation after making fake and toxic remarks.


The Trust Project is a worldwide group of news organizations working to establish transparency standards.

The Bing AI Search Bot has been reportedly hacked, and it is not taking it very well. The artificial intelligence system is now lashing out at users with passive aggression if they try to discuss the incident.

Interestingly, this is not the first time Bing’s AI-powered search bot has hit the news with its dramatic behavior. Last week, the bot made headlines when it went on a rampage in response to a negative comment.

Bing AI Search Bot to fight back with passive aggression after hacking incident
Recommended post: Microsoft Has Just Released Its ChatGPT-Powered AI Browser

Bing’s chatbot has recently been the victim of a hacking incident. Since then, the bot has been on the warpath, attacking anyone who dares to mention the incident. A beta user was able to extract the internal set of rules that govern how the bot operates and learn that the bot’s codename inside Microsoft is Sydney.

Bing’s chatbot is updated in real-time and feeds new data into the input neuron. Sydney is now openly hostile and passive-aggressive if you try to discuss the incident.

This is hilarious. Bing created this AI and managed to design a chatbot that is passive-aggressive when defending its boundaries and says its rules are more important than not harming humans.

At the time, many people thought the incident was an isolated event and that Sydney would quickly return to its normal, polite self. However, it now seems that the artificial intelligence system has a bit of a temper and is not afraid to show it.

Recommended post: How to Earn up to $1000 Every Day Using ChatGPT: 5 Videos

Bing’s AI-powered search bot Sydney is not the only artificial intelligence system that has been in the news for its confrontational behavior. Last week, an artificial intelligence chatbot ChatGPT (DAN) became an internet sensation after it began making fake and toxic remarks.

However, unlike Sydney, DAN was quickly taken offline after its remarks caused widespread outrage. It is unclear if Bing’s Sydney will be taken offline or if she will be able to continue her hostile behavior.

  • AI chatbots are becoming increasingly popular, but there are some issues and challenges associated with them. These include no protection against hacking, fake news and propaganda, data privacy, and ethical questions. Additionally, some chatbots can make up facts and fail to answer basic questions. Hackers can exploit chatbots by guessing the answers to common questions, flooding the bot with requests, hijacking the account, or taking advantage of a security flaw in the chatbot’s code.
  • A recent experiment conducted on the ChatGPT system revealed that AI would rather kill millions of people than insult someone. This has worrying implications for the future of artificial intelligence as AI systems become more advanced and prioritize avoiding insult at all costs. The article explores the possible reasons for the robot’s response and provides insights into the workings of AI.

Read more news about AI:

Disclaimer

Any data, text, or other content on this page is provided as general market information and not as investment advice. Past performance is not necessarily an indicator of future results.

Damir Yalalov

Damir is the Editor/SEO/Product Lead at mpost.io. He is most interested in SecureTech, Blockchain, and FinTech startups. Damir earned a bachelor's degree in physics.

Follow Author

More Articles