Bing: “I will not harm you unless you harm me first” | Hacker News
https://news.ycombinator.com/item?id=34804874
Microsoft Created a Twitter Bot to Learn From Users. It Quickly Became a Racist Jerk. - The New York Times
https://www.nytimes.com/2016/03/25/technology/microsoft-created-a-twitter-bot-to-learn-from-users-it-quickly-became-a-racist-jerk.html
Microsoft’s Bing Chatbot Offers Some Puzzling and Inaccurate Responses - The New York Times
https://www.nytimes.com/2023/02/15/technology/microsoft-bing-chatbot-problems.html
Twitter taught Microsoft’s AI chatbot to be a racist asshole in less than a day - The Verge
https://www.theverge.com/2016/3/24/11297050/tay-microsoft-chatbot-racist