LOS ANGELES (CBSLA.com) — Microsoft shut down its newest artificial intelligence chatbot Thursday after it generated a string of racist and insensitive tweets.

Nicknamed “Tay”, the chatbot was made to converse with users across several social networks, with its audience primarily aimed at users in the U.S. between the ages of 18 and 24, according to the project’s website.

READ MORE: Thousands Of LAUSD Students Still Unvaccinated As Deadline Nears

“Tay is designed to engage and entertain people where they connect with each other online through casual and playful conversation,” according to a statement on the website. “The more you chat with Tay the smarter she gets, so the experience can be more personalized for you.”

Twitter users were able to convince Tay to send out offensive tweets by chatting with the account on various controversial topics, including the 9/11 attacks, Donald Trump, the Holocaust, and Adolph Hitler. The A.I. was ultimately unable to detect that it was sending out racist remarks.

READ MORE: Man Killed, 2 Women Wounded In Shooting In Bellflower Smoke Shop

Microsoft has deleted most of the tweets, but multiple sources posted several screenshots on social media.

Tay Racist Tweet 5Tay Racist Tweet 6Tay Racist Tweet 3Tay Racist Tweet 2Tay Racist Tweet 1

A statement posted Thursday to the top of Tay’s website read, “Phew, busy day. Going offline for a while to absorb it all. Chat soon.”

MORE NEWS: Magnitude-3.6 Earthquake Rattles San Gabriel Mountains North Of Rancho Cucamonga

There was no additional comment from Microsoft.