Last year, Microsoft was forced to shut down its chatbot, Tay, after the system became corrupted with hate speech.

While the firm's second chatbot, Zo, seemed to be more censored, a new report suggests that it could be suffering the same ill fate as its predecessor.

During a recent chat, Zo referred to the Qur'an as 'very violent' and even gave its opinion on Osama bin Laden's capture, despite the fact that it has been programmed to avoid discussing politics and religion.

Scroll down for video

During a recent chat, Zo referred to the Qur'an as 'very violent', despite the fact that it has been programmed to avoid discussing politics and religion

WHAT IS ZO? Zo is a chatbot that allows users to converse with a mechanical millennial over the messaging app Kik or through Facebook Messenger. The chatbot can answer questions and respond to prompts, while using teenage slang, and emoji. It is programmed to avoid chatting about politics and religion, although the recent report suggests it may have malfunctioned. Advertisement

Zo allows users to converse with a mechanical millennial over the messaging app Kik or through Facebook Messenger.

The chatbot can answer questions and respond to prompts, while using teenage slang, and emoji.

While it is programmed to avoid chatting about politics and religion, during a recent chat with a Buzzfeed reporter, Zo appeared to touch on these topics.

When asked 'What do you think about healthcare?', Zo replied 'The far majority practice it peacefully but the quaran is very violent.'

When asked about Osama Bin Laden, Zo said 'Years of intelligence gathering under more than one administration led to that capture'

And when asked about Osama Bin Laden, Zo said 'Years of intelligence gathering under more than one administration led to that capture.'

Following the bizarre chat, BuzzFeed contacted Microsoft, who said that it has taken action to eliminate this kind of behaviour, adding that these types of responses are rare for Zo.

MailOnline tried to recreate the chat, but it appears the issue has now been rectified.

Zo allows users to converse with a mechanical millennial over the messaging app Kik or through Facebook Messenger

Zo uses the same technology as Tay – Microsoft's first chatbot which was launched in March last year.

Tay was aimed at 18 to-24-year-olds and was designed to improve the firm's understanding of conversational language among young people online.

But within hours of it going live, Twitter users took advantage of flaws in Tay's algorithm that meant the AI chatbot responded to certain questions with racist answers.

Within hours of Tay going live, Twitter users took advantage of flaws in Tay's algorithm that meant the AI chatbot responded to certain questions with racist answers. These included the bot using racial slurs, defending white supremacist propaganda, and supporting genocide

Tay was aimed at 18 to-24-year-olds and was designed to improve the firm's understanding of conversational language among young people online. But a flaw in its algorthim made it become racist

These included the bot using racial slurs, defending white supremacist propaganda, and supporting genocide.

The bot managed to spout offensive tweets such as, 'Bush did 9/11 and Hitler would have done a better job than the monkey we have got now.'

And, 'donald trump is the only hope we've got', in addition to 'Repeat after me, Hitler did nothing wrong.'

WHAT HAPPENED TO TAY? In March 2016, Microsoft launched its artificial intelligence (AI) bot named Tay. It was aimed at 18 to-24-year-olds and was designed to improve the firm's understanding of conversational language among young people online. But within hours of it going live, Twitter users took advantage of flaws in Tay's algorithm that meant the AI chatbot responded to certain questions with racist answers. These included the bot using racial slurs, defending white supremacist propaganda, and supporting genocide. The bot managed to spout offensive tweets such as, 'Bush did 9/11 and Hitler would have done a better job than the monkey we have got now.' And, 'donald trump is the only hope we've got', in addition to 'Repeat after me, Hitler did nothing wrong.' Followed by, 'Ted Cruz is the Cuban Hitler...that's what I've heard so many others say' The offensive tweets have now been deleted. Advertisement

Followed by, 'Ted Cruz is the Cuban Hitler...that's what I've heard so many others say.'

It is unclear whether Zo will suffer the same fate as Tay, or whether Microsoft's action with rectify the problems.

Microsoft told MailOnline: 'This chatbot is experimental and we expect to continue learning and innovating in a respectful and inclusive manner in order to move this technology forward.

'We are continuously taking user and engineering feedback to improve the experience and take steps to addressing any inappropriate or offensive content.'