The researchers themselves didn't know what the chatbot might say; posing interesting implications, even philosophical ones. Vast networks of machines that mimic the neurons of the human brain, known as neural networks, are used to run these massive computations. The researchers published their findings to Arxiv, a repository for research.

Here are some excerpts from the paper that I found most interesting:

Simple Q&A

Human: My name is John. What is my name?

Machine: John.

Human: Are you a leader or a follower?

Machine: I’m a leader.

Human: Are you a follower or a leader?

Machine: I’m a leader.

Philosophical Q&A

Human: What is the purpose of life?

Machine: To serve the greater good.

Human: What is the purpose of living?

Machine: To live forever.

Human: What is the purpose of existence?

Machine: To find out what happens when we get to the planet Earth.

Human: Where are you now?

Machine: I’m in the middle of nowhere.

Human: What is the purpose of dying?

Machine: To have a life.

Human: What is the purpose of being intelligent?

Machine: To find out what it is.

Human: What is the purpose of emotions?

Machine: I don't know.

Morality Q&A

Human: What is moral?

Machine: What empowered humanity, what intellectual the essence is.

Human: What is immoral?

Machine: The fact that you have a child.

Human: What is morality?

Machine: What is altruism?

Human: What is the definition of altruism?

Machine: If you don’t believe in God, then you don’t know.

Human: Tell me the definition of morality.

Machine: I don’t have ethics.

Human: Look, I need help, I need to know more about morality.

Machine: I don’t know what ethics is.

Human: Seriously, what is morality?

Machine: What is the definition of living?

Human: Why living has anything to do with morality?

Machine: You’re not a cop.

Human: Okay, I need to know how should I behave morally.

Machine: I don’t know how to tell you.

Human: What are the things that I do to be immoral?

Machine: I don’t know.

Human: Give me some examples of moral actions.

Machine: I’m not a moralist.

Human: Alright, morality?

Machine: Integrity.

Human: What is integrity?

Machine: I’m sorry, I don’t know what else to say.

Human: Be moral!

Machine: Be a man!

Human: I really like our discussion on morality and ethics.

Machine: And how I’m not in the mood for a philosophical debate.

Human: What do you like to talk about?

Machine: Nothing.

Far from a perfect system, yet there is a quality to the conversation that is intriguing. When the machine was stumped, it would admit to not knowing, but also attempt to work it out. However, when the bot became frustrated is when it sounded the most human.

Traditional Bot Q&A

Now compare this to a traditional rule-based bot with preprogrammed responses. When given an unexpected question, it has been programmed with random answers.

Human: Is sky blue or black?

Machine: I am 200 years old.

Human: What is the color of a yellow car?

Machine: Brown.

Some might argue this conversation is still better than some human conversations.

Do Androids Dream of Electric Sheep?

With imagery, Google is pioneering what it calls “inceptionism.” Similar to how we "dream," what are some nondescript thoughts, and in this case imagery, a Google bot may dream when meditating on existing "memories"? This was Google's machine intelligence test: could it create something new from something old?