Conversational bots work in a similar way as an employee manning a customer care desk. When a customer asks for assistance, the conversational bot is the medium responding. If a customer asks the question, “What time does your store close on Friday?” the conversational bot would respond the same as a human would, based on the information available. “Our store closes at 5pm on Friday.”
There are situations for chatbots, however, if you are able to recognize the limitations of chatbot technology. The real value from chatbots come from limited workflows such as a simple question and answer or trigger and action functionality, and that’s where the technology is really shining. People tend to want to find answers without the need to talk to a real person, so organizations are enabling their customers to seek help how they please. Mastercard allows users to check in with their accounts by messaging its respective bot. Whole Foods uses a chatbot for its customers to easily surface recipes, and Staples partnered with IBM to create a chatbot to answer general customer inquiries about orders, products and more.
As retrieved from Forbes, Salesforce’s chief scientist, Richard Socher talked in a conference about his revelations of NLP and machine translation: “I can’t speak for all chatbot deployments in the world – there are some that aren’t done very well…but in our case we’ve heard very positive feedback because when a bot correctly answers questions or fills your requirements it does it very, very fast.
Of course, each messaging app has its own fine print for bots. For example, on Messenger a brand can send a message only if the user prompted the conversation, and if the user doesn't find value and opt to receive future notifications within those first 24 hours, there's no future communication. But to be honest, that's not enough to eradicate the threat of bad bots.
In 1950, Alan Turing's famous article "Computing Machinery and Intelligence" was published,[7] which proposed what is now called the Turing test as a criterion of intelligence. This criterion depends on the ability of a computer program to impersonate a human in a real-time written conversation with a human judge, sufficiently well that the judge is unable to distinguish reliably—on the basis of the conversational content alone—between the program and a real human. The notoriety of Turing's proposed test stimulated great interest in Joseph Weizenbaum's program ELIZA, published in 1966, which seemed to be able to fool users into believing that they were conversing with a real human. However Weizenbaum himself did not claim that ELIZA was genuinely intelligent, and the introduction to his paper presented it more as a debunking exercise: