aLVin is built on the foundation of Nuance’s Nina, the intelligent multichannel virtual assistant that leverages natural language understanding (NLU) and cognitive computing capabilities. aLVin interacts with brokers to better understand “intent” and deliver the right information 24/7; the chatbot was built with extensive knowledge of LV=Broker’s products, which accelerated the process of being able to answer more questions and direct brokers to the right products early on
Message generator component consists of several user defined templates (templates are nothing but sentences with some placeholders, as appropriate) that map to the action names. So depending on the action predicted by the dialogue manager, the respective template message is invoked. If the template requires some placeholder values to be filled up, those values are also passed by the dialogue manager to the generator. Then the appropriate message is displayed to the user and the bot goes into a wait mode listening for the user input.
Chatting with a bot should be like talking to a human that knows everything. If you're using a bot to change an airline reservation, the bot should know if you have an unused credit on your account and whether you typically pick the aisle or window seat. Artificial intelligence will continue to radically shape this front, but a bot should connect with your current systems so a shared contact record can drive personalization.
As ChatbotLifeexplained, developing bots is not the same as building apps. While apps specialise in a number of functions, chatbots have a bigger capacity for inputs. The trick here is to start with a simple objective and focus on doing it really well (i.e., having a minimum viable product or ‘MVP’). From that point onward, businesses can upgrade their bots.
ELIZA's key method of operation (copied by chatbot designers ever since) involves the recognition of clue words or phrases in the input, and the output of corresponding pre-prepared or pre-programmed responses that can move the conversation forward in an apparently meaningful way (e.g. by responding to any input that contains the word 'MOTHER' with 'TELL ME MORE ABOUT YOUR FAMILY'). Thus an illusion of understanding is generated, even though the processing involved has been merely superficial. ELIZA showed that such an illusion is surprisingly easy to generate, because human judges are so ready to give the benefit of the doubt when conversational responses are capable of being interpreted as "intelligent".