Easy-to-use platform even for non-techies
Understand what your customers say in any language
Connect to your digital ecosystem with free Cognigy.AI Extensions
Integrate with any CRMs, ERPs, RPA, and more
Enhance customer interactions with AI on the frontline
Connect customers to the right agent on the first try
Give your agents the confidence to solve complex requests
Boost your customers’ experience and agents’ efficiency
Improve customer satisfaction with detailed insights
Engage with your customers on the platform of their choice
Drive appropriate actions with intelligent processes
Integrate easily with existing enterprise infrastructure
Increase operational efficiency and reduce agent attrition
Dive deep into the world of conversational AI
Learn everything about voice bots and virtual agents
Become an expert in Conversational AI and automation
See Conversational AI in action through demos
Explore Cognigy.AI and get support from the community
Join our mission to make Conversational AI more accessible
Find us at live conferences or virtual meetups
Help us achieve our vision with your talents and skills
Be the first to know about all the latest news
Reach out to our experts and get your questions answered
See how we help your organization secure sensitive data and comply with applicable laws and regulations.
An in-depth guide into the trusted use of AI in customer service automation
Find out everything you need to know about establishing "Explainable AI" based on the AIC4 criteria catalogue.
Amazon recently introduced APL (Alexa Presentation Language). APL enables developers to build interactive and multimodal Skills that support screens. Thanks to the generic Alexa directive support built in to Cognigy, you can already leverage this new capability.
All you need to do is:
Enabled APL Interface in the Alexa developer console
To activate this feature, go to the Alexa developer console, select your Skill and enable the Alexa Presentation Language in the Interfaces section.
Amazon’s APL authoring tool
There you can create visuals (using a sample template or starting from scratch), export the final document, and copy its “document” and “datasources” properties for your Alexa response. The JSON structure for the APL RenderDocument Skill directive also requires a type and a custom token.
The JSON looks roughly like this:
Alexa directive in a Cognigy Say Node
The last step is to create a Say-Node in Cognigy, select the Alexa tab, and add the JSON document as a new directive.
You also can send APL commands to already-rendered APL documents using the Alexa.Presentation.APL.ExecuteCommands directive. To improve this further, you can distinguish between devices with and without APL support by checking for the Alexa.Presentation.APL property in your conversation flow (ci.data.context.System.device.supportedInterfaces).
You can also handle user events which are sent to your Skill. To do this, check the Alexa.Presentation.APL.UserEvent request type which can be found under ci.data.request.type.
See our Alexa quick start video and start your Alexa Skill now using the free Community Edition. We can't wait to see what you will build.
Sascha Wolter is a professional developer and user experience enthusiast. His true passion is to improve the human-computer interaction:
He loves to build conversational and multimodal experiences with text (chatbots) and voice (aka Alexa) and, as such, he works as Senior UX Consultant and Principal Technology Evangelist at Cognigy.
Stay up to date with the latest Conversational AI tips and news.