APL for Audio Reference
Alexa Presentation Language (APL) for Audio lets you create rich audio experiences for use within your custom skills. Use the
APLA document type to define these experiences.
Use the following links to access the reference material relevant to APL for audio.
APL for audio uses the APLA Document format for sending an audio response to the device. The APLA document defines and arranges a set of audio clips. You build these audio clips from text-to-speech and audio files using APL components. All Alexa devices can play the audio response defined by these clips.
Data sources and data-binding
APL for audio uses data-binding to incorporate user-provided data, include audio resources, and conditionally inflate components based on data. To use data-binding in your document, you write data-binding expressions, which are JSON strings with an embedded
See the following topics for details about data sources and data-binding for APL for audio:
APL for audio uses APL components to define audio clips. For example, a
Speech component defines text to convert to speech. The
Audio component identifies an audio file to play. Other components let you mix and arrange the speech and audio in different ways.
All components have the same Base Component Properties.
Components that can render multiple child components also have the common multi-child component properties: Multiple Child Component Properties.
APL for audio supports the following primitive components:
You can use Filters to modify an audio component. For more details, see Filters.
Skill request and response reference
APL for audio supports directives in the
Use the authoring tool for APL for audio
The developer console provides an authoring tool to build and preview APL documents. You can write your document and data sources in JSON, and then play the results, and then copy the JSON code into your skill.
For details about using the authoring tool, see Build an audio response.
Last updated: Jan 03, 2022