Technology in the car keeps drivers informed and connected. Logitech, a global leader in personal computing and accessories, is using voice control and the Alexa Voice Service (AVS) to help drivers stay focused on the road.
The company built ZeroTouch, a product that enables drivers to use their voice to control mobile-based applications on their Android phones or access cloud-based services via Amazon Alexa.
ZeroTouch is a combination of hardware and software that creates a hands-free voice experience in any car. The magnetic hardware car mount holds a phone in place either on the car dash or an air vent. The ZeroTouch mobile application launches via BluetoothⓇ and its voice control feature is activated with the wave of a hand.
Click the link below to read more about the development process.[Read More]
In 2012, brothers Maurice and Marcel Eisterhues built a smartphone app for their father. TorAlarm—German for GoalAlert—had a simple purpose: to help dad keep up with the scores for his favorite football teams. (That’s soccer for readers in the USA.)
What started as a fun project turned into a true opportunity for the two German entrepreneurs. TorAlarm’s popularity grew steadily, until in 2014, the brothers and their father founded a company with the same name. Today, TorAlarm is among Germany’s most popular apps for tracking the scores and schedules of football matches across the country,with over a million users in Germany alone.
Maurice and Marcel knew instantly voice would be the next step in TorAlarm’s evolution when they saw the upcoming launch of Amazon Echo in Germany.
“We were both totally amazed when we first saw the Amazon Echo,” says Maurice. “We’re always interested in new technology, so we decided very quickly we wanted to be part of this launch.”[Read More]
A few months ago we introduced Flask-Ask, a new Python framework for rapid Alexa skill development created by Alexa Champion John Wheeler. Today, due to popular demand, John shares how you can deploy your Alexa skills built with Flask-Ask to AWS Lambda, a service that lets you run code without provisioning or managing servers, which you can use to build serverless applications. Check out John’s technical tutorial below, connect with him on Twitter, and hear more about Flask-Ask in the Alexa Dev Chat podcast episode 10.
In our first post, Flask-Ask and ngrok were used to rapidly create a memory game skill and test it locally. This post shows how to use Flask-Ask with the Zappa framework to quickly deploy skills to AWS Lambda. As of this writing, AWS Lambda supports Python 2.7. This tutorial assumes Python 2.7 is installed on your Windows, Mac, or Linux system.
Zappa, a serverless Python framework, uses a combination of AWS components to emulate the WSGI environment on Lambda that Python web frameworks require. Since Flask-Ask is a Flask extension and Flask requires a WSGI environment, Zappa is the perfect fit for deploying Flask-Ask skills to AWS Lambda. To demonstrate, we'll create an Alexa skill that uses the GitHub API to return how many stars, watchers, and forks a repository has.
Let's get started![Read More]
Andy Huntwork has worked at Amazon for over 10 years, the last three as a principal engineer. He’s developed front-end and backend services for technologies ranging from websites to payment systems and everything in between. But when the Amazon Echo came out in 2015, he saw a new doorway open. Alexa was an exciting way to bring voice-based experiences to the world, and Andy wanted to part of it.
“So I joined the Alexa team,” Andy says, “and immediately started playing around with the Alexa Skills Kit (ASK).” Only a few months after Amazon released ASK, Andy and his wife, Laura, created their first skill. The skill recited public domain works, like Abraham Lincoln’s speeches and The Jungle Book, but the Huntworks wanted to build something more interactive and engaging.
Laura recalls wondering, “What would you ask Alexa to open that’s exciting, even magical? Wouldn’t it be fun to open a magic door?”
From that simple idea grew The Magic Door, an adventure with Alexa guiding you through a growing number of original, interactive stories. Today, The Magic Door skill is a sophisticated adventure framework, hosting 10 adventure storylines, 30,000 spoken words, numerous character voices and hundreds of sounds effects.
To enter a faraway land of magical creatures, perplexing riddles and hidden prizes, just say Alexa, open The Magic Door. Suddenly, you’re off on an adventure with Alexa as your personal guide.[Read More]
When Amazon first introduced the Echo, Nick Schwab was intrigued. He’d always loved voice commands in his car, but he wasn’t sure he wanted to buy another cool device just yet. Then the Echo Dot came out, and once again, Nick couldn’t resist a good deal. He ordered his own Dot, dug into the Alexa Skills Kit (ASK). Right away, he started working on Bargain Buddy, an Alexa skill to relieve him of a daily surf to find daily deals.
Two days after the Bargain Buddy was certified, Nick received his Echo Dot in the mail—his first Alexa device. That’s right, he developed, tested and released his first Alexa skill, before he even had his first Echo Dot.
That was early in 2016. These days, Nick has become a force to be reckoned in the Alexa developer community.[Read More]
Today a speaker is not just a speaker—it’s a connected device we use to play music when we want, where we want, and from whatever service we prefer. New cloud-based, streaming media providers have given us greater control over our music libraries, creating magical and memorable experiences every time we hit “play”.
More than ever, convenience matters, and that’s where Linkplay excels.
Linkplay is a turnkey Wi-Fi audio solution provider that works with device manufacturers to build connected speaker products at various price points and for a growing number of use cases. Whether you want to build a high-end product for the home or a durable portable speaker for camping trips, Linkplay offers customizable product solutions, complete with the latest technologies, to meet your needs.[Read More]
In November 2016 we collaborated with Capital One to accelerate the pace of voice technology innovation with a $10,000 Alexa skill contest for AWS re:Invent attendees. In the contest we challenged attendees to build innovative voice experiences using the Alexa Skills Kit. Individuals or teams of up to four competed to create a unique skill that a customer could use every day. [Read More]
Just Eat has grown a lot since its humble beginnings in a Danish basement in 2001. Now headquartered in London, Just Eat is listed on the London Stock Exchange and is the world’s leading marketplace for online food ordering and delivery. Its goal, simply put, is to revolutionize the way people find, order and enjoy food.
Just Eat is making good on that mission. Today, it connects more than 62,000 restaurants across 100 cuisines in 15 countries, with an audience of over 15 million people.
Craig Pugsley is a principal designer in Just Eat’s Product Research team. He says the UK has a long tradition of delivery and takeout meals. Just Eat’s apps let diners explore exciting new cuisines at nearby restaurants. With menus for over 27,000 restaurants in the UK alone, it’s easy to find a new favorite flavor anytime.
Research quickly showed Pugsley’s team that diners tend to order their favorites again and again. So when Amazon brought Echo and Alexa to the UK, Just Eat saw a new opportunity. The Just Eat Alexa skill would make reordering a tasty new fave even easier, with just a few words:
“Alexa, tell Just Eat to re-order Dim sum.”
No phone calls. No fumbling for a smartphone app. And no digging out credit card details. Just quick delivery of your favorite comfort food.[Read More]
Earlier in the summer, Ashwin Karuhatty reached out to a group of connected home integration professionals in the Custom Electronics Design and Installation Association (CEDIA). Karuhatty, part of Amazon’s Smart Home business development team, wanted to encourage integrators to develop new Alexa skills for the connected home. CEDIA’s annual conference was an ideal place to start.[Read More]
Watchers of Jeopardy!, America’s favorite TV game show, just can’t get enough of the show. For nearly 35 years, Jeopardy! has tested the trivia knowledge of contestants and viewers of all ages. It presents contestants first with answers (clues), then contestants frame their guesses in the form of a question. Home audiences have become so engaged in the TV game’s play, they often shout out the answers to their televisions.
Public fascination with Jeopardy! has led to a long line of off-the-air versions. These started with board games and card games, then electronic versions for game consoles and personal computers, and more recently mobile apps. In all these formats, though, one magical component was always missing. They lacked a way to let players answer out loud (in the form of a question, of course) and have that response validated.
Up until December of last year.
One year ago, on December 30, 2015, Sony Picture Television launched a new version for Amazon Echo users, Jeopardy! J6, built with the Alexa Skills Kit (ASK). On that day, Alexa became the new host of the first all-voice version of Jeopardy!
Sony Television’s games division wanted a way to keep its Jeopardy! fan base growing and engaged, especially younger audiences. To do that, they created a new online version of the game, called Jeopardy! J6, or “J!6” for short.
The premise of J!6 is simple. In addition to the five clues presented for each quiz category on the show, the writers create a backup sixth clue. Most of these clues are never used on the show. With J!6, however, die-hard Jeopardy! fans can play those extra clues. And since the J!6 categories change along with those on the show, playing online feels like an authentic extension of the episode the player just watched.
Geremie Camara, head of the Games Group at Sony Pictures Television, says using high-caliber material from the show makes the J!6 experience authentic and engaging. However, many online quiz games present a multiple-choice list of possible guesses. Somehow, they never quite feel the same as the show…
When Alexa came along, Camara said they’d found the missing piece —a way to present the studio’s high-quality material in an authentic, interactive experience.
The idea for the skill first came from an intern’s summer project. After Amazon shipped two Echo devices to Sony’s R&D group, a intern built a rapid prototype in two short weeks. Though it was full of hard-coded clues and questions, Camara said there was no doubt: they were onto something big.
“We were all blown away at how good the Echo and Alexa technology was,” says Camara, “but also at how good it felt for the brand. We worked very closely with Jeopardy!’s Supervising Producer Rocky Schmidt to ensure that the Echo experience would live up to the TV show’s high standards."
Though an intern did the first pass, a senior engineer built Jeopardy J!6 skill from the ground up. It is written in Java and runs on AWS Lambda, which makes it simple to manage, scalable and very lightweight.[Read More]
EDF Energy is one of the UK’s largest energy companies and its largest producer of low-carbon electricity. It produces around one-fifth of the nation's electricity from its nuclear power stations, wind farms, coal and gas power stations.
Bhavesh Limani is a project manager at Blue Lab, EDF Energy’s innovation accelerator near Brighton in the UK. Launched in 2015, Blue Lab monitors emerging technologies that help shape EDF Energy’s customer experience. One of its primary focus areas is the connected home, including how customers can manage their energy accounts and energy consumption.
When Amazon Echo launched in the United States, it grabbed Blue Lab’s attention. In collaboration with EDF Energy’s R&D UK Centre, the Blue Lab team obtained two Echo units in late 2015. It then began to explore linking voice technology to energy account functionality. Blue Lab wanted to be ready whenever Amazon released Echo and Alexa in the UK.
When Amazon started shipping Echo to UK customers on 28th September, EDF Energy was one of the first UK-specific skills made available to UK customers.
Over the last few years, EDF Energy has worked to give customers more direct access and control of their energy accounts. They first created an online sales and service portal, followed by smartphone apps for iOS and Android users.
“Our customers expect digital solutions now,” says Stuart Roberts, Head of Digital Operations at EDF Energy. “We used Alexa as an opportunity to develop a voice channel to extend the online account management experience to voice.”
As the EDF Energy project team refined their proof of concept, they identified four use cases to meet core customer needs and provide a stand-out experience:
The EDF Energy team established an initial voice user interface (VUI) framework and collaborated with Amazon to refine the VUI. Investing time up front was key to minimizing changes and risks later in development.
“I would say most of our voice interface was well-developed from our first cycle,” says Bhavesh. “The Amazon team was absolutely brilliant in helping us to evaluate the various options.”[Read More]
Today's guest post comes from Jim Kresge from Capital One Engineering.
In March 2016, Capital One became the first company to offer its customers a way to interact with their financial accounts through Alexa devices. With the Capital One skill for Alexa, customers can access in real time all of their Capital One accounts -- from credit cards to bank accounts, to home and auto loans. The skill is highly rated on the Alexa app, with 4/5 stars.
The Capital One team has continued to update the skill since launch, including a recent update to the skill called “How much did I spend?” With the update, Capital One customers can access their recent spending history at more than 2,000 merchants. Customers who have enabled the skill can now ask Alexa about their spending for the past six months--by day, month, or a specific date range--through questions posed in natural language such as:
Q: Alexa, ask Capital One, how much did I spend last weekend?
A: Between December 9th and December 11th, you spent a total of $90.25 on your Venture Card.
Q: Alexa, ask Capital One, how much did I spend at Starbucks last month?
A: Between November 1st and November 30th, you spent a total of $43.00 at Starbucks on your Quicksilver Card.
Q: Alexa, ask Capital One, how much did I spend at Amazon between December 1 and December 15?
A: Between December 1st and December 15th, you spent a total of $463.00 at Amazon on your Quicksilver Card.
The building of the skill was a collaborative effort between product development, engineering and design teams at Capital One. I have the privilege of representing the great work of the entire team in this blog post to give a behind the scenes look at the building of the Capital One skill.
In summer 2015, a group of engineers at Capital One recognized the potential to develop a skill for accessing financial accounts using Amazon Echo. We got together for a hackathon, worked our way through several possibilities, and began building the skill. The Beta version included a server-side account linking mechanism that we built ourselves. We were able to use an enhanced beta version of the Capital One mobile app to provide the account linking interface and created some AWS infrastructure to support it. We then demoed the Beta at the AWS re:Invent conference in October 2015.
Having proved out the Beta version of the skill, we became really driven and focused on building the first skill for Alexa that would enable people to interact with their financial accounts.
We began working on a production version in December, 2015, with the goal of delivering a product by March, 2016. Working in an iterative design model, we found that coding the skill for Capital One financial accounts was relatively straightforward. But, as with anything game-changing, we realized that what we were attempting involved some things no one had done before. First, we were attempting to integrate sensitive data with Alexa, which no company with a skill on Alexa had done yet. It was also the first time we had built a conversational UI. And, the Ask Alexa software was still maturing and evolving as we were building the skill, which meant that we needed to be flexible in quickly making adjustments to code.
We started with the premise that in the first iteration, Capital One credit card and bank customers can ask Alexa things like their current account balance, their recent transactions, and when their next bill is due.
Data security is always top of mind for us, as was creating an experience for customers that was friction-free and simple.
With Amazon, we worked through possible solutions within the Alexa infrastructure to build in a security layer that ensures data integrity while still providing a simple, hands-free experience. In addition to using OAuth to securely link accounts, we added a security solution that involves an in-channel spoken “personal key.” As users set up the Capital One skill and pair their accounts using OAuth, Alexa asks the user if they would like to add a “personal key,” a 4-digit personal identification code.
In addition to wanting users to be able to control access to their account information, we wanted the language Alexa uses in her conversations with customers to be warm and humorous at times. We learned a lot through testing and are using that feedback as we fine tune tone and wording along the way.
We built the Capital One skill using node.js. We also use AWS to host our skill and internal APIs to get customer account information. The basic engineering work is straightforward and the Amazon developer portal documentation makes it easy to learn. Here are a few of the creative technical solutions we added on top of the basic engineering work to help us move fast with high quality:
The Capital One utterance compiler
We created a tool that automatically generates an expansive set of utterances from just a few input parameters. This allows us to avoid maintaining a huge list of individual utterances for our skill. For example, in our "AccountBalance" intent, we have many ways of asking for the balance on an account. To this already long list we then added account types (e.g. checking, savings, etc). After that we added product names (e.g. Venture credit card, Quicksilver credit card). Our list of utterances for that intent is now huge when you incorporate all the different ways customers can ask for their balance across account types and product names. Our utterance compiler makes it simple to generate and maintain all these utterances.[Read More]
On November 18, the first episode of The Grand Tour series marked the most-watched premiere in Amazon’s video streaming service’s history. British car enthusiasts Jeremy Clarkson, Richard Hammond, and James May returned to the screen for an all-new series of globetrotting adventures. Each episode takes Amazon Prime Video viewers to another exotic location.
For Amazon Alexa users, watching The Grand Tour is only half the fun. Prior to the series premiere, Amazon debuted a companion skill built by PullString on the Alexa Store, available to its US and UK customers.
Each Thursday, prior to the show’s Friday airtime, The Grand Tour skill provides a new clue about what to watch for in the upcoming video episode. On Saturday, if viewers are truly “on the tour” and answer three trivia questions correctly, they’ll unlock exclusive video content.
The fun aside, what makes the skill unique is another first: the PullString Platform on which it was developed.
Mike Houlahan, head of PullString’s enterprise partner program, explains Oren Jacob and Martin Reddy co-founded the company in 2011. The two Pixar Animation veterans’ vision was to build lasting emotional connections between characters and audiences using two-way computer conversations. They noted an absence of professional toolsets for building conversational experiences between a character and its audience, and they set about filling that gap.
Now, the company makes the power of the PullString Platform available to Alexa developers. “We are very excited to launch The Grand Tour skill,” Houlahan said. “We are simultaneously announcing the availability of PullString for the Alexa Developer Community to build their own Alexa skills.”
The PullString Platform includes:
With the PullString Platform, a creative writer can prototype, develop, test and deploy an entire skill without writing a single line of code. That’s just what Danielle Frimer did.
Frimer is the creative writer who scripted the voice interaction model (VUI) for The Grand Tour Alexa skill using PullString. She worked with Amazon Prime Video to get the show’s actors into the recording booth to record dialog, and put it all together using the PullString Platform.
“I am not a developer in any way,” says Frimer. “With the platform, I could focus my attention on the creative aspects of it—the lines, the flow of things, the overall design—not on the underlying nuts and bolts of it.”
The skill’s design mimics the flow of The Grand Tour’s episode rollout. The voice interaction, of course, is peppered with the recorded dialog, making the experience even more engaging.
Frimer says PullString’s templates and documentation give developers a quick-start on different types of conversation projects. In all cases, it relieves both authors and developers of the complicated logic involved with a complex VUI model.[Read More]
Tushar Chugh is a graduate student at the Robotics Institute at Carnegie Mellon University (CMU). There he studies the latest in robotics, particularly how computer vision devices perceive the world around them.
One of his favorite projects was a robot named Andy. Besides having arms, Andy could discern colors and understand spatial arrangement. Andy could also respond to voice commands, like “pick up the red block and place it on top of the blue block.” Andy’s speech recognition, a CMU framework, was about to change.
When Amazon came to give some lectures at CMU, they had a raffle drawing. Chugh won the drawing and took home a new Amazon Echo as a prize. Over three days and nights without sleep, he completely integrated Andy and Alexa using the Alexa Skills Kit (ASK).
When he saw Hackster’s 2016 Internet of Voice challenge, he knew he had to enter. And in August 2016, Chugh’s Smart Cap won the prize for the Best Alexa Skills Kit with Raspberry Pi category.
According to Chugh, there are about 285 million visually-impaired people in the world. In 2012, he worked on a project to help the visually impaired navigate inside a building. His device, a belt with embedded sensing tiles, won a couple of prizes, including a Wall Street Journal Technology Innovation Award. It was ahead of its time, though, and it wasn’t yet practical to develop the technology into a commercial product.
A lot can change in four years, including Chugh’s discovery of Alexa. Besides dabbling with Alexa and Andy the robot, he has also worked with Microsoft Cognitive Services for image recognition. Chugh now saw a chance to bring a new and better “seeing device” to light.
“When I saw Alexa, I thought we can extend it and integrate [Alexa] as a separate component,” says Chugh. “I talked with a couple of organizations for the blind in India, and they agreed this kind of system would be very, very useful. That was my main motivation.”
Chugh says the hardware for the Smart Cap is basic. He used a Raspberry Pi (RPi), a battery pack, a camera and a cap on which to mount it. As for the software, it included:
The goal was straightforward. A visually-impaired user could ask Alexa what is in front of them. Alexa would vocalize the scene, allowing the person to navigate safely wherever he or she may be.
How do the pieces all fit together?
Chugh says there are two distinct parts.
First, the image capture and analysis:
Now comes the Alexa skill:
Barely a year later, Mandy attended the Technica Hack ‘15—and won the JP Morgan Chase prize for best mobile app. For Mandy, that event started a new career. After participating alongside so many eager and helpful programmers, she knew what she wanted to do.
Just 11 months later, Mandy has won three more hackathons, including NY TechCrunch ‘16 and Manhattan AngelHack ’16. But unlike her first contest, all these prizes were for Alexa skills.
Mandy first discovered Amazon Echo while attending a 2016 developer conference in San Francisco. She watched a team working with an Echo, and it instantly appealed to her interests in both back-end software development and artificial intelligence. “It was like having my code right in front of me. I talked to my code, and the code kind of talked back to me,” says Mandy.
The day before TechCrunch New York in May, Mandy dove into all the online Alexa Skills Kit documentation she could find. The next day, nervous but determined, she created the prototype that became the Dr. Speech skill—and won the Best Use of Alexa prize.
Mandy, originally from Hong Kong, wanted to help others improve their pronunciation of challenging words, so she created a skill called Dr. Speech, which helps non-native English speakers pronounce words accurately, thereby giving them more confidence without expensive speech therapy sessions.
Mandy gets tweets from people around the world thanking her for how Dr. Speech has been instrumental in improving their pronunciation and has also inspired other developers to build self-improvement skills. Similarly, a user of Mood Journal—another of Mandy’s Alexa skills—wrote to say it helped him in battling anxiety and depression.
Humbled, Mandy repeats she loves to write software that helps people. “Every skill I write is an extension of me. Dr. Speech is about improving speech, because I strive to be a great speaker. I never imagined how my skills would have touched so many people.”[Read More]