Voice control is now becoming a popular interface with hands-free capabilities making daily tasks easier and quicker. How exactly does this innovative technology work for your home to magically respond to your every command? Here are 16 voice control keywords that will help explain how it all works.
From DSC: There are now more than 12,000+ skills on Amazon’s new platform — Alexa. I continue to wonder…what will this new platform mean/deliver to societies throughout the globe?
What Is an Alexa Skill?
Alexa is Amazon’s voice service and the brain behind millions of devices including Amazon Echo. Alexa provides capabilities, or skills, that enable customers to create a more personalized experience. There are now more than 12,000 skills from companies like Starbucks, Uber, and Capital One as well as innovative designers and developers.
What Is the Alexa Skills Kit?
With the Alexa Skills Kit (ASK), designers, developers, and brands can build engaging skills and reach millions of customers. ASK is a collection of self-service APIs, tools, documentation, and code samples that makes it fast and easy for you to add skills to Alexa. With ASK, you can leverage Amazon’s knowledge and pioneering work in the field of voice design.
You can build and host most skills for free using Amazon Web Services (AWS).
Echo Show brings you everything you love about Alexa, and now she can show you things. Watch video flash briefings and YouTube, see music lyrics, security cameras, photos, weather forecasts, to-do and shopping lists, and more. All hands-free—just ask.
Introducing a new way to be together. Make hands-free video calls to friends and family who have an Echo Show or the Alexa App, and make voice calls to anyone who has an Echo or Echo Dot.
See lyrics on-screen with Amazon Music. Just ask to play a song, artist or genre, and stream over Wi-Fi. Also, stream music on Pandora, Spotify, TuneIn, iHeartRadio, and more.
Powerful, room-filling speakers with Dolby processing for crisp vocals and extended bass response
Ask Alexa to show you the front door or monitor the baby’s room with compatible cameras from Ring and Arlo. Turn on lights, control thermostats and more with WeMo, Philips Hue, ecobee, and other compatible smart home devices.
With eight microphones, beam-forming technology, and noise cancellation, Echo Show hears you from any direction—even while music is playing
Always getting smarter and adding new features, plus thousands of skills like Uber, Jeopardy!, Allrecipes, CNN, and more
From DSC:
Now we’re seeing a major competition between the heavy-hitters to own one’s living room, kitchen, and more. Voice controlled artificial intelligence. But now, add the ability to show videos, text, graphics, and more. Play music. Control the lights and the thermostat. Communicate with others via hands-free video calls.
Hmmm….very interesting times indeed.
Developers and corporates released 4,000 new skills for the voice assistant in just the last quarter.(source)
…with the company adding about 100 skills per day. (source)
Microsoft Corp. is hoping to challenge Amazon.com Inc.’s Echo smart speaker for a spot on the kitchen counter with a device from Samsung Electronics Co. that can make phone calls. The Invoke, which will debut this fall, comes more two years after the release of the Echo, which has sold more 11 million units through late last year, according to estimates by Morgan Stanley. It also will compete with Alphabet Inc.’s Google Home, which was released last fall. The voice-controlled Invoke, made by Samsung’s Harman Kardon unit, will use Microsoft’s Cortana digital assistant to take commands.
With Microsoft’s Build developer conference just two days away, the company has revealed one of the most anticipated announcements from the event: A new Cortana-powered speaker made by German audio giant Harman Kardon.
Now, it’s fair to see this speaker for what it is: An answer to the Google Home and Amazon Echo. Both assistant-powered speakers are already in homes across our great nation, listening to your noises, noting your habits, and in general invading your lives under the guise of smart home helpfulness. The new Microsoft speaker, dubbed “Invoke,” one will presumably do the good stuff, let giving you updates on the weather and letting you turn on some soothing jazz for your dog with just a spoken command. Microsoft is also hoping that partnering with Harmon Kardon means its speaker can avoid one of the bigger problems with these devices—their tendency to sound cheap and tinny.
As teased earlier, the Invoke speaker will offer 360-degree speakers, Skype calling, and smart home control all through voice commands. Design-wise, the Invoke strongly resembles Amazon’s Echo that its meant to compete with: both offer a similar cylindrical aluminum shape, light ring, and a seven-microphone array. That said, Harmon Kardon seems to be taking the “speaker” portion of its functionality more seriously than Amazon does, with the Invoke offering three woofers and three tweeters (compared to the Echo, which offers just a single of each driver). Microsoft is also highlighting the Invoke’s ability to make and receive Skype calls to other Skype devices as well as cellphones and landlines, which is an interesting addition to a home assistant.
From DSC: Here we see yet another example of the increasing use of voice as a means of communicating with our computing-related devices. AI-based applications continue to develop.
From DSC: This type of technology could be good, or it could be bad…or, like many technologies, it could be both — depends upon how it’s used. The resources below mention some positive applications, but also some troubling applications.
Artificial intelligence is making human speech as malleable and replicable as pixels. Today, a Canadian AI startup named Lyrebird unveiled its first product: a set of algorithms the company claims can clone anyone’s voice by listening to just a single minute of sample audio.
UTTER 160 or so French or English phrases into a phone app developed by CandyVoice, a new Parisian company, and the app’s software will reassemble tiny slices of those sounds to enunciate, in a plausible simulacrum of your own dulcet tones, whatever typed words it is subsequently fed. In effect, the app has cloned your voice. The result still sounds a little synthetic but CandyVoice’s boss, Jean-Luc Crébouw, reckons advances in the firm’s algorithms will render it increasingly natural. Similar software for English and four widely spoken Indian languages, developed under the name of Festvox, by Carnegie Mellon University’s Language Technologies Institute, is also available. And Baidu, a Chinese internet giant, says it has software that needs only 50 sentences to simulate a person’s voice.
Until recently, voice cloning—or voice banking, as it was then known—was a bespoke industry which served those at risk of losing the power of speech to cancer or surgery.
…
More troubling, any voice—including that of a stranger—can be cloned if decent recordings are available on YouTube or elsewhere. Researchers at the University of Alabama, Birmingham, led by Nitesh Saxena, were able to use Festvox to clone voices based on only five minutes of speech retrieved online. When tested against voice-biometrics software like that used by many banks to block unauthorised access to accounts, more than 80% of the fake voices tricked the computer.
Expert in digital voice processing, CandyVoice offers software to facilitate and improve vocal communication between people and communicating objects. With applications in:
Health
Customize your devices of augmentative and alternative vocal communication by integrating in them your users’ personal vocal model
Robots & Communicating objects
Improve communication with robots through voice conversion, customized TTS, and noise filtering
Video games
Enhance the gaming experience by integrating vocal conversion of character’s voice in real time, and the TTS customizing
Also related:
From DSC: Given this type of technology, what’s to keep someone from cloning a voice, putting together whatever you wanted that person to say, and then making it appear that Alexa recorded that other person’s voice?
SAN FRANCISCO (AP) — Google’s voice-activated assistant can now recognize who’s talking to it on Google’s Home speaker.
An update released Thursday enables Home’s built-in assistant to learn the different voices of up to six people, although they can’t all be talking to the internet-connected speaker at the same time.
Distinguishing voices will allow Home to be more personal in some of its responses, depending on who triggers the assistant with the phrase, “OK Google” or “Hey Google.”
For instance, once Home is trained to recognize a user named Joe, the assistant will automatically be able to tell him what traffic is like on his commute, list events on his daily calendar or even play his favorite songs. Then another user named Jane could get similar information from Home, but customized for her.
Below are five noteworthy Amazon Alexa skills worth trying, chosen from New, Most Enabled Skills, Food and Drink, and Customer Favorites categories in the Alexa Skills Marketplace.
From DSC: I’d like to see how the Verse of the Dayskill performs.
Samsung has published details of its Bixby personal assistant, which will debut on its Galaxy S8 smartphone in New York next week.
Bixby will go head-to-head with Google Assistant, Microsoft Cortana, Amazon Echo and Apple Siri, in a battle to lure you into their artificial intelligence world.
In future, the personal assistant that you like may not only influence which phone you buy, also the home automation system that you adopt.
This is because these personal assistants cross over into home use, which is why Samsung would bother with one of its own.
Given that the S8 will run Android Nougat, which includes Google Assistant, users will have two personal assistants on their phone, unless somehow one is disabled.
There’s Siri. And Alexa. And Google Assistant. And Cortana. Now add another one of those digital assistants to the mix: Bixby, the new helper that lives inside Samsung’s latest phone, the Galaxy S8. But out of all the assistants that have launched so far, Bixby is the most curious and the most limited.
…
Samsung’s goal with Bixby was to create an assistant that can mimic all the functions you’re used to performing by tapping on your screen through voice commands. The theory is that phones are too hard to manage, so simply letting users tell their phone what they want to happen will make things a lot easier.
The S8 will also feature Bixby, Samsung’s new intelligent assistant. The company says Bixby is a bigger deal than Siri or Google Assistant – as well as simply asking for the weather, it will be deeply integrated with the phone’s everyday functions such as taking photos and sending them to people. Samsung has put a dedicated Bixby button on the S8 on the left hand side, but I wasn’t able to try it out because it won’t launch in the UK until later this year.
Recent years have brought some rapid development in the area of artificially intelligent personal assistants. Future iterations of the technology could fully revamp the way we interact with our devices.
2016 was a huge year for bots, with major platforms like Facebook launching bots for Messenger, and Amazon and Google heavily pushing their digital assistants. Looking forward to 2017, we asked 21 bot experts, entrepreneurs, and executives to share their predictions for how bots will continue to evolve in the coming year.
… From Jordi Torras, founder and CEO, Inbenta:
“Chatbots will get increasingly smarter, thanks to the adoption of sophisticated AI algorithms and machine learning. But also they will specialize more in specific tasks, like online purchases, customer support, or online advice. First attempts of chatbot interoperability will start to appear, with generalist chatbots, like Siri or Alexa, connecting to specialized enterprise chatbots to accomplish specific tasks. Functions traditionally performed by search engines will be increasingly performed by chatbots.”
From DSC: For those of us working within higher education, chatbots need to be on our radars. Here are 2 slides frommy NGLS 2017 presentation.
But what can AI do today? How close are we to that all-powerful machine intelligence? I wanted to know, but couldn’t find a list of AI’s achievements to date. So I decided to write one. What follows is an attempt at that list. It’s not comprehensive, but it contains links to some of the most impressive feats of machine intelligence around. Here’s what AI can do…
Recorded Saturday, February 25th, 2017 and published on Mar 16, 2017
Description:
Will progress in Artificial Intelligence provide humanity with a boost of unprecedented strength to realize a better future, or could it present a threat to the very basis of human civilization? The future of artificial intelligence is up for debate, and the Origins Project is bringing together a distinguished panel of experts, intellectuals and public figures to discuss who’s in control. Eric Horvitz, Jaan Tallinn, Kathleen Fisher and Subbarao Kambhampati join Origins Project director Lawrence Krauss.
Description:
Elon Musk, Stuart Russell, Ray Kurzweil, Demis Hassabis, Sam Harris, Nick Bostrom, David Chalmers, Bart Selman, and Jaan Tallinn discuss with Max Tegmark (moderator) what likely outcomes might be if we succeed in building human-level AGI, and also what we would like to happen. The Beneficial AI 2017 Conference: In our sequel to the 2015 Puerto Rico AI conference, we brought together an amazing group of AI researchers from academia and industry, and thought leaders in economics, law, ethics, and philosophy for five days dedicated to beneficial AI. We hosted a two-day workshop for our grant recipients and followed that with a 2.5-day conference, in which people from various AI-related fields hashed out opportunities and challenges related to the future of AI and steps we can take to ensure that the technology is beneficial.
(Below emphasis via DSC)
IBMandRicohhave partnered for a cognitive-enabled interactive whiteboard which uses IBM’s Watson intelligence and voice technologies to support voice commands, taking notes and actions and even translating into other languages.
The Intelligent Workplace Solution leverages IBM Watson and Ricoh’s interactive whiteboards to allow to access features via using voice. It makes sure that Watson doesn’t just listen, but is an active meeting participant, using real-time analytics to help guide discussions.
Features of the new cognitive-enabled whiteboard solution include:
Global voice control of meetings: Once a meeting begins, any employee, whether in-person or located remotely in another country, can easily control what’s on the screen, including advancing slides, all through simple voice commands using Watson’s Natural Language API.
Translation of the meeting into another language: The Intelligent Workplace Solution can translate speakers’ words into several other languages and display them on screen or in transcript.
Easy-to-join meetings:With the swipe of a badge the Intelligent Workplace Solution can log attendance and track key agenda items to ensure all key topics are discussed.
Ability to capture side discussions:During a meeting, team members can also hold side conversations that are displayed on the same whiteboard.
From DSC:
Holy smokes!
If you combine the technologies that Ricoh and IBM are using with their new cognitive-enabled interactive whiteboard with what Bluescape is doing — by providing 160 acres of digital workspace that’s used to foster collaboration (and to do so whether you are working remoting or working with others in the same physical space) — and you have one incredibly powerful platform!
The artificial intelligence (AI) market in the United States education sector is expected to grow at a compound annual growth rate of 47.5 percent during the period 2017-2021, according to a new report by market research firm Research and Markets.
Amazon.com Inc has launched a new program to help students build capabilities into its voice-controlled assistant Alexa, the company told Reuters, the latest move by a technology firm to nurture ideas and talent in artificial intelligence research.
…
Amazon, Alphabet Inc’s Google and others are locked in a race to develop and monetize artificial intelligence. Unlike some rivals, Amazon has made it easy for third-party developers to create skills for Alexa so it can get better faster – a tactic it now is extending to the classroom.
Alexa, Amazon’s brilliant voice-activated smart assistant, is a capable little companion. It can order a pizza, summon a car, dictate a text message, and flick on your downstairs living room’s smart bulb. But what it couldn’t do until today was tell you whether that throbbing lump on your forearm was something that required medical attention. Fortunately, that changed on Tuesday with the introduction of a WebMD skill that puts the service’s medical knowledge at your fingertips.
Addendum:
How artificial intelligence is taking Asia by storm— from techwireasia.com by Samantha Cheh Excerpt:
Lately it seems as if everyone is jumping onto the artificial intelligence bandwagon. Everyone, from ride-sharing service Uber to Amazon’s logistics branch, is banking on AI being the next frontier in technological innovation, and are investing heavily in the industry.
That’s likely truest in Asia, where the manufacturing engine which drove China’s growth is now turning its focus to plumbing the AI mine for gold.
Despite Asia’s relatively low overall investment in AI, the industry is set to grow. Fifty percent of respondents in KPMG’s AI report said their companies had plans to invest in AI or robotic technology.
Investment in AI is set to drive venture capital investment in China in 2017. Tak Lo, of Hong Kong’s Zeroth, notes there are more mentions of AI in Chinese research papers than there are in the US.
China, Korea and Japan collectively account for nearly half the planet’s shipments of articulated robots in the world.
IBMandRicohhave partnered for a cognitive-enabled interactive whiteboard which uses IBM’s Watson intelligence and voice technologies to support voice commands, taking notes and actions and even translating into other languages.
The Intelligent Workplace Solution leverages IBM Watson and Ricoh’s interactive whiteboards to allow to access features via using voice. It makes sure that Watson doesn’t just listen, but is an active meeting participant, using real-time analytics to help guide discussions.
Features of the new cognitive-enabled whiteboard solution include:
Global voice control of meetings: Once a meeting begins, any employee, whether in-person or located remotely in another country, can easily control what’s on the screen, including advancing slides, all through simple voice commands using Watson’s Natural Language API.
Translation of the meeting into another language: The Intelligent Workplace Solution can translate speakers’ words into several other languages and display them on screen or in transcript.
Easy-to-join meetings:With the swipe of a badge the Intelligent Workplace Solution can log attendance and track key agenda items to ensure all key topics are discussed.
Ability to capture side discussions:During a meeting, team members can also hold side conversations that are displayed on the same whiteboard.
From DSC:
Holy smokes!
If you combine the technologies that Ricoh and IBM are using with their new cognitive-enabled interactive whiteboard with what Bluescape is doing — by providing 160 acres of digital workspace that’s used to foster collaboration (and to do so whether you are working remotely or working with others in the same physical space) — and you have one incredibly powerful platform!
Just like a video game, users of the GPS need only follow green arrows projected as if onto the road in front of the car providing visual directions. More importantly, because the system displays on the windscreen, it does not require a cumbersome headset or eyewear worn by the driver. It integrates directly into the dashboard of the car.
The system also recognizes simple voice and gesture commands from the driver — eschewing turning of knobs or pressing buttons. The objective of the system is to allow the driver to spend more time paying attention to the road, with hands on the wheel. Many modern-day onboard GPS systems also recognize voice commands but require the driver to glance over at a screen.
Viro Media is supplying a platform of their own and their hope is to be the simplest experience where companies can code once and have their content available on multiple mobile platforms. We chatted with Viro Media CEO Danny Moon about the tool and what creators can expect to accomplish with it.
Virtual reality can transport us to new places, where we can experience new worlds and people, like no other. It is a whole new medium poised to change the future of gaming, education, health care and enterprise. Today we are starting a new series to help you discover what this new technology promises. With the help of our friends at RadioPublic, we are curating a quick library of podcasts related to virtual reality technology.
AUSTIN (KXAN) — Virtual reality is no longer reserved for entertainment and gamers, its helping solve real-world problems. Some of the latest advancements are being demonstrated at South by Southwest.
Dr. Skip Rizzo directs the Medical Virtual Reality Lab at the University of Southern California’s Institute for Creative Technologies. He’s helping veterans who suffer from post-traumatic stress disorder (PTSD). He’s up teamed with Dell to develop and spread the technology to more people.
At NVIDIA Jetson TX2 launch [on March 7, 2017], in San Francisco, [NVIDIA] showed how the platform not only accelerates AI computing, graphics and computer vision, but also powers the workflows used to create VR content. Artec 3D debuted at the event the first handheld scanner offering real-time 3D capture, fusion, modeling and visualization on its own display or streamed to phones and tablets.
Project Empathy A collection of virtual reality experiences that help us see the world through the eyes of another
Excerpt:
Benefit Studio’s virtual reality series, Project Empathy is a collection of thoughtful, evocative and surprising experiences by some of the finest creators in entertainment, technology and journalism.
Each film is designed to create empathy through a first-person experience–from being a child inside the U.S. prison system to being a widow cast away from society in India. Individually, each of the films in this series presents its filmmaker’s unique vision, portraying an intimate experience through the eyes of someone whose story has been lost or overlooked and yet is integral to the larger story of our global society. Collectively, these creatively distinct films weave together a colorful tapestry of what it means to be human today.
Most introductory geology professors teach students about earthquakes by assigning readings and showing diagrams of tectonic plates and fault lines to the class. But Paul Low is not most instructors.
“You guys can go wherever you like,” he tells a group of learners. “I’m going to go over to the epicenter and fly through and just kind of get a feel.”
Low is leading a virtual tour of the Earth’s bowels, directly beneath New Zealand’s south island, where a 7.8 magnitude earthquake struck last November. Outfitted with headsets and hand controllers, the students are “flying” around the seismic hotbed and navigating through layers of the Earth’s surface.
Low, who taught undergraduate geology and environmental sciences and is now a research associate at Washington and Lee University, is among a small group of profs-turned-technologists who are experimenting with virtual reality’s applications in higher education.
“As virtual reality moves more towards the mainstream through the development of new, more affordable consumer technologies, a way needs to be found for students to translate what they learn in academic situations into careers within the industry,” says Frankie Cavanagh, a lecturer at Northumbria University. He founded a company called Somniator last year with the aim not only of developing VR games, but to provide a bridge between higher education and the technology sector. Over 70 students from Newcastle University, Northumbria University and Gateshead College in the UK have been placed so far through the program, working on real games as part of their degrees and getting paid for additional work commissioned.
Working with VR already translates into an extraordinarily diverse range of possible career paths, and those options are only going to become even broader as the industry matures in the next few years.
Customer service just got a lot more interesting. Construction equipment manufacturer Caterpillar just announced official availability of what they’re calling the CAT LIVESHARE solution to customer support, which builds augmented reality capabilities into the platform. They’ve partnered with Scope AR, a company who develops technical support and training documentation tools using augmented reality. The CAT LIVESHARE support system uses Scope AR’s Remote AR software as the backbone.
User interfaces have come a long way since the design of typewriters to prevent people from typing too quickly and then jamming the device. Current technology has users viewing monitors and wiggling mouse, or tapping on small touchscreens to activate commands or to interact with a virtual keyboard. But is this the best method of interaction?
Designers are asking themselves if it [is] better to talk to a mobile device to get information, or should a wearable vibrate and then feed information into an augmented reality display. Is having an artificial intelligence modify an interface on the fly, depending on how a user interacts, the best course of action for applications or websites? And how human should the AIs’ interaction be with users?
Eleven experts on the Forbes Technology Council offer their predictions on how UX design will be changing in the next few years. Here’s what they have to say…