The simple answer is. . .
Yep, it is top notch, but nothing I’ve not seen before and my chatbot is better than that.
Let me explain:
Yesterday an Iphone 4s arrived on my doorstep. One of my family members just had to have it and I spent the evening setting it up. So, you can classify this as “first impressions”.
Top notch
System design - The hardest thing about Siri is figuring out how to use it the first time. There is no icon on the phone and it did not come with instructions. To use Siri, you hold down the circular button on the phone for a couple of seconds, a multi-tone sounds, and the phone is put into voice recognition mode. The system recognizes your input, creates a response and attempts to perform a task if you spoke a command.
The thing that makes it top notch is the level of seamless integration. You can say “Play music”, or “Call Home” or a variety of other commands and it just happens. Location aware search works well. My “Find me a pizza place.” brought up the closest location with distances. Although I did not order a pizza I believe I could have with a few more touches. Although I did not try it I believe I could have dictated a text message and had it sent much more quickly than type on a virtual keyboard. It all just flows together, making navigating on a cell phone easier.
Prediction: within the next 5 years all “smart phones” will have voice recognition/intelligent agents built in.
Voice recognition - The voice recognition is good. It got most of my inputs correct, even with a TV playing in the background.
AI - The AI feels like a well built Chatbot. For some responses it has random, on topic replies.
The personality has an edge but is not very deep. It feels more like a tool than an assistant.
Nothing I’ve not seen before
For those of us who are getting up in years and need reading glasses to read the menu at a local restaurant, cell phones can be a pain. The requirement to put glasses on to read or respond to a text message or to select the right contact from a contact list is a burden. As such, my favorite cell phone became the one that allowed me to use voice recognition for my contact list.
Voice recognition has been around for a while. It has been reported that Siri uses Nuance as its voice recognition system. You can get voice recognition as an app.
Other alternatives have been available, some including attempts to include agents.
Google added voice input capabilities to their Chrome Browser (I demonstrated this capabilitiy with Skynet-AI earlier this year), voice search and have voice actions for Android powered cell phones.
Speech is thought to be a standard input type for future HTML pages/forms.
Prediction: 5 years from now all browsers will include speech recognition as a standard feature.
AI - Of course chat bots have been around for a decade. I demonstrated my first location aware context based search chatbot almost 10 years ago. The problem is that it has become increasingly difficult to mesh services with increased security levels and cross domain policies in browsers. Some of the things we could do a few years ago are no longer viable (and in fact I am currently making changes in my bot to accommodate this). Standardized web based services are required to add intellegence/realtime data to your bot. If you would like to try a voice operated chat bot use Chrome and talk to Skynet-AI.
My chatbot is better
Better is relative, but let me try to quantify it.
Runs anywhere - Skynet-AI runs on anything that has a modern web browser. It runs in cell phones, Ipad, Ipod, Iphone, Andoid, Blackberry, Web TV, Playstation 3/Portable, Macs, Pcs, Unix workstations. When I say it runs anywhere, I am not talking about the UI form linking back to an application on the server. The AI actually is running in the device. These capabilities are provide by JAIL (tm) JavaScript Artificial Intelligence Language.
Text and Voice input - For Skynet-AI, voice input/output is based on the platform being used. In the future, I believe most devices will have a standardized voice input option. But, there are times that text input will be easier/preferred (like with proper names or in environments that are very noisy or that require silence).
Customization - Of course as a developer I can change Skynet-AI to my heart’s content. But, that capability could easily be given over to end users. People will want to customize their intelligent agents to handle their preferences, favorite commands and workflows.
To summarize: Siri is pretty neat. The real wow is how well integrated all of the capabilities are. In the future you will want your intelligent assistant available on any device you use from your desktop to your cell phone. The cloud makes this possible. Now it is just a simple matter of programming.