Tag Archives: Voice First

User Interface Design is the new black!

14 Dec

capture

 

LinkedIn Unveils The Top Skills That Can Get You Hired In 2017

 

Number 5: USER INTERFACE DESIGN!

“User interface design is the new black:

UI Design, which is designing the part of products that people interact with, is increasingly in-demand among employers. It ranked #14 in 2014, #10 last year, and #5 this year (second largest jump on this year’s Global Top Skills of 2016 list). Data has become central to many products, which has created a need for people with user interface design skills who can make those products easy for customers to use.”

 

capture

Read all about it here.

https://blog.linkedin.com/2016/10/20/top-skills-2016-week-of-learning-linkedin

Amazon Alexa Developers Conference (10 Oct, London)

7 Oct

Exciting times for Amazon Alexa!

capture

Amazon Echo

  • Alexa has crossed over 3,000 skills in the US Skill Store, up from 1,000 in June;
  • Alexa has just started shipping in the UK and is coming soon to Germany too. Echo and Echo Dot were just made available in the UK, whereas in Germany they are available by invitation for those who want to help shape Alexa as she evolves—the devices will start shipping next month.
  • Amazon has also announced a new Echo Dot,  enabled with new Echo Spatial Perception (ESP) which allows devices to determine which device a user is talking to (meaning that only one will respond, when multiple devices hear the user). The Dot will increase the number of devices Alexa can talk to in the home, creating an innovative customer experience. It will retail for £49.99 and Echo for £149.99.

Here are 2 neat little YouTube videos showing Alexa in action.

 

 

In this context, Amazon is bringing their Alexa training events to Europe in October. Hello Alexa London is on Monday 10th October. 

  • Developers, engineers, QA/testers and anyone who wants to learn how to build skills can participate in the full-day agenda from 8:30am – 4:30pm (+ Happy Hour afterwards!)
  • Business development, strategy and planning, VUX/UX /VUI, account teams, producers and project management professionals can participate in the Alexa Business Development for Agencies session later in the day from 3pm – 4:30pm (and then of course join the Happy Hour!). They can also join the breakfast session, Welcome to Alexa (a 45-minute keynote) and Hello Alexa (a 1-hour session on the basics of creating a skill: what goes into the design, build, test, publish) from 8:30am – 10:45am.
  • Click here to register (although the event is already sold out by now!)  and hope to see you there!

I am really excited at how ubiquitous speech recognition is becoming! It was already ubiquitous as we were dragging it around on our smartphones (Apple SIRI, Google Now), but now it’s penetrated our homes too, crossing over work/personal/family lives. The future is omni-channel but unimodal?!

A.I.: from Sci-Fi to Science reality

17 Jan

Just found this very brief and illustrated History of Artificial Intelligence at LiveScience.com and I couldn’t help but share it!

We have come a long way! (e.g. getting a Chatbot to pass the Turing Test and having to come up with a new test now!) But we also still have a long way to go until the dreaded Singularity moment comes!

A timeline of developments in computers and robotics.

Source:LiveScience

METALOGUE: Building a Multimodal Spoken Dialogue Tool for teaching Call Centre Agents metacognitive skills

21 Jul

METALOGUE logo

Since November 2013, I’ve had the opportunity to actively participate in the new EU-funded FP7 R & D project, METALOGUE, through my company DialogCONNECTION Ltd, which is one of the 10 Consortium Partners. The project aims to develop a natural, flexible, and interactive Multi-perspective and Multi-modal Dialogue system with metacognitive abilities, i.e. a system that can monitor, reason about, and provide feedback on its own behaviour, intentions and strategies, as well as the dialogue itself, guess those of its interlocutor,  and accordingly plan the next step in the dialogue. The goal is to dynamically adapt both its strategy and its behaviour (speech and non-verbal aspects) in order to influence the interlocutor’s reaction, and hence the progress of the dialogue over time, and thereby also achieve its own goals in the most advantageous way for both sides. The project runs for 3 years (until Oct 2016) and has a budget of € 3,749,000 (EU contribution: € 2,971,000). METALOGUE brings together 10 Academic and Industry partners from 5 EU countries (Germany, Netherlands, Greece, Ireland, and UK).

metalogue_replay

The METALOGUE research focuses on interactive and adaptive coaching situations where negotiation skills play a key role in the decision-making processes. Reusable and customisable software components and algorithms will be developed, tested and integrated into a prototype platform, which will provide learners with a rich and interactive environment that will help them develop metacognitive skills, support motivation, and stimulate creativity and responsibility in the decision-making, argumentation, and negotiation process. The project will produce virtual trainers capable of engaging in natural interaction in English, German, and Greek, using gestures, facial expressions, and body language.

 

Pilot systems will be developed for 2 different Sectors: Government and Industry. The corresponding user scenaria that have been selected are: a) Youth Parliamentarian Debating Skill Tutoring (for the Hellenic Youth Parliament) and b) Call Centre Agent Training (for multilingual UK Call Centres). Particularly for Call Centres, we have identified the following broad CC Agent training goals:

CC training goals

CC training goals B

These training goals translate into the following metacognitive skills that a Call Centre Agent needs to learn and which will be taught through the METALOGUE system:

CC training goals C

To this effect, DialogCONNECTION Ltd and myself are looking for UK-based Call Centres, preferably with multilingual agents, that would like to participate in the course of the project.

 

What we need from you

Ideally, we would get access to real-world Call Centre Agent-Caller/Customer recordings. However, simulated Trainer – Trainee phone calls that are used for situational Agent training are also acceptable (either already available or collected specifically for the project). 2 hours of audio (and video if available) would suffice for the 1st year of the project (needed by October 2014). By the 2nd year (Dec 2015) we would need a total of 15 hours of audio. The audio will be used to train the METALOGUE speech recognisers and the associated acoustic and language models, as well as its metacognitive models. We are looking for Call Centres that are either small and agile (serving multiple small clients) or large (and probably plagued by the well-known agent burn-out syndrome). Strict EU Guidelines for data protection will be applied on all the collected and all the published data (e.g. caller anonymisation, sensitive data redaction) and ultimately YOU determine what can and what cannot be published both during and after the project has ended.

 

What’s in it for you

  • Participation, input in, early access & evaluation of all intermediate pilots as they are developed (No need to wait until the end of the project like the rest of the world!)
  • Provide feedback, express wishes with regards to your own requirements for features and functionality (i.e. the pilots and the end system will be customised to your own needs!)
  • Full METALOGUE system at the end for free, customised to your needs and requirements (no source code, or speech recogniser, just the system as-is)

 

If I have sparked your interest, please get in touch by leaving a comment to this post or contact us through our company website. Here is a handy PDF with the invitation to Call Centres (METALOGUE Poster).

 

You can get updates on the progress of the METALOGUE project by connecting with us on LinkedIn and on Twitter. Watch the future happen now!

 

EU FP7 logo

Develop your own Android voice app!

26 Dec

Voice application Development for Android

My colleague Michael F. McTear has got a new and very topical book out! Voice Application Development for Android, co-authored with Zoraida Callejas. Apart from a hands-on step-by-step but still condensed guide to voice application development, you get the source code to develop your own Android apps for free!

Get the book here or through Amazon. And have a look at the source code here.

Exciting times ahead for do-it-yourself Android speech app development!

The AVIxD 49 VUI Tips in 45 Minutes !

6 Nov

Image

 

 

The illustrious Association for Voice Interaction Design (AVIxD) organised a Workshop in the context of SpeechTEK in August 2010, whose goal was “to provide VUI designers with as many tips as possible during the session“. Initially the goal was 30 Tips in 45 minutes. But they got overexcited and came up with a whooping 49 Tips in the end! The Session was moderated by Jenni McKienzie, and the panelists were David Attwater, Jon Bloom, Karen Kaushansky, and Julie Underdahl. This list dates back 3 years now, but it’s by no means outdated. This is the most sound advice you will find in designing better voice recognition IVRs and I hated it being buried in a PDF!

So I am audaciously plagiarising and bringing you here: the 49 VUI Tips for Better Voice User Interface Design! Or go and read the .PDF yourselves here:

Image

Image

Image

Image

Image

Image

Image

Image

Image

Image

And finally ….

Image

 

Have you got a VUI Tip you can’t find in this list that you’d like to share? Tell us here!

 

XOWi: The wearable Voice Recognition Personal Assistant

30 Oct

I just found out about the new venture of my colleagues, Ahmed Bouzid and Weiye Ma, and I’m all excited and want to spread the word!

They came up with the idea of a Wearable and hence Ubiquitous Personal Voice Assistant, XOWi (pronounced Zoe). The basic concept is that XOWi is small and unintrusive (you wear it like a badge or pin it somewhere near you) but still connects to your smartphone and through that to all kinds of apps and websites for communicating with people (Facebook, Twitter, Ebay) and controlling data and information (selecting TV channels, switching the aircon on). Moreover, it is completely voice-driven, so it is completely hands- and eyes-free. This means that it won’t distract you (if you’re driving, reading, working) and if you have any vision impairment or disability, you are still completely connected and communicable. So, XOWi truly turns Star Trek into reality! The video below explains the concept:


The type of application context is exemplified by the following diagram.

XOWi architecture

And here is how it works:

Ahmed and Weiye have turned to Kickstarter for crowdfunding. If they manage to get $100,000 by 21st November, XOWi will become a product and I will get one for my birthday in March 2014! 😀 Join the Innovators and support the next generation in smart communicators!