Archive | Design & Development Standardisation RSS feed for this section

Human-Machine Interaction in Translation (NLPCS 2011)

21 Aug

For a few years now I have been in the Programme Committee of the International Workshop on Natural Language Processing and Cognitive Science (NLPCS), organised by a long-time colleague and friend, Dr. Bernadette Sharp from Staffordshire University. The aim of this annual workshop is “to bring together researchers and practitioners in Natural Language Processing (NLP) working within the paradigm of Cognitive Science (CS)“.

The overall emphasis of the workshop is on the contribution of cognitive science to language processing, including conceptualisation, representation, discourse processing, meaning construction, ontology building, and text mining.”

There have been NLPCS  Workshops in Porto (2004), Miami (2005), Paphos (2006), Funchal (2007), Barcelona (2008), Milan (2009) and Funchal (2010).

Copenhagen Business School

Copenhagen Business School

This year’s 8th International NLPCS Workshop just took place this weekend in Copenhagen, Denmark (20-21 Aug 2011). The Workshop topic was: “Human-Machine Interaction in Translation“, focussing on all aspects of human and machine translation, and human-computer interaction in translation, including:  translators’ experiences with CAT tools, human-machine interface design, evaluation of interactive machine translation, user simulation and human factors. Thus, the topics were approached from a number of different perspectives:

  • from full automation by machines for machine (traditional NLP or HLT)
  • semi-automated processing, i.e. machine-mediated processing (programs assisting people in their tasks),
  • but also simulation of human cognitive processes

I had the opportunity once again to review a few of the paper submissions and can therefore highly recommend reading the full Proceedings of the NLPCS 2011 Workshop that have just been made available.

I found particularly interesting the following 3 contributions:

  • Valitutti, A. “How Many Jokes are Really Funny? A New Approach to the Evaluation of Computational Humour Generators”
  • Nilsson, M. and J. Nivre. “Entropy-Driven Evaluation of Models of Eye Movement Control in Reading” 

and

  • Finch, A., Song, W., Tanaka-Ishii, K. and E. Sumita. “Source Language Generation from Pictures for Machine Translation on Mobile Devices”

Enjoy!

Advertisements

Speech Interaction on Mobile Devices at SpeechTEK 2011 (New York)

7 Aug

Today sees the launch of the Joint AVIxD / IxDA Workshop on Speech Interaction on Mobile Devices that kick-starts the mother of Voice Solutions Fairs, SpeechTEK 2011 in New York next week (8-10 Aug).

AVIxD

AVIxD is the Association for Voice Interaction Design, a professional organisation that aims to

“eliminate apathy and antipathy toward the need for good design of automated voice services”, 

which has become my favourite VUI mantra!

IxDA is the Interaction Design Association, a much bigger professional “un-organisation” which  intends to:

“improve the human condition by advancing the discipline of Interaction Design”

A very worthy cause indeed, especially since it is true that “the human condition is increasingly challenged by poor experiences. “!

IxDA

Today’s Joint Workshop in New York aims to bring together interaction design practitioners from across the voice, interactive, and digital areas to identify the issues and challenges involved in  speech interaction design on mobile devices, such as smartphones and tablets, and to come up by the end of the day with ways to approach them or even tackle them. A very ambitious format that, however, really does work!

AVIxD organised another Workshop this year on Cross-linguistic & Cross-cultural Voice Interaction Design, which was also the 1st European Workshop, just before SpeechTEK Europe in London this May past. See what we all came up with in those 6 hours in the SpeechTEK Europe PDF presentation below.

And if you don’t manage to take part in today’s workshop, make sure you go to the SpeechTEK Conference and Exhibition itself that starts tomorrow and runs until Wednesday the 10th. Listen to presentations and see or even try for yourself market-ready products relating to:

  • multimodal applications
  • cross-channel applications
  • speech analytics
  • speaker identification and verification
  • in-car systems
  • natural language and say-anything technologies
  • speech translation
  • voice-enabled personal assistants
  • as well as the latest speech recognition techniques and technologies

I particularly recommend the Keynote Panel on “Mobility — A Game-Changer for Speech?” on Tuesday on how smartphones are dramatically changing how customers interact with businesses and with the devices themselves. Some really interesting issues and questions will be raised, such as:

* How voice user interfaces will be integrated with graphical user interfaces?

or

* Will users embrace voice as they have embraced keypads on mobile devices? 

Sadly I am in the UK today and next week, so I’m going to miss it all. But if you are lucky enough to be in or near New York, make sure you go and enjoy!

SpeechTEK 2011 New York

SpeechTEK Europe 2011 – The Voice Solutions Showcase

20 May

(update at the end)

SpeechTEK Europe 2011 takes place in London next week (25 – 26 May 2011, Copthorne Tara Hotel, London, UK) and I am participating very actively! Firstly, I am co-chairing the Workshop on Cross-linguistic & Cross-cultural Voice Interaction Design organised by the Association for Voice Interaction Design (AVIxD). I have already written a blog post on that. Then, I will be presenting the outcome of our discussions at the Workshop in the Main SpeechTEK Conference itself, on Wednesday 25th May (2:45 p.m. – 3:30 p.m) during Session B104: Speech organisations speak out. It should be a challenge as the Workshop runs from 1-7pm the previous day, so I will have a very busy evening after dinner trying to prepare a coherent and comprehensive presentation!!

And finally, on both days of the Main Conference (Wed 25 – Thu 26 May), I will be holding the free consultancy one-to-one appointments in the context of the brand new for this year Meet the Consultants Clinic.  I am one of the “5 global speech tech experts” available “to discuss your speech tech needs and challenges“. Maybe you need to check out my older blog post on speech recognition (for dummies!) to get an idea of what I will be chatting about with everyone. You may also want to check out my presentation slides from last year and from 2007. Get them from these older blog posts: ““The Eternal Battle Between the VUI Designer and the Customer“ and “Does Your Customer Know What They are Signing off??“. Although you do need to pre-book, these appointments are free for registered conference delegates or Expo visitors, so I’m looking forward to meeting some of you in person!

There’s still time to sign up for the SpeechTEK Europe Conference and Free Entry Expo. Use the following link to register and we’ll see you in London next week! http://www.speechtek.com/europe2011/Registration.aspx

Here’s a quick round-up of what’s happening:

  • Conference Keynotes by Google‘s Engineering Director, Dave Burke, who tells SpeechTEK Europe about Google’s plans for cloud-based speech recognition, and Professor Alex Waibel who describes and demonstrates how speech technology is helping to overcome language and cultural barriers. Free entry for Expo visitors too.
  • Learn from over 50 global expert speakers sharing their experiences – both good and bad – and enabling you to build the ultimate multimodal experience for your customers, saving you money and improving your service.
  • Network with colleagues from all over the world, who have already implemented successful strategies. Companies attending include ABN Amro Bank, Apple, Barclays Bank, Microsoft, Orange, Lloyds Bank, Dell, Cap Gemini and more.
  • Identify, evaluate, integrate, and optimise the latest speech technology solutions from world-leading providers at SpeechTEK Europe’s Expo.

SpeechTEK Europe features over 50 speakers from around the world, and from a wide range of business environments including Google, Barclays Bank, Deutsche Telekom, Nuance, Loquendo, Openstream, Voxeo, Belgian Railways, Telecom Italia, Cable & Wireless, and Westpac.

LEARN ABOUT

Business strategies – Speech biometrics – Multichannel applications – Multilingual applications – Multimodal applications – Assistive technologies – Analytics and Measurement – Voice User Interaction design – Speech application development tools and languages – Case studies, panel discussions and more …

UPDATE

SpeechTEK Europe 2011 has come and gone and I’ve got many interesting things to report (as I have been tweeting through my @dialogconnectio Twitter account).

But first, here are the slides for my presentation at the main conference on the outcome of the AVIxD Workshop on Cross-linguistic & Cross-cultural Voice Interaction Design organised by the Association for Voice Interaction Design (AVIxD). I only had 12 hours to prepare them – including sleep and London tube commute – so I had to practically keep working on them until shortly before the Session! Still I think the slides capture the breadth and depth of topics discussed or at least touched upon at the Workshop. There are several people now writing up on all these topics and there should be one or more White papers on them very soon (by the end of July we hope!). So the slides did their job after all!

Get the slides in PDF here:  Maria Aretoulaki – SpeechTEK Europe 2011 presentation.

TECHGRUMPS: technology addictions & the rise of a new social (un)conscience

26 Apr

On Easter Sunday (24 April 2011), I was happy and honoured to take part in the live recording of the latest TECHGRUMPS podcast, Techgrumps 27: Non geeks go raw like sushi (sic!).  80 minutes of whinging about the latest technology trends, as well as the uses of said technology.

My contribution to the grump world is complaining about the social terror of checking your smartphone notifications every 5 minutes, whatever the (social) context, and the de facto new social media exhibitionism regarding all facets of your personal life through the various social media (a stark contrast to my earlier blog posts on the Social Media Scenes in Manchester and London!). Hear me from the 10th to the 32nd minute complain about:

  •  people spending more time updating their current location and taking photos and videos at a gig rather than dancing, singing and enjoying said gig (check the phone screens in the two photos below I took from a Jamiroquai gig earlier this month)

Jamiroquai at MEN Arena Manchester (19 Apr 2011)

Jamiroquai at MEN Arena Manchester (19 Apr 2011)

  • people checking their Facebook or Twitter notifications on their phone in the middle of a philosophical conversation (usually initiated by the person without a smartphone ;))
  • people checking their phone every 5 minutes in the middle of a film at the cinema, just in case someone has texted them or has posted a witticism on Twitter or Facebook (and that’s even when the film is NOT horrible)
  • people needing to offload very personal information and details on their daily routines every hour of the day on their wide social media audiences, which consist mainly of remote acquaintances rather than close friends (who are usually not remotely interested in said details either)

This excessive notification checking, irrespective of the current social situation, is of course partly due to the availability of the technology itself, i.e. integration of Facebook or Twitter on your phone, internet on-the-go, dedicated notification sounds for texts, Facebook, Twitter, chat etc. So, in all fairness, it is hard not to check your phone when you do get a notification (sound). For all you know, it could be a missed call from a loved one who has been in an accident, or an email confirming that new contract. Nevertheless, it seems that we are all sucked up in a world of instantly available information and an overflow of personal and less personal data that we don’t seem able to escape from. As a result, we are missing the NOW, the experience of the current moment and of the person(s) standing opposite us in real life. This obsessive behaviour can be construed as  rude and anti-social by the people in the immediate surroundings not checking their phones, but – more than anything – it indicates a shift in general social conscience and social mores, whereby the remote online acquaintance in the US you have never met in your life  is allocated by default the same or more (potential?) value than the close offline friend sitting next to you here and now. So new types of shallow relationships are cropping up. Whether someone has retweeted you is becoming more important than whether someone actually lends an open ear to you at a cafe to discuss your problems over a cup of  coffee.

This need to connect and be “approved” by as many people as possible, whether real close friends, Facebook “friends” or Twitter followers you are not even remotely interested in, must have its roots at the basic human need for love, approval and the sense of belonging (in the right groups). Still, it seems that our whole lives are run by this new need for exhibitionism and we are practically controlled indirectly by our ubiquitous and international audience who is or may be reading.

Having suffered the social media notification terror myself when sitting at my laptop, I refuse to use that functionality or indeed the internet on my (admittedly palaeolithic) phone.  Even the thought of getting a free smartphone scares me! My time when I’m away from my laptop is my treasured time OFFLINE and I want it to remain that way! I have already spent thousands of invaluable hours chained to my laptop obsessing over emails and notifications in the past 20 years, hours that have been sadly subtracted off MY LIFE! So this is not a rant about Social Media – which often really help in the democratisation of Governments, processes and opinion. This is a rant about Social Media abuse and their infliction onto others as well as onto ourselves.

It sounds very heavy but the whole podcast is actually full of witty jokes and hearty laughter! And there are several more techy topics covered, as you can see on the podcast page: from the “native” IE to Firebug, Wikimedia, LaTeX, and the latest iphone personal information storage scare.  Enjoy!

Techgrumps 27: Non geeks go raw like sushi

Cross-linguistic & Cross-cultural Voice Interaction Design

31 Jan

(update at the end)

2010 saw the first SpeechTEK Conference to have taken place outside of the US, SpeechTEK Europe 2010 in London. This year’s European Conference, SpeechTEK Europe 2011, will take place again in London (25 – 26 May 2011), but this time it will be preceded on Tuesday 24th May by a special Workshop on Cross-linguistic & Cross-cultural Voice Interaction Design organised by the Association for Voice Interaction Design (AVIxD). The main goal of AVIxD is to bring together voice interaction and experience designers from both Industry and Academia and, among other things, to “eliminate apathy and antipathy toward the need for good design of automated voice services” (that’s my favourite!). This is the first AVIxD Workshop to take place in Europe and I am honoured to have been appointed Co-Chair alongside Caroline Leathem-Collins from EIG.

Participation is free to AVIxD members and just £25 for non-members (which may be applied towards AVIxD membership). However in order to participate in the workshop, you need to submit a brief position paper in English (approx. 500 words) on any of the special topics of interest of the Workshop (See CFP below). The deadline for electronic submissions is Friday 25 March, so you need to hurry if you want to be part of it!

Here’s the full Call for (Position) Papers from the AVIxD site:

Call for Position Papers

First European AVIxD Workshop

Cross-linguistic & Cross-cultural Voice Interaction Design

Tuesday, 24 May 2011 (just prior to SpeechTEK Europe 2011), 1 – 7 PM

London, England

The Association for Voice Interaction Design (AVIxD) invites you to join us for our first voice interaction design workshop held in Europe, Cross-linguistic & Cross-cultural Voice Interaction Design. The AVIxD workshop is a hands-on day-long session in which voice user interface practitioners come together to debate a topic of interest to the speech community. The workshop is a unique opportunity for them to meet with their peers and delve deeply into a single topic.

As in previous years with the AVIxD Workshops held in the US, we will write papers based on our discussions which we will then publish on www.avixd.org. Please visit our website to see papers from previous workshops, and for more details on the purpose of the organization and how you can be part of it.

In order to participate in the workshop, individuals must submit a position paper of approximately 500 words in English. Possible topics to touch upon in your submission (to be discussed in depth during the workshop) include:

  1. Language choice and user demographics
  2. Presentation of the language options to the caller and caller preference
  3. Creation and (co-)maintenance of dialogue designs, grammars, prompts across languages
  4. Political and sociolinguistic issues in system prompt choices and recognition grammars, such as code-switching, formal versus informal registers
  5. Guidelines for application localization, translation, and interpretation
  6. Setting expectations regarding availability of multilingual agents, Language- and culture-sensitive persona definition
  7. Coordinating usability testing and tuning across diverse linguistic / cultural groups
  8. Language choice and modality preference

We always encourage the use of specific examples from applications you’ve worked on in your position paper.

Participation is free to AVIxD members; non-members will be charged £25, which may be applied towards AVIxD membership at the workshop. Please submit your position papers via email no later than Friday 25 March 2011 to cfp@avixd.org. Letters of acceptance will be sent out on 30 March 2011.

We look forward to engaging with the European speech design community to discuss the particular challenges of designing speech solutions for users from diverse linguistic and cultural backgrounds. Feel free to contact either of the co-chairs below, if you have any questions.

Caroline Leathem-Collins, EIG  (caroline {at} eiginc {dot} com)

Maria Aretoulaki, DialogCONNECTION Ltd (maria {at} dialogconnection {dot} com)

UPDATE

SpeechTEK Europe 2011 has come and gone and I’ve got many interesting things to report (as I have been tweeting through my @dialogconnectio Twitter account).

But first, here are the slides for my presentation at the main conference on the outcome of the AVIxD Workshop on Cross-linguistic & Cross-cultural Voice Interaction Design organised by the Association for Voice Interaction Design (AVIxD). I only had 12 hours to prepare them – including sleep and London tube commute – so I had to practically keep working on them until shortly before the Session! Still I think the slides capture the breadth and depth of topics discussed or at least touched upon at the Workshop. There are several people now writing up on all these topics and there should be one or more White papers on them very soon (by the end of July we hope!). So the slides did their job after all!

Get the slides in PDF here:  Maria Aretoulaki – SpeechTEK Europe 2011 presentation.

The eternal battle between the VUI Designer & the Customer

7 Dec

I promised some time ago to put up the slides of my presentation at this year’s SpeechTEK Europe 2010 in London, the first SpeechTEK to have taken place outside of the US. My presentation, “The Eternal Battle Between the VUI Designer and the Customer“, was on Wednesday 26th May 2010 and opened the “Voice User Interface Design: Major Issues” Session.  It went down really well, and I had afterwards several people in the audience tell me about their own experience and asking me for tips on how to deal with similar issues.

Here is a PDF with the presentation slides:

Maria Aretoulaki – “The Eternal Battle Between the VUI Designer and the Customer” (SpeechTEK Europe 2010 presentation)

Maria Aretoulaki – “The Eternal Battle Between the VUI Designer and the Customer” (SpeechTEK Europe 2010 presentation)

Maria Aretoulaki – SpeechTEK Europe 2010 presentation UPDATED ppt

And here’s the gist of it:

VUI Design is preoccupied with the conception, the design, the implementation, the testing, and the tuning of solutions that work in the most efficient, secure and non-irritating for the user manner. Well, realistically that’s what VUI Design can achieve. In an ideal world, the VUI Designer would actually strive to create speech applications that – apart from taking into consideration the customer’s financial and brand requirements – would also fit the caller’s needs, goals and preferences. The initial Requirements analysis should bring both in focus. So much is already known and accepted both amidst the VUI Designers and the customers.

The problems start just after they all leave the meeting room and start working on the implementation: Call flow design, system persona development and prompt crafting, but even recognition grammars, all seem to fall victim of a war of words and attitudes between the VUI Design expert who has seen systems being developed and spurned before, and the customer with his tech-savvy business team and their technical architects and programming geniuses, who all think they know what callers want and how call flows should be structured, prompt wording crafted and grammars written, just because they have got strong opinions! Even the results of Usability tests are liable to different interpretations by each side.

This presentation pinpoints common pitfalls in the communication between a VUI Designer and customer employees and recommends ways to resolve conflicts and disagreements on the application design and implementation.

Credits:

SpeechTEK Europe 2010 was organised by:

Information Today, Inc.
143 Old Marlton Pike
Medford NJ 08055 U.S.A.
Phone 1 (609) 654-6266.
http://www.infotoday.com

The voice-activated lift won’t do Scottish! (Burnistoun S1E1 – ELEVEN!)

28 Jul

Voice recognition technology? …  In a lift? … In Scotland? … You ever TRIED voice recognition technology? It don’t do Scottish accents!

Today I found this little gem on Youtube and I thought I must share it, as apart from being hilarious, it says a thing or two about speech recognition and speech-activated applications. It’s all based on the urban myth that speech recognisers cannot understand regional accents, such as Scottish and Irish.

Scottish Elevator – Voice Recognition – ELEVEN!

(YouTube – Burnistoun – Series 1 , Episode 1 [ Part 1/3 ])

What? No Buttons?!

These two Scottish guys enter a lift somewhere in Scotland and find that there are no buttons for the floor selection, so they quickly realise it’s a “voice-activated elevator“, as the system calls itself. They want to go to the 11th floor and they first pronounce it the Scottish way:

/eh leh ven/

That doesn’t seem to work at all.

You need to try an American accent“, says one of them, so they try to mimic one, sadly very unsuccessfully:

/ee leh ven/

Then they try a quite funny, Cockney-like English accent:

/ä leh ven/

to no avail.

VUI Sin No. 1: Being condescending to your users

The system prompts them to “Please speak slowly and clearly“, which is exactly what they had been doing up to then in the first place! Instead, it should have said something along the lines of “I’m afraid I didn’t get that. Let’s try again.” and later “I’m really sorry, but I don’t seem to understand what you’re saying. Maybe you would like to try one more time?“. Of course, not having any buttons in the lift means that these guys could be stuck in there forever! That’s another fatal usability error: Both modalities, speech and button presses, should have been allowed to cater for different user groups (easy accents, tricky accents) and different use contexts (people who have got their hands full with carrier bags vs people who can press a button!).

I’m gonna teach you a lesson!

One of them tries to teach the system the Scottish accent: “I keep saying it until she understands Scottish!“, a very reasonable expectation, which would work particularly well with a speaker-dependent dictation system of the kind you’ve got on your PC, laptop or hand-held device. This speaker-independent one (‘cos you can’t really have your personal lift in each building you enter!) will take a bit more time to learn anything from a single conversation! It requires time analysing the recordings, their transcriptions and semantic interpretations, comparing what the system understood with what the user actually said and using those observations to tune the whole system. We are talking at least a week in most cases. They would die of dehydration and starvation by then!

VUI Sin No.2: Patronising your users until they explode

After a while, the system makes it worse by saying what no system should ever dare say to a user’s face: “Please state which floor you would like to go to in a clear and calm manner.” Patronising or what! The guys’ reaction is not surprising: “Why is it telling people to be calm?! .. cos Scottish people would be going out for MONTHS at it!“.

Well, that’s not actually true. These days off-the-shelf speech recognition software is optimised to work with most main accents in a language, yes, including Glaswegian! Millions of real-world utterances spoken by thousands of people with all possible accents in a language (and this for many different languages too) are used to statistically train the recognition software to work equally well with most of them and for most of the time. These utterances are collected from applications that are already live and running somewhere in the world for the corresponding language. The more real-world data available, the better the software can be tuned and the more accurate the recognition of “weird” pronunciations will be, even when you take the software out of the box.

VUI Best Practice: Tune your application to cater for YOUR user population

An additional safeguarding and optimising technique is tuning the pronunciations for a specific speech recognition application.  So when you already know that your system will be deployed in Scotland, you’d better add the Scottish pronunciation for each word explicitly in the recognition lexicon.  This includes manually adding /eh leh ven/ , as the standard /ee leh ven/ pronunciation is not likely to work very well. Given that applications are usually restricted to a specific domain anyway (selecting floors in a lift, getting your bank account balance, choosing departure and arrival train times etc.), this only needs to be done for the core words and phrases in your application, rather than the whole English, French, or Farsi language! So do not despair, there’s hope for freedom (of speech) even for the Scottish! 🙂

For a full transcript of the video, check out EnglishCentral.