Do Speech Assistants like Alexa have difficultly with Emotion?

The 26th Irish Conference on Artificial Intelligence and Cognitive Science was hosted by the Trinity College Dublin on the 6th and 7th of December in the Long Room Hub. The conference covered an interesting variety of topics across computer science, psychology and neuroscience. This led to interesting presentations including one asking the question whether Amazon’s Alexa suffers from Alexithymia?

 

QxLab was well represented with two papers: Rahul, who is funded by a scholarship from the SFI CONNECT research centre for future networks, presented his work on voice over IP quality monitoring with a presentation entitled: “The Sound of Silence: How Traditional and Deep Learning Based Voice Activity Detection Influences Speech Quality Monitoring”. In his paper he investigated how different voice activity detection strategies can influence the outputs from a speech quality prediction model.  In his experiment used a dataset with labelled subjective quality judgments.

Alessandro also had a paper in the proceeding and gave an interesting talk on exploring a DNN-based fusion for speech separation. Alessandro’s PhD is funded by the SFI Insight Centre for data analytics.

QxLab picks up the inaugural CONNECT Research Impact Award

QxLab’s Dr Andrew Hines accepted the inaugural CONNECT Research Impact Award in recognition of a technique, developed in collaboration with Google, which is now used to assess the audio quality of YouTube’s most popular videos.

He was presented with the award by CONNECT Director, Prof. Luiz DaSilva. CONNECT is the Science Foundation Ireland research centre for future networks and communications. Dr Hines, a Funded Investigator at the centre.

“This area of research is called ‘Quality of Experience and it is a key pillar in CONNECT’s research platform,’” Andrew explains. “The research leading to this development was conducted during my time in UCD, DIT, and Trinity College Dublin.

“Our work has also been used in the development of the OPUS codec which is used to stream over 1 billion minutes of audio per week through the Chrome browser. The model has been freely shared and is used for research and development across other Google research divisions for helping applications ranging from trying to restore speech to silent movies to lipreading at cocktail parties. Other industry leaders such as Apple have also used it in headphone assessment.”