Google improves the quality of the calls on the Duo thanks to a new model of artificial intelligence

Share on facebook
Share on twitter
Share on linkedin
Share on reddit
Share on email
Share on whatsapp

Table of Contents

Google tends to explain quite often how to work with different machine learning models to improve the performance of your applications. He recently did explaining how they worked the radars of Soli and now explains how to operate the new improvements in audio quality of the Google Duo.

What has been achieved with a new model called the WaveNetEQ, a system of generation based on the technology of DeepMind able to complete the packets lost in the waves of voice. Let’s see how he manages to Google such a feat.


Improved Google quality audio Duo via a generative model

Image 2020 04 02 12 56 11 Google account, when you send packets of data over the internet, it is easy to have mistakes at the time of receiving them.

Google account that when you transmit a call through internet the packages have some quality problems. These problems are due to excessive fluctuations or delays in the network, and may lose as much as 8% of the total content due to e these.

Surely you’ve ever suffered voices canned or robotic in a video call. The main reason is the loss of quality in the process of sending and receiving packets, if lost several, low quality

To make sure that the communication works correctly in real time, Google has created WaveNetEQa PLC system (programmable logic controller) trained with a generous base of voice data. What makes this model? The explanation is quite technical and complex, so that we will summarize it in the simplest way.

Image 2020 04 02 12 57 01 The Google system analyses the audio packets to try to predict what you want to say. With this information processed in neural networks, complete the spectrum of wave that is missing.

WaveNetEQ is a generative model that allows to synthesize the waves of voice even though they have lost parts of the wave. Surely you’ve ever had bad quality on video call and have heard a sound robotic or metallic. This is because when the lack of packages is high (there is a lot of latency) and the sound may not be reproduced with quality.

Using neural networks Google is able to give continuity to the signal in real time, minimizing the loss of quality. Basically, thanks to the speech database, the model “guess” what you intend to say and complete the wave with fragments missing.

Image 2020 04 02 12 59 56 Blue line original audio. In orange, the lines that Google predicts.

These improvements will begin to be applied already in the Google Duo Google Pixel 4although the company ensures that the model will come to the rest of the terminals in a littleit will be an improvement of the application, not only of the devices.

More information | Google


The news

Google improves the quality of the calls on the Duo thanks to a new model of artificial intelligence

it was originally published in

Xataka Android

by
Ricardo Aguilar

.

Google improves the quality of the calls on the Duo thanks to a new model of artificial intelligence 1

Google improves the quality of the calls on the Duo thanks to a new model of artificial intelligence 2