Affiliate Disclosure
If you buy through our links, we may get a commission. Read our ethics policy.

Google Assistant response speed getting improved by on-device processing

Last updated

The Google Assistant will be quicker to respond to user queries faster by performing processing on the device, the search company revealed at Google I/O, with the next generation of the virtual assistant able to quickly switch between apps and help with multitasking in Android than for previous versions.

The promise of on-device processing of digital assistants offer the possibility of faster response times for queries based on locally-held data, without necessarily requiring an internet connection. It also has the prospect of fending off any potential privacy issues caused through dispatching requests to the cloud, where the data may be analyzed and used for other purposes.

At Google I/O, the search company suggested it was getting closer to the idea of on-device virtual assistants, with the next version capable of processing requests up to ten times faster.

An on-stage demonstration showed someone rattling through multiple queries in quick succession, with a small corner transcript of what was being heard revealing a lot of the processing was being carried out at a very high speed. Much faster than what would normally be expected by offloading processing to a separate server.

According to Google, it has reduced the size of the artificial intelligence models used for speech recognition and language understanding down from hundreds of gigabytes in size to less than half a gigabyte, making it more manageable to store on a smartphone. This enables the next generation Assistant to perform near real-time transcription without an Internet connection.

The Assistant also includes query features like Continued Conversation, where multiple queries could be made in a row without requiring the use of "Hey Google" every time. The examples also showed how a user could write a message out verbally and insert an image from a separate app, all without requiring any finger presses or vocal prompts to change task.

Google anticipates the next-generation Assistant will be arriving initially on its new Pixel smartphones later this year. It is unclear when it will roll out to other Android device, or even to the iOS app.

Google is not the only company grappling with the idea of on-device processing for virtual assistants.

Siri, Apple's digital assistant that is on a par with Google Assistant in terms of worldwide usage, could have an offline mode if a November patent application comes to fruition. Titled "Offline personal assistant," the Apple filing suggests the use of speech-to-text processing and validation to occur on the device, and if a connection was available, could perform both local and remote processing for two possible interpretations of a query, which could be used to further confirm a transcription as correct.



21 Comments

gatorguy 13 Years · 24627 comments

Everyone will appreciate a feature available starting today:
"Users in English-speaking countries will be able to stop ringing alarms and timers on Google Home and Smart Display devices simply by saying "stop," no hotword required."

jeffythequick 6 Years · 269 comments

Duh...
Microsoft figured this out in 2005 with Voice Command, which existed on the Windows Phones.

Streaming requests that can be handled locally (i.e. "Call Julie on Mobile", or "what appointments do I have today?") over the Internet is just dumb, inefficient, and exposes privacy concerns.  Items such as "Who put the bop in the bop-she-bop-she-bop" or "What movies are playing on El Cajon Blvd theaters tonight?" are those that parsed on the device, then a web query can be used to search the Internet.

A central place for all of my, and everyone else's queries are the ingredients for 1984.

avon b7 20 Years · 8046 comments

We all knew this capacity was coming but it's good to see it 'go live' so to speak. Things like music playback control by voice should be doable via assistants without requiring an internet connection.

Two years ago we were told onboard AI would be a major advance for low latency, ultra fast response situations and where privacy is paramount. It's becoming reality.

genovelle 16 Years · 1481 comments

Duh...

Microsoft figured this out in 2005 with Voice Command, which existed on the Windows Phones.

Streaming requests that can be handled locally (i.e. "Call Julie on Mobile", or "what appointments do I have today?") over the Internet is just dumb, inefficient, and exposes privacy concerns.  Items such as "Who put the bop in the bop-she-bop-she-bop" or "What movies are playing on El Cajon Blvd theaters tonight?" are those that parsed on the device, then a web query can be used to search the Internet.

A central place for all of my, and everyone else's queries are the ingredients for 1984.

Interesting. Device and media control has been handled on device since 2001 on macs. The original iPhone and even some iPods before it could as well via voice commands. Apple announced for iOS 11 that most Siri request were being processed on device. Only database or internet queries would require a connection. So, Google is announcing they are going to do something Apple implemented 2 years ago, and the article pretends Apple is working on it, when in fact they did it in 2017. Odd.  Here’s an article about it from Sept 2017. https://www.fastcompany.com/40443055/apple-explains-how-its-making-siri-smart-without-endangering-user-privacy

gatorguy 13 Years · 24627 comments

genovelle said:
Duh...

Microsoft figured this out in 2005 with Voice Command, which existed on the Windows Phones.

Streaming requests that can be handled locally (i.e. "Call Julie on Mobile", or "what appointments do I have today?") over the Internet is just dumb, inefficient, and exposes privacy concerns.  Items such as "Who put the bop in the bop-she-bop-she-bop" or "What movies are playing on El Cajon Blvd theaters tonight?" are those that parsed on the device, then a web query can be used to search the Internet.

A central place for all of my, and everyone else's queries are the ingredients for 1984.
Interesting. Device and media control has been handled on device since 2001 on macs. The original iPhone and even some iPods before it could as well via voice commands. Apple announced for iOS 11 that most Siri request were being processed on device. Only database or internet queries would require a connection. So, Google is announcing they are going to do something Apple implemented 2 years ago, and the article pretends Apple is working on it, when in fact they did it in 2017. Odd.  Here’s an article about it from Sept 2017. https://www.fastcompany.com/40443055/apple-explains-how-its-making-siri-smart-without-endangering-user-privacy

As far as I know Siri voice requests are still processed "in the cloud" rather than on-device. What Google is planning is something improved from that aren't they? 

https://www.youtube.com/watch?time_continue=7&v=FDdorhY7fb8