How Smart Sound works
Smart Sound is often referred to as what makes the Google Home Max speaker “smarter” than its smaller siblings, the Home and Home Mini, but that hardly explains what’s happening under the hood.
Smart Sound’s main component is a feature called Room EQ, which was designed to help mitigate muddled sound resulting from nearby walls and objects. Room EQ uses each of the Max’s six onboard microphones to measure the low frequencies of the music playing. Those measurements are then fed to a machine-learning model that adjusts the bass output so that the sound is balanced relative to its environment.
Room EQ is made possible entirely by a machine-learning algorithm, which Google trained by placing the device in thousands of different demonstration rooms and real-life homes provided by the employees enrolled in the company’s “dogfooding” program. “We collected an immense amount of data throughout the whole product development cycle,” explained Frances Kwee, the lead acoustics engineer, who helped design the microphone and speaker systems on the Home Max. This data was used to help determine everything about the Max, including how it performed in different temperature environments and whether it could still hear commands at the loudest volume.
Even tuning the speaker’s sound involved hard quantitative data. “Having a seamless listening window for the Google Home Max required a lot of measurements from different angles, different heights,” said Kwee. “No one just stands in front of the speaker. People are walking around and living their life in their home, and we wanted to make sure it sounded balanced no matter what angle you’re at in front of the Google Home Max.”
Can Smart Sound make music sound better?
Some audio professionals aren’t entirely sold on the idea of machine learning as the cure-all for bad room acoustics. Ethan Winer, author of The Audio Expert: Everything You Need to Know About Audio, compares the Max’s Smart Sound feature to digital room correction (DRC), which relies on the same concept of digital filters to help reduce noise and increase fidelity. “It can help a little bit, but there’s a lot of things it can’t help. There’s nothing you can do electronically inside a box to counter echoes — echoes happen outside the box,” he explained. “There’s also a thing called ‘nulls,’ where certain frequencies get canceled out because of room acoustics, and there’s nothing these things can do to fix that, either.”
Winer is also suspicious about the lack of specifications available on the Max’s audio capabilities. “They say ‘technical specs,’ but it’s really features,” said Winer, noting that there is no available information for the Max’s frequency or decibel (dB) range, the main criteria for audio fidelity. He later clarified in an email what an appropriate listing would look like by sending over the owner’s manual for another high-end speaker. “It says the speaker is uniform, within 1.5 dB over most of its range, then at the top of the second page, it shows that it’s 3 dB softer at 49 Hz and 22 KHz. The only important thing missing is the amount of distortion. Most consumers don’t need all the rest shown, though savvy consumers would want that!”
Less skeptical about Smart Sound’s abilities is Daniel Rothmann, a machine-learning engineer who has programmed AI for audio processing and is currently working on a project in collaboration with Dynaudio A/S and Aarhus University in Denmark. “AI definitely promises us better audio-processing technologies that are more flexible and situationally aware,” he wrote in an email. “Up to this point, we’ve had a bit of trouble understanding audio as well as we do images, so that’s probably why we’re only now seeing technologies like Google’s Smart Sound emerging. But in a simple sense, AI technologies are becoming a reality for audio processing.”
Despite his optimism, however, Rothmann admits that there’s still work to be done before the AI in a speaker like the Max can be truly situationally aware. “It is still really difficult to understand ‘auditory scenes,’ or situations with several sounds going on at once. So for now, we deal mostly with the broader idea of a sound ‘profile’ or ‘fingerprint’ from which we can choose a processing setting.”
As for whether audiophiles will ever find a smart speaker like the Google Home Max a compelling product, Rothman thinks it will take time. “As techniques improve in the future, I think they will start to like it.”
Buy it because it’s smart
The Google Home Max’s Smart Sound is certainly smart: Google poured a ton of research into the product in an effort to help make it stand out from its competition. But whether there’s a keen difference between the Max and other Hi-Fi speakers is as of yet indeterminable. Fortunately, it has plenty of other worthy features to rely on, most notably its integration with the Google Assistant. And if you’ve seen the wave of “Make Google do it” commercials, then you may have realized it’s the main reason to consider buying one of these speakers in the first place.
If you specifically want the best-sounding speaker for playing music, Winer suggested going the old-fashioned route by heading to a brick-and-mortar store like Best Buy and listening to a few of your favorite songs on the demo models. After a couple of listens, pick the one that sounds best to your ears. “If you want a stereo, you’re better off getting a real boombox.”