Macworld (USA)

How Apple’s Xnor.ai acquisitio­n could boost Siri

It’s not about cameras. It’s about smarts.

- BY MICHAEL SIMON

Apple buys a lot of companies throughout the course of a year, with only a couple of them rising to the level of intriguing news. Last year’s purchase of Intel’s smartphone modem business certainly qualifies, as does the 2018 acquisitio­n of Shazam (go.macworld.com/bysz), but for the most part, Apple scopes out companies that we’ve never heard of for reasons we’ll never know.

Its most recent acquisitio­n might be

different. The company, Xnor.ai, might not be one you’ve ever heard of, but they’re hardly unknown. Since last summer, the Seattlebas­ed startup’s tech has been the brains behind the popular Wyze cam’s marquee feature: people detection. Simply put, it allowed the $20 camera to distinguis­h between faces, pets, and dust, and vastly improved its abilities, putting it on a somewhat level playing field with the far-more-expensive Ring and Nest cams of the world.

But it’s not just that Xnor.ai’s engine worked on a budget cam, it’s how it worked. Not only did it vastly improve the capability of the pint-sized recorder, but it also did it with privacy in mind. Using something called Edge AI, Xnor.ai was able to process its algorithm engine on the camera itself, meaning it didn’t need to transmit images to a far-away cloud.

That cuts to Apple’s main privacy argument. We’ve long suspected that the reason why Siri lags Google Assistant and Amazon Alexa is that Apple doesn’t collect the same kind of informatio­n that those companies do and is thus at a disadvanta­ge. Google and Amazon may offer the ability to toggle privacy settings, but the core business model relies on data collection. It’s easier to improve AI processing when you have a mountain of data to work with, especially when you’re dealing with millions of users. But maybe it doesn’t have to be that way.

SIRI SAFE AND SOUND

That’s where Xnor.ai comes in, and likely why Apple deemed it worthy of several million dollars. I don’t think Siri’s developmen­t (or lack thereof) is the result of malaise or a lack of focus from Apple, but rather the capabiliti­es of the AI engine. Apple wants to process as much as it can on the device, but the reality is that it’s just

not possible on Siri’s scale, at least not without a little help.

It’s not for a lack of processing power. Apple’s A-series Bionic chips are certainly capable. The A13 on the iphone 11 has both a faster-than-ever Neural Engine and a set of machine learning accelerato­rs that handle more tasks than ever, but it’s still limited to practical applicatio­ns specific to the device, like battery efficiency, power consumptio­n, and graphics accelerati­on.

But Xnor.ai’s Edge AI engine could be the thing that brings everything together. It’s unlikely that we’ll see any fruits from Apple’s purchase in the iphone 12 or even the iphone 13, but Apple’s incredible silicon advancemen­ts, coupled with the kind of on-device AI processing that Xnor. ai brings, could boost Siri’s capabiliti­es in a big way. By embedding Edge AI into Apple’s own chip via the Neural Engine or a new co-processor, Siri could be faster and far more capable, learning from what you do and prioritizi­ng tasks in kind. And it could all work offline, tapping into the tremendous power of Apple’s system-on-chip and doing the work of a powerful cloud right on the device.

Xnor.ai estimates that Edge AI runs 10 times as fast with 15X memory than cloudbased systems, and a responsive assistant dedicated to each specific phone could finally let Apple build a voice recognitio­n system with near-perfect accuracy. We’ve been waiting years for Siri to be able to do basic things like distinguis­hing between different users’ voices, and Edge AI could bring a great understand­ing of each user’s particular cadence. After all, we already know that Xnor.ai’s Wyze AI was able to distinguis­h people from pets, so using it to differenti­ate voices shouldn’t be all that difficult. That alone would go a long way toward closing the gap between Siri and

Google Assistant and Amazon Alexa. By building a powerful AI engine directly on our

phones, Apple could do the kinds of things with Siri we want it to do without compromisi­ng our privacy.

That same engine could be applied to speech patterns. Siri dictation isn’t bad at all, but saying “period” and “comma” gets tedious. Edge AI could recognize our vocal patterns, so when we pause a certain way it adds a period, or if we change our inflection it adds a question mark.

Let’s go a step further. Say you’re texting with a friend and they ask about seeing a movie. That could trigger Siri to quietly suggest upcoming showtimes inside your conversati­on. Or if you copy a link in Safari, a smart suggestion could instantly present a series of apps before you even press the share button. These are the kinds of things Apple would never do in the cloud—that whole what-happens-on-youriphone-stays-on-your-iphone thing ( go.macworld.com/wthp)— but by using Edge AI, Apple could bring those kinds of interactio­ns to the iphone itself, which opens up Siri to a new world of capabiliti­es.

And if it’s learning on the device, then it could know what app we’re in and respond accordingl­y. So if we’re in Photos, we could say, “Share this with my wife” and it wouldn’t need any extra clarificat­ion. Or if we’re reading a news article in Safari, we could say, “Tell me more about this,” and it would make the appropriat­e search. Or maybe an automatic routine could be suggested based on our app habits.

Bottom line: there’s a lot that Siri doesn’t quite know how to do, and Xnor. ai’s Edge AI engine could help teach it. All while keeping Apple’s privacy promise intact. Unfortunat­ely for Wyze cam owners, Xnor’s Edge AI system has been yanked ( go.macworld.com/yank), but their loss could be Siri’s gain. ■

 ??  ??
 ??  ?? Wyze’s people detection ran completely on the device, so there was no fear of video being hacked.
Wyze’s people detection ran completely on the device, so there was no fear of video being hacked.
 ??  ?? The A13 Bionic chip already has a powerful Neural Engine, but Edge AI could bring it to the next level.
The A13 Bionic chip already has a powerful Neural Engine, but Edge AI could bring it to the next level.
 ??  ?? If Xnor.ai could transform a tiny $20 camera with Edge AI, imagine what it could do for Siri and the iphone.
If Xnor.ai could transform a tiny $20 camera with Edge AI, imagine what it could do for Siri and the iphone.

Newspapers in English

Newspapers from Australia