手順 12を翻訳中
-
Next we pry out this hunk of ... stuff, which turns out to be an earpiece speaker, mic, ambient light sensor (AMS TMD3702VC), and the Soli chip, for interpreting your gestures using the power of radar.
-
Google calls this implementation of its in-house Project Soli Motion Sense.
-
Although radar technology has been in use for a long time and seems simple enough on paper, we're at a loss as to how Google stuffed the entire system into a tiny featureless rectangle with no moving parts.
-
The Soli chip then studies the reflected waves and analyzes their time delay, frequency shift, and other data to learn the characteristics of the object that reflected them—how big it is, how fast it's moving, in which direction, etc.
-
Soli then runs that data against its known gesture database to determine what action, if any, needs to be performed in the OS.
-
TL;DR: magic rectangle knows your every move.
クリエイティブコモンズのオープンソース著作権のもと、あなたの投稿は著作権の対象となります。