You are currently viewing Google is working on a new tech that can read your body language without using cameras- Technology News, FP

Google is working on a new tech that can read your body language without using cameras- Technology News, FP


There is no point in denying it, but automation is the future. Imagine a world where your TV pauses the movie or the show that you’re watching when it senses that you’ve stood up to fetch a fresh bowl of popcorn, and resumes playing the content when you return. Or how about a computer that senses you’re stressed out at work and starts playing some mellow and relaxing tunes?

Google is working on a new tech that can read your body language without using cameras

Well, as futuristic as these ideas seem, most of these things are happening now. However, one of the biggest reasons why it hasn’t taken off with a bang, is that these systems use cameras to record and analyse user behaviour. The problem with using cameras in such systems is that it raises a ton of privacy concerns. After all, people are actually paranoid about their computers and smartphones, keeping an eye on them.

Google is actually working on a new system, that records and analyses users’ movement and behaviour, without using cameras. Instead, the new tech uses radar to read your body movements and understand your mood and intentions, and then act accordingly.

The basic idea for the new system is, that a device will use radar to create spatial awareness, and will monitor the space for any changes, and then send out instructions in compliance with what the user would want the system to do.

This isn’t the first time that Google has played with the idea of using spatial awareness-based stimuli for its devices. In 2015, Google unveiled the Soli sensor, which used radar-based electromagnetic waves to pick up precise gestures and movements. Google first used the sensor in Google Pixel 4, when it used simple hand gestures for various inputs, like snoozing alarms, pausing music, taking screenshots etc. Google has also used the radar-based sensor, in the Nest Hub smart display, to study the movement and breathing patterns of a person sleeping next to it.

Studies and experiments around the Soli sensor are now enabling computers to recognize our everyday movements and make new kinds of choices.

The new study focuses on proxemics, the study of how people use space around them to mediate social interactions. This assumes that devices such as computers and mobile phones have their own personal space. 

So when there are any changes in the personal space, the radar picks this up and sends out instructions. For example, a computer can boot up, without you needing to press a button. 

Google is working on a new tech that can read your body language without using cameras

The final frontier for large scale automation has been private, end users, and households. If Google is able to finalise this tech and make it mainstream, it will be a massive win for automation.





Source link

Leave a Reply