Why Eye Tracking is a Huge Deal for VR/AR
A game changer in visual fidelity, interface design and interactivity
Last year, Japanese company FOVE released the world’s first VR headset with built-in eye tracking — the technology showed a lot of promise, and in the months that followed, Facebook, Apple & Google all acquired eye-tracking startups to incorporate the technology into their respective XR devices.
So what’s the big deal with eye-tracking, and how can it impact the VR/AR industry?
Better Performance & Natural Focus
Eye tracking allows developers to optimize the performance of VR/AR experiences by focusing system resources specifically where the user is currently looking. This not only lowers VR’s high barrier to entry, but also gives creators the ability to create breathtaking visuals by using their processing resources wisely.
Another major visual improvement comes from the fact that eye-tracking technology can simulate natural focus realistically — a feature that has remained thoroughly absent from VR headsets so far.
A New Way to Design User Interfaces and UX
With the screen-based devices we use today, whenever we want to perform any action we need to tell our device what we want it to do. Usually we do this by touching a certain area of the screen (touch screen interactions), or by pointing at things with a cursor (using a mouse).
Before doing any of those things, however, we always look at what we’re about to interact with, and this is where eye-tracking comes in.
It cuts out the middleman, allowing us to engage with content by simply looking at it. This will give rise to new ways of building User Interfaces that feel natural and are incredibly accurate, completely replacing the need for cursors and most touch based interactions altogether. Eye-tracking interactivity is also discrete by nature, and may allow us to use immersive computers in small public spaces — possibly answering one of the biggest design questions in VR/AR today.
An Analytics Oasis
Eye-tracking will allow VR/MR creators to have access to an unprecedented level of usage analytics — not only they’ll know exactly what users have looked at or ignored throughout an experience, they’ll also be able to accurately measure engagement through pupil tracking.
You may have heard that human pupils dilate on physical attraction: but it goes much further than that. Pupil expansion betrays not only physical attraction, but also mental strain and emotional engagement. It can even go as far as to predict actions of a user seconds before they do it (explored and explained in detail in my article about the future of immersive education).
All of this will be immensely powerful for developers, and will allow them to combine these bits of data to create immersive software that’s 100% reactive to a user’s emotions and truly understands what’s going through their mind as they go further into the experience.
New Gameplay Mechanics and Interactions
Eye-tracking will also give way to a number of new interactions and game-play mechanics that were never possible before — virtual characters will now be aware of when you’re looking at them, even going as far as to cross-examine what you’re looking at and why.
Users will be able to aim with their eyes, make narrative choices by simply gazing at an object, and meaningfully change the world around them with almost subconscious gestures, opening up a number of new opportunities for creative storytelling and interaction design.
These are some of the reasons why Eye-Tracking will be a huge deal for the immersive industry going forward. Did I miss anything? Let me know your thoughts in the comments — and follow me on Twitter!
Thanks for reading!
Lucas Rizzotto is an award-winning Immersive Experience Designer, Artist and Creator.