Man versus machine reaches symbiotic state; eye tracking and gesture applications to revolutionize patient care
Jul 11, 2016
Gesture, eye tracking, and proximity sensor technologies will mark the next stage of innovation in machine design, finds ABI Research. A broader and more competitive ecosystem spurred by smartphone and tablet sensor integration—forecast to hit close to $5 billion in 2016—will create massive opportunities in automotive, consumer electronics, and healthcare. Healthcare, in particular, shows the largest, untapped opportunity for eye tracking and gesture applications in patient care.
“The same way that touchscreens eclipsed the PC mouse, gesture and eye tracking sensors will transform the way people interact with machines, systems, and their environment,” says Jeff Orr, Research Director for ABI Research. “Healthcare professionals are relying on these sensors to move away from subjective patient observations and toward more quantifiable and measurable prognoses, revolutionizing patient care.”
Eye tracking sensors can help detect concussions and head trauma, identify autism in children before they are speaking, and enable vision therapy programs for early childhood learning challenges to retrain the learned aspects of vision. Similarly, gesture sensors are translating sign language into speech, providing doctors a means to manipulate imaging hands-free during surgical procedures, and providing a natural means to navigate through virtual experiences.
Both established and startup companies are involved in the human-machine interface revolution. Sensor innovation is stemming from Hillcrest Labs, NXP, and Synaptics, among others. Atheer, Bluemint Labs, eyeSight, Google, Intel, Leap Motion, Microsoft, Nod Labs, RightEye, and Tobii Group also all recently announced creative gesture, proximity, and eye tracking solutions.
“Healthcare is only one industry poised to benefit from reinventing the user interface,” adds Orr. “The larger competitive ecosystem for perceptual sensors is forging opportunities in consumer appliances, autonomous driving, musical instruments, gaming, retail, and even hazardous locations.”
These findings are from ABI Research’s Eye Tracking, Gestures and Proximity Sensor Applications and Human-Machine Interfaces webinar. This report is part of the company’s Wearables & Devices sector, which includes research, data, and analyst insights.