- Jay Inslee floats Megan Rapinoe as his secretary of state pick 4 Years Ago
- The cast list for the ‘Kingsman’ prequel movie looks totally nuts 4 Years Ago
- The best Prime Day deals to heat up your kitchen 4 Years Ago
- YouTuber Emily Hartridge killed in electric scooter crash Today 10:50 AM
- Is Lashana Lynch really playing 007 in the new Bond movie? Today 10:33 AM
- Trump demands apology after his racist tweets Today 10:21 AM
- Prime Day deals that’ll make you grateful for your Amazon membership Today 9:51 AM
- Netflix’s ‘4L’ takes a long road trip to the Sahara Today 9:04 AM
- Air Force says it’s ready to ‘protect’ Area 51 amid Facebook event buzz Today 9:02 AM
- Get 50% off 23andMe DNA tests today for Prime Day Today 9:00 AM
- The best Prime Day streaming deals Today 8:47 AM
- ‘Veronica Mars’ season 4 is a triumphant return to form Today 8:19 AM
- Indie drama ‘The Farewell’ beats ‘Avengers: Endgame’ box office record Today 7:44 AM
- Join the Unicorn Moms if you’re sick of being judged by other parents online Today 7:00 AM
- Get the Amazon Fire TV Stick for just $15 on Prime Day Today 6:51 AM
Will eye-controlled prosthetics enable the next wave of human productivity?
Using nothing more than blinks and glances, it’s possible to manipulate a robotic arm well enough to paint a simple picture.
Dr. Aldo Faisal of Imperial College London’s Departments of Computing and Bioengineering successfully demonstrated a system that lets a human user control a paintbrush-wielding robot arm using only his or her eyes. He suggests this might lay the foundation for humanity’s future of robot-enabled multitasking.
Using technology that tracks the user’s eyes, the robot is able to receive and understand a variety of commands, even to clean the paintbrush and prepare it with a different color. Once the robot is holding a paint-dabbed brush, it’s apparently pretty intuitive to use. Sabine Dziemian, a postgraduate in Faisal’s research group, says, “If I want to draw a straight line, I look at the start point and the end point, and the robot moves the brush across that line.”
Blinking three times puts the robot in color selection mode, in which it moves the brush over to a variety of pre-dispensed colors. At that point, the user only needs to look at the color he or she wants to use next, and the arm applies the color to the brush.
“Since time immemorial, human imagination has sparked the idea of having additional arms,” says Faisal. He invokes the multiarmed Hindu goddess Shiva, often a symbol of transformation, to suggest we might one day “do the dishes while taking a phone call or [holding] your baby and [preparing] the food at the same time, because you have just that extra pair of hands attached to you.”
The future of robotic prosthetics is developing rapidly. Last month, researchers at Sandia National Laboratories made significant progress in wiring a prosthetic device to the human brain, while leading manufacturer Ossur has developed a battery-powered knee that optimizes movements based on the wearer’s natural gait.
Taken as a whole, these developments could not only help make able-bodied humans more productive, but they also have immediately obvious potential for the disabled, who might use remotely controlled robotic arms to maintain more independence.
“If you’re controlling a third arm, you want it to be natural, seamless, and you don’t want to think about it,” Faisal says.
Dylan Love is an editorial consultant and journalist whose reporting interests include emergent technology, digital media, and Russian language and culture. He is a former staff writer for the Daily Dot, and his work has been published by Business Insider, International Business Times, Men's Journal, and the Next Web.