Teleoperating Robots With Virtual Reality
October 12, 2017 | MITEstimated reading time: 4 minutes
Certain industries have traditionally not had the luxury of telecommuting. Many manufacturing jobs, for example, require a physical presence to operate machinery.
But what if such jobs could be done remotely? Last week researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) presented a virtual reality (VR) system that lets you teleoperate a robot using an Oculus Rift headset.
The system embeds the user in a VR control room with multiple sensor displays, making it feel like they’re inside the robot’s head. By using hand controllers, users can match their movements to the robot’s movements to complete various tasks.
“A system like this could eventually help humans supervise robots from a distance,” says CSAIL postdoc Jeffrey Lipton, who was the lead author on a related paper about the system. “By teleoperating robots from home, blue-collar workers would be able to tele-commute and benefit from the IT revolution just as white-collars workers do now."
The researchers even imagine that such a system could help employ increasing numbers of jobless video-gamers by “gameifying” manufacturing positions.
The team used the Baxter humanoid robot from Rethink Robotics, but said that it can work on other robot platforms and is also compatible with the HTC Vive headset.
Lipton co-wrote the paper with CSAIL Director Daniela Rus and researcher Aidan Fay. They presented the paper at the recent IEEE/RSJ International Conference on Intelligent Robots and Systems in Vancouver.
There have traditionally been two main approaches to using VR for teleoperation.
In a direct model, the user's vision is directly coupled to the robot's state. With these systems, a delayed signal could lead to nausea and headaches, and the user’s viewpoint is limited to one perspective.
In a cyber-physical model, the user is separate from the robot. The user interacts with a virtual copy of the robot and the environment. This requires much more data, and specialized spaces.
The CSAIL team’s system is halfway between these two methods. It solves the delay problem, since the user is constantly receiving visual feedback from the virtual world. It also solves the the cyber-physical issue of being distinct from the robot: Once a user puts on the headset and logs into the system, they’ll feel as if they’re inside Baxter’s head.
The system mimics the homunculus model of the mind — the idea that there’s a small human inside our brains controlling our actions, viewing the images we see, and understanding them for us. While it’s a peculiar idea for humans, for robots it fits: Inside the robot is a human in a virtual control room, seeing through its eyes and controlling its actions.
Using Oculus’ controllers, users can interact with controls that appear in the virtual space to open and close the hand grippers to pick up, move, and retrieve items. A user can plan movements based on the distance between the arm’s location marker and their hand while looking at the live display of the arm.
To make these movements possible, the human’s space is mapped into the virtual space, and the virtual space is then mapped into the robot space to provide a sense of co-location.
The system is also more flexible compared to previous systems that require many resources. Other systems might extract 2-D information from each camera, build out a full 3-D model of the environment, and then process and redisplay the data. In contrast, the CSAIL team’s approach bypasses all of that by simply taking the 2-D images that are displayed to each eye. (The human brain does the rest by automatically inferring the 3-D information.)
To test the system, the team first teleoperated Baxter to do simple tasks like picking up screws or stapling wires. They then had the test users teleoperate the robot to pick up and stack blocks.
Users successfully completed the tasks at a much higher rate compared to the direct model. Unsurprisingly, users with gaming experience had much more ease with the system.
Tested against current state-of-the-art systems, CSAIL’s system was better at grasping objects 95 percent of the time and 57 percent faster at doing tasks. The team also showed that the system could pilot the robot from hundreds of miles away; testing included controling Baxter at MIT from a hotel’s wireless network in Washington.
"This contribution represents a major milestone in the effort to connect the user with the robot's space in an intuitive, natural, and effective manner." says Oussama Khatib, a computer science professor at Stanford University who was not involved in the paper.
The team eventually wants to focus on making the system more scalable, with many users and different types of robots that can be compatible with current automation technologies.
The project was funded, in part, by the Boeing Company and the National Science Foundation.
Subscribe
Stay ahead of the technologies shaping the future of electronics with our latest newsletter, Advanced Electronics Packaging Digest. Get expert insights on advanced packaging, materials, and system-level innovation, delivered straight to your inbox.Subscribe now to stay informed, competitive, and connected.
Suggested Items
Neusoft Smart Go, Tencent Cloud Partner on AI-Powered Intelligent Cockpit Ecosystem
05/01/2026 | PRNewswireThe company now aims to become a global leading provider in full-domain upper-body electronics solutions for intelligent vehicles.
Sikorsky, Robinson Unmanned Win U.S. Marine Corps Autonomous Logistics Contract
04/28/2026 | Lockheed MartinThe United States Marine Corps awarded a $15.5 million contract to Sikorsky, a Lockheed Martin company, for the Medium Aerial Resupply Vehicle – Expeditionary Logistics (MARV-EL) Increment 2 program.
Skoltech Team Develops Modulator for Compact Photonic Integrated Circuits
04/28/2026 | SkoltechResearchers at Skoltech have developed an ultra-compact electro-optic modulator based on silicon photonics and plasmonics that enables high-efficiency optical signal control within a small device footprint.
American Made Advocacy: Rebuilding America’s Military Stockpiles Begins With Microelectronics
04/28/2026 | Shane Whiteside -- Column: American Made AdvocacyCurrent world events demonstrate the fragility of long-distance supply chains transiting multiple zones of conflict. The U.S. military is currently drawing down supplies of key munitions and other electronic systems at unprecedented rates.1Every one of those systems is powered by printed circuit boards. The American PCB industry has kept pace with peacetime demand for the defense industry, but will now be called upon to increase production to a wartime footing at rates not seen in decades.
Global Electronics Association and FED Open Call for Abstracts for 2027 Pan-European Design Conference
04/30/2026 | Global Electronics AssociationThe German Electronics Design and Manufacturing Association (FED) and the Global Electronics Association are officially opening the Call for Abstracts for the Third Pan-European Electronics Design Conference (PEDC).