Indian origin team develops model for safer self-driving cars

SAN FRANCISCO — A team of Indian American researchers has developed a novel model that uses human inputs to uncover Artificial Intelligence (AI) "blind spots" in self-driving cars, so that the vehicles can avoid dangerous errors in the real world.

The model developed by MIT and Microsoft researchers identifies instances in which autonomous systems have "learned" from training examples that don't match what's actually happening in the real world.

Engineers could use this model to improve the safety of AI systems, such as driverless vehicles and autonomous robots.

"The model helps autonomous systems better know what they don't know," said first author Ramya Ramakrishnan from Computer Science and Artificial Intelligence Laboratory at MIT.

"Many times, when these systems are deployed, their trained simulations don't match the real-world setting [and] they could make mistakes, such as getting into accidents.

"The idea is to use humans to bridge that gap between simulation and the real world, in a safe way, so we can reduce some of those errors," explained Ramakrishnan.

Indian origin team develops model for safer self-driving cars

Ramya Ramakrishnan 

The AI systems powering driverless cars are trained extensively in virtual simulations to prepare the vehicle for nearly every event on the road.

But sometimes the car makes an unexpected error in the real world because an event occurs that should, but doesn't, alter the car's behavior.

The researchers validated their method using video games, with a simulated human correcting the learned path of an on-screen character.

The next step is to incorporate the model with traditional training and testing approaches for autonomous cars and robots with human feedback.

Co-authors on the papers are Julie Shah, an associate professor in the Department of Aeronautics and Astronautics and head of the CSAIL's Interactive Robotics Group; and Ece Kamar, Debadeepta Dey, and Eric Horvitz -- all from Microsoft Research.

"When the system is deployed into the real world, it can use learned model to act more cautiously and intelligently," said Ramakrishnan.

— Indo-Asian News Service 

(0) comments

Welcome to the discussion.

Keep it Clean. Please avoid obscene, vulgar, lewd, racist or sexually-oriented language.
PLEASE TURN OFF YOUR CAPS LOCK.
Don't Threaten. Threats of harming another person will not be tolerated.
Be Truthful. Don't knowingly lie about anyone or anything.
Be Nice. No racism, sexism or any sort of -ism that is degrading to another person.
Be Proactive. Use the 'Report' link on each comment to let us know of abusive posts.
Share with Us. We'd love to hear eyewitness accounts, the history behind an article.