Human vision inspired supercomputer may drive cars someday
Washington, Sep 16 (ANI): Inspired by our remarkable ability to quickly navigate through the street recognizing cars, other people, trees and lampposts instantaneously, researchers at Yale have developed a supercomputer based on the human visual system that could one day operate cars on its own.
Dubbed NeuFlow, the system, developed by Eugenio Culurciello of Yale's School of Engineering and Applied Science, takes its inspiration from the mammalian visual system, mimicking its neural network to quickly interpret the world around it.
The system uses complex vision algorithms developed by Yann LeCun at New York University to run large neural networks for synthetic vision applications. One idea-the one Culurciello and LeCun are focusing on, is a system that would allow cars to drive themselves.
In order to be able to recognize the various objects encountered on the road-such as other cars, people, stoplights, sidewalks, not to mention the road itself-NeuFlow processes tens of megapixel images in real time.
The system is also extremely efficient, simultaneously running more than 100 billion operations per second using only a few watts (that's less than the power a cell phone uses) to accomplish what it takes bench-top computers with multiple graphic processors more than 300 watts to achieve.
"One of our first prototypes of this system is already capable of outperforming graphic processors on vision tasks," said Culurciello.
Culurciello embedded the supercomputer on a single chip, making the system much smaller, yet more powerful and efficient, than full-scale computers.
"The complete system is going to be no bigger than a wallet, so it could easily be embedded in cars and other places," said Culurciello.
Beyond the autonomous car navigation, the system could be used to improve robot navigation into dangerous or difficult-to-reach locations, to provide 360-degree synthetic vision for soldiers in combat situations, or in assisted living situations where it could be used to monitor motion and call for help should an elderly person fall, for example.
The study was presented at the High Performance Embedded Computing (HPEC) workshop in Boston, Mass. (ANI)
-
India vs New Zealand T20 World Cup 2026 Final: Five Positive Signs Favouring India Before Title Clash -
IND vs NZ Final Live: When and Where to Watch India vs New Zealand T20 World Cup 2026 Title Clash -
Ind vs NZ T20 World Cup 2026: New Zealand Needs 256 Runs To Beat India And Win The World Cup -
UAE Attacks Iran, Becomes 5th Nation To Enter War; Reports Suggest Strike On Iranian Facility -
ICC T20 World Cup 2026 Final: Ricky Martin, Falguni Pathak To Perform At Closing Ceremony, How To Watch -
Who Is Nishant Kumar: Education, Personal Life and Possible Political Role -
IND vs NZ T20 WC Final: New Zealand Win Toss, Opt To Chase; Why Batting First Could Be A Tough Call For India -
Gold Rate Today 8 March 2026: IBJA Issues Fresh Gold Rates; Tanishq, Malabar, Kalyan, Joyalukkas Prices -
From Kerala Boy To World Cup Hero: Sanju Samson’s 89-Run Blitz, His Birth, Religion, Wife And Inspiring Story -
Hyderabad Gold Silver Rate Today, 8 March, 2026: Latest Gold Prices And Silver Rate In Nizam City -
Panauti Stadium? Is Narendra Modi Stadium an Unlucky Venue for India National Cricket Team? -
Storm Over West Bengal Govt's 'Snub' To President Droupadi Murmu












Click it and Unblock the Notifications