Site icon TechNewsBoy.com

Location key to improved autonomous vehicle vision

A data visualization showing two Ford self-driving research vehicles driving through Dearborn, MI. Credit: Ford Motor Company

QUT robotics researchers working with Ford Motor Company have found a way to tell an autonomous vehicle which cameras to use when navigating.

Professor Michael Milford, Joint Director of the QUT Center for Robotics and Australian Research Council Laureate Fellow and senior author, said the research comes from a project looking at how cameras and LIDAR sensors, commonly used in autonomous vehicles, can better understand the world around them.

“The key idea here is to learn which cameras to use at different locations in the world, based on previous experience at that location,” Professor Milford said.

“For example, the system might learn that a particular camera is very useful for tracking the position of the vehicle on a particular stretch of road, and choose to use that camera on subsequent visits to that section of road.”

Dr. Punarjay (Jay) Chakravarty is leading the project on behalf of the Ford Autonomous Vehicle Future Tech group.

“Autonomous vehicles depend heavily on knowing where they are in the world, using a range of sensors including cameras,” says Dr. Chakravarty.

“Knowing where you are helps you leverage map information that is also useful for detecting other dynamic objects in the scene. A particular intersection might have people crossing in a certain way.

“This can be used as prior information for the neural nets doing object detection and so accurate localization is critical and this research allows us to focus on the best camera at any given time.”

To make progress on the problem, the team has also had to devise new ways of evaluating the performance of an autonomous vehicle positioning system.

Joint lead researcher Dr. Stephen Hausler said, “We’re focusing not just on how the system performs when it’s doing well, but what happens in the worst-case scenario.”

This research took place as part of a larger fundamental research project with Ford looking at how cameras and LIDAR sensors, commonly used in autonomous vehicles, can better understand the world around them.

This work has just been published in IEEE Robotics and Automation Letters.


New AI creates bird’s-eye view map faster, brings safer autonomous vehicles a step closer


More information:
Stephen Hausler et al, Improving Worst Case Visual Localization Coverage via Place-Specific Sub-Selection in Multi-Camera Systems, IEEE Robotics and Automation Letters (2022). DOI: 10.1109/LRA.2022.3191174
Provided by
Queensland University of Technology


Citation:
Location key to improved autonomous vehicle vision (2022, August 8)
retrieved 8 August 2022
from https://techxplore.com/news/2022-08-key-autonomous-vehicle-vision.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.

For all the latest Technology News Click Here 

 For the latest news and updates, follow us on Google News

Read original article here

Denial of responsibility! TechNewsBoy.com is an automatic aggregator around the global media. All the content are available free on Internet. We have just arranged it in one platform for educational purpose only. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials on our website, please contact us by email – abuse@technewsboy.com. The content will be deleted within 24 hours.
Exit mobile version