New data published to support safe operation of autonomous vehicles in adverse weather
The Heriot-Watt University-led Radiate project has published a new dataset that will support the development of autonomous vehicles that can operate safely in adverse weather conditions.
Credit: Heriot-Watt University
Sensors that operate in Scotland’s rain, snow and fog are providing data that could help autonomous vehicles to see and operate safely in adverse weather.
The Radiate project – led by Edinburgh’s Heriot-Watt University – has published a new dataset that includes three hours of radar images and 200,000 tagged road actors, including other vehicles and pedestrians. The dataset solves a problem that has been facing manufacturers and researchers of autonomous vehicles.
Until now, almost all of the available, labelled data has been based on sunny, clear days. This meant that there was no public data available to help develop autonomous vehicles that can operate safely in adverse weather conditions. It has also relied primarily on data collected from optical sensors, which, much like human vision, don’t work as well during bad weather.
Heriot-Watt University’s Professor Andrew Wallace and Dr Sen Wang have been collecting the data since 2019, when they equipped a van with light detection and ranging (LiDAR), radar and stereo cameras and geopositioning devices. They drove the car around Edinburgh and the Scottish Highlands to capture urban and rural roads at all times of day and night, purposefully chasing bad weather.
Professor Wallace said: “Datasets are essential to developing and benchmarking perception systems for autonomous vehicles. We’re many years from driverless cars being on the streets, but autonomous vehicles are already being used in controlled circumstances or piloting areas. We’ve shown that radar can help autonomous vehicles to navigate, map and interpret their environment in bad weather, when vision and LiDAR can fail.”
The team says that, by labelling all of the objects that their system spotted on the roads, they have provided another step forward for researchers and manufacturers.
Dr Sen Wang said: “We labelled over 200,000 road objects in our dataset – bicycles, cars, pedestrians, traffic signs and other road actors. We could use this data to help autonomous vehicles to predict the future and navigate safely. When a car pulls out in front of you, you try to predict what it will do – will it swerve, will it take off? That’s what autonomous vehicles will have to do, and now we have a database that can put them on that path, even in bad weather.”
The team is based at Heriot-Watt University’s Institute of Sensors, Signals and Systems, which has already developed classical and deep learning approaches to interpreting sensory data. They say that their ultimate goal is to improve perception capability.
Wallace said: “We need to improve the resolution of the radar, which is naturally fuzzy. If we can combine high-resolution optical images with the weather-penetrating capability of enhanced radar, that takes us closer to autonomous vehicles being able to see and map better and, ultimately, navigate more safely.”