ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events
Autonomous vehicles are advanced driving systems that revolutionize transportation, but their vulnerability to adversarial attacks poses significant safety risks. Consider a scenario in which a slight perturbation in sensor data causes an autonomous vehicle to fail unexpectedly, potentially leading...
Saved in:
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2024-01-01
|
Series: | IEEE Open Journal of Intelligent Transportation Systems |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10714436/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832590360290787328 |
---|---|
author | Aizaz Sharif Dusica Marijan |
author_facet | Aizaz Sharif Dusica Marijan |
author_sort | Aizaz Sharif |
collection | DOAJ |
description | Autonomous vehicles are advanced driving systems that revolutionize transportation, but their vulnerability to adversarial attacks poses significant safety risks. Consider a scenario in which a slight perturbation in sensor data causes an autonomous vehicle to fail unexpectedly, potentially leading to accidents. Current testing methods often rely on computationally expensive active learning techniques to identify such vulnerabilities. Rather than actively training complex adversaries by interacting with the environment, there is a need to first intelligently find and reduce the search space to only those states where autonomous vehicles are found to be less confident. In this paper, we propose a black-box testing framework ReMAV that uses offline trajectories first to efficiently identify weaknesses of autonomous vehicles without the need for active interaction. To this end, we introduce a three-step methodology which i) uses offline state action pairs of any autonomous vehicle under test, ii) builds an abstract behavior representation using our designed reward modeling technique to analyze states with uncertain driving decisions, and iii) uses a disturbance model for minimal perturbation attacks where the driving decisions are less confident. Our reward modeling creates a behavior representation that highlights regions of likely uncertain autonomous vehicle behavior, even when performance seems adequate. This enables efficient testing without computationally expensive active adversarial learning. We evaluated ReMAV in a high-fidelity urban driving simulator across various single- and multi-agent scenarios. The results show substantial increases in failure events compared to the standard behavior of autonomous vehicles: 35% in vehicle collisions, 23% in road object collisions, 48% in pedestrian collisions, and 50% in off-road steering events. ReMAV outperforms two baselines and previous testing frameworks in effectiveness, efficiency, and speed of identifying failures. This demonstrates ReMAV’s capability to efficiently expose autonomous vehicle weaknesses using simple perturbation models. |
format | Article |
id | doaj-art-d4d6bc308f054a9dbcde440653cd283c |
institution | Kabale University |
issn | 2687-7813 |
language | English |
publishDate | 2024-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Open Journal of Intelligent Transportation Systems |
spelling | doaj-art-d4d6bc308f054a9dbcde440653cd283c2025-01-24T00:02:53ZengIEEEIEEE Open Journal of Intelligent Transportation Systems2687-78132024-01-01566969110.1109/OJITS.2024.347909810714436ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure EventsAizaz Sharif0https://orcid.org/0000-0003-1860-813XDusica Marijan1https://orcid.org/0000-0001-9345-5431Department of VIAS, Simula Research Laboratory, Oslo, NorwayDepartment of VIAS, Simula Research Laboratory, Oslo, NorwayAutonomous vehicles are advanced driving systems that revolutionize transportation, but their vulnerability to adversarial attacks poses significant safety risks. Consider a scenario in which a slight perturbation in sensor data causes an autonomous vehicle to fail unexpectedly, potentially leading to accidents. Current testing methods often rely on computationally expensive active learning techniques to identify such vulnerabilities. Rather than actively training complex adversaries by interacting with the environment, there is a need to first intelligently find and reduce the search space to only those states where autonomous vehicles are found to be less confident. In this paper, we propose a black-box testing framework ReMAV that uses offline trajectories first to efficiently identify weaknesses of autonomous vehicles without the need for active interaction. To this end, we introduce a three-step methodology which i) uses offline state action pairs of any autonomous vehicle under test, ii) builds an abstract behavior representation using our designed reward modeling technique to analyze states with uncertain driving decisions, and iii) uses a disturbance model for minimal perturbation attacks where the driving decisions are less confident. Our reward modeling creates a behavior representation that highlights regions of likely uncertain autonomous vehicle behavior, even when performance seems adequate. This enables efficient testing without computationally expensive active adversarial learning. We evaluated ReMAV in a high-fidelity urban driving simulator across various single- and multi-agent scenarios. The results show substantial increases in failure events compared to the standard behavior of autonomous vehicles: 35% in vehicle collisions, 23% in road object collisions, 48% in pedestrian collisions, and 50% in off-road steering events. ReMAV outperforms two baselines and previous testing frameworks in effectiveness, efficiency, and speed of identifying failures. This demonstrates ReMAV’s capability to efficiently expose autonomous vehicle weaknesses using simple perturbation models.https://ieeexplore.ieee.org/document/10714436/Autonomous vehicle testingdeep reinforcement learningbehavior modelinginverse reinforcement learning |
spellingShingle | Aizaz Sharif Dusica Marijan ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events IEEE Open Journal of Intelligent Transportation Systems Autonomous vehicle testing deep reinforcement learning behavior modeling inverse reinforcement learning |
title | ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events |
title_full | ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events |
title_fullStr | ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events |
title_full_unstemmed | ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events |
title_short | ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events |
title_sort | remav reward modeling of autonomous vehicles for finding likely failure events |
topic | Autonomous vehicle testing deep reinforcement learning behavior modeling inverse reinforcement learning |
url | https://ieeexplore.ieee.org/document/10714436/ |
work_keys_str_mv | AT aizazsharif remavrewardmodelingofautonomousvehiclesforfindinglikelyfailureevents AT dusicamarijan remavrewardmodelingofautonomousvehiclesforfindinglikelyfailureevents |