dc.contributor.author | Marzari, Luca | |
dc.contributor.author | Cicalese, Ferdinando | |
dc.contributor.author | Farinelli, Alessandro | |
dc.contributor.author | Amato, Christopher | |
dc.contributor.author | Marchesini, Enrico | |
dc.date.accessioned | 2025-10-03T18:57:36Z | |
dc.date.available | 2025-10-03T18:57:36Z | |
dc.date.issued | 2025-09-30 | |
dc.identifier.issn | 2157-6904 | |
dc.identifier.uri | https://hdl.handle.net/1721.1/162893 | |
dc.description.abstract | Ensuring safety in reinforcement learning (RL) is critical for deploying agents in real-world applications. During training, current safe RL approaches often rely on indicator cost functions that provide sparse feedback, resulting in two key limitations: (i) poor sample efficiency due to the lack of safety information in neighboring states, and (ii) dependence on cost-value functions, leading to brittle convergence and suboptimal performance. After training, safety is guaranteed via formal verification methods for deep neural networks (FV), whose computational complexity hinders their application during training. We address the limitations of using cost functions via verification by proposing a safe RL method based on a violation value---the risk associated with policy decisions in a portion of the state space. Our approach verifies safety properties (i.e., state-action pairs) that may lead to unsafe behavior, and quantifies the size of the state space where properties are violated. This violation value is then used to penalize the agent during training to encourage safer policy behavior. Given the NP-hard nature of FV, we propose an efficient, sample-based approximation with probabilistic guarantees to compute the violation value. Extensive experiments on standard benchmarks and real-world robotic navigation tasks show that violation-augmented approaches significantly improve safety by reducing the number of unsafe states encountered while achieving superior performance compared to existing methods. | en_US |
dc.publisher | ACM | en_US |
dc.relation.isversionof | http://dx.doi.org/10.1145/3770068 | en_US |
dc.rights | Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. | en_US |
dc.source | Association for Computing Machinery | en_US |
dc.title | Verifying Online Safety Properties for Safe Deep Reinforcement Learning | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Luca Marzari, Ferdinando Cicalese, Alessandro Farinelli, Christopher Amato, and Enrico Marchesini. 2025. Verifying Online Safety Properties for Safe Deep Reinforcement Learning. ACM Trans. Intell. Syst. Technol. | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Laboratory for Information and Decision Systems | en_US |
dc.relation.journal | ACM Transactions on Intelligent Systems and Technology | en_US |
dc.identifier.mitlicense | PUBLISHER_POLICY | |
dc.eprint.version | Final published version | en_US |
dc.type.uri | http://purl.org/eprint/type/JournalArticle | en_US |
eprint.status | http://purl.org/eprint/status/PeerReviewed | en_US |
dc.date.updated | 2025-10-01T07:58:03Z | |
dc.language.rfc3066 | en | |
dc.rights.holder | The author(s) | |
dspace.date.submission | 2025-10-01T07:58:04Z | |
mit.license | PUBLISHER_POLICY | |
mit.metadata.status | Authority Work and Publication Information Needed | en_US |