Moor: Model-based offline policy optimization with a risk dynamics model
Abstract Offline reinforcement learning (RL) has been widely used in safety-critical domains by avoiding dangerous and costly online interaction. A significant challenge is addressing uncertainties and risks outside of offline data. Risk-sensitive offline RL attempts to solve this issue by risk aver...
Saved in:
Main Authors: | Xiaolong Su, Peng Li, Shaofei Chen |
---|---|
Format: | Article |
Language: | English |
Published: |
Springer
2024-11-01
|
Series: | Complex & Intelligent Systems |
Subjects: | |
Online Access: | https://doi.org/10.1007/s40747-024-01621-x |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Stealthy data poisoning attack method on offline reinforcement learning in unmanned systems
by: ZHOU Xue, et al.
Published: (2024-12-01) -
Reinforcement Learning-Based Autonomous Soccer Agents: A Study in Multi-Agent Coordination and Strategy Development
by: Biplov Paneru, et al.
Published: (2025-01-01) -
HPRS: hierarchical potential-based reward shaping from task specifications
by: Luigi Berducci, et al.
Published: (2025-02-01) -
Tactical intent-driven autonomous air combat behavior generation method
by: Xingyu Wang, et al.
Published: (2024-12-01) -
Reinforcement learning for deep portfolio optimization
by: Ruyu Yan, et al.
Published: (2024-09-01)