Poster
QPRL : Learning Optimal Policies with Quasi-Potential Functions for Asymmetric Traversal
Jumman Hossain · Nirmalya Roy
West Exhibition Hall B2-B3 #W-700
Many real-world tasks—like a delivery robot climbing steep streets or a rescue drone squeezing through one-way passages—cost much more effort in one direction than the other, and some actions cannot be undone at all. Today’s learning algorithms usually ignore this imbalance, so they may choose routes that waste energy or put the robot in a trap. Our research introduces Quasi-Potential Reinforcement Learning (QPRL), a method that lets an agent recognize and plan around these one-way or “uneven-effort” situations. QPRL splits the cost of every move into two parts: a reusable “potential” map (showing, for example, how steep a hill is) and an extra penalty for truly irreversible steps (such as dropping off a ledge). A built-in safety check ensures the agent never takes a step whose future cost could exceed a small, user-set limit. Across several simulated navigation and control tasks, QPRL learns faster, reaches goals more reliably, and commits about four times fewer irreversible errors than existing approaches. These ideas could help future robots and self-driving vehicles to act more safely and efficiently whenever undoing a decision is hard or impossible.
Live content is unavailable. Log in and register to view live content