CSE510 Deep Reinforcement Learning (Lecture 23)
Due to lack of my attention, this lecture note is generated by ChatGPT to create continuations of the previous lecture note.
Offline Reinforcement Learning Part II: Advanced Approaches
Lecture 23 continues with advanced topics in offline RL, expanding on model-based imagination methods and credit assignment structures relevant for offline multi-agent and single-agent settings.
Reverse Model-Based Imagination (ROMI)
ROMI is a method for augmenting an offline dataset with additional transitions generated by imagining trajectories -backwards- from desirable states. Unlike forward model rollouts, backward imagination stays grounded in real data because imagined transitions always terminate in dataset states.
Reverse Dynamics Model
ROMI learns a reverse dynamics model:
Parameter explanations:
- : learned reverse transition model.
- : parameter vector for the reverse model.
- : next state (from dataset).
- : action that hypothetically leads into .
- : predicted predecessor state.
ROMI also learns a reverse policy to sample actions that likely lead into known states:
Parameter explanations:
- : reverse policy distribution.
- : action sampled for backward trajectory generation.
- : state whose predecessors are being imagined.
Reverse Imagination Process
To generate imagined transitions:
- Select a goal or high-value state from the offline dataset.
- Sample from .
- Predict from .
- Form an imagined transition .
- Repeat backward to obtain a longer imagined trajectory.
Benefits:
- Imagined states remain grounded by terminating in real dataset states.
- Helps propagate reward signals backward through states not originally visited.
- Avoids runaway model error that occurs in forward model rollouts.
ROMI effectively fills in missing gaps in the state-action graph, improving training stability and performance when paired with conservative offline RL algorithms.
Implicit Credit Assignment via Value Factorization Structures
Although initially studied for multi-agent systems, insights from value factorization also improve offline RL by providing structured credit assignment signals.
Counterfactual Credit Assignment Insight
A factored value function structure of the form:
can implicitly implement counterfactual credit assignment.
Parameter explanations:
- : global value function.
- : individual component value for agent or subsystem .
- : mixing function combining components.
- : environment state.
- : action taken by entity .
In architectures designed for IGM (Individual-Global-Max) consistency, gradients backpropagated through isolate the marginal effect of each component. This implicitly gives each agent or subsystem a counterfactual advantage signal.
Even in single-agent structured RL, similar factorization structures allow credit flowing into components representing skills, modes, or action groups, enabling better temporal and structural decomposition.
Model-Based vs Model-Free Offline RL
Lecture 23 contrasts model-based imagination (ROMI) with conservative model-free methods such as IQL and CQL.
Forward Model-Based Rollouts
Forward imagination using a learned model:
Parameter explanations:
- : learned forward dynamics model.
- : parameters of the forward model.
- : predicted next state.
- : current state.
- : action taken in current state.
Problems:
- Forward rollouts drift away from dataset support.
- Model error compounds with each step.
- Leads to training instability if used without penalties.
Penalty Methods (MOPO, MOReL)
Augmented reward:
Parameter explanations:
- : penalized reward for model-generated steps.
- : uncertainty score of model for state-action pair.
- : penalty coefficient.
- : original reward.
These methods limit exploration into uncertain model regions.
ROMI vs Forward Rollouts
- Forward methods expand state space beyond dataset.
- ROMI expands -backward-, staying consistent with known good future states.
- ROMI reduces error accumulation because future anchors are real.
Combining ROMI With Conservative Offline RL
ROMI is typically combined with:
- CQL (Conservative Q-Learning)
- IQL (Implicit Q-Learning)
- BCQ and BEAR (policy constraint methods)
Workflow:
- Generate imagined transitions via ROMI.
- Add them to dataset.
- Train Q-function or policy using conservative losses.
Benefits:
- Better coverage of reward-relevant states.
- Increased policy improvement over dataset.
- More stable Q-learning backups.
Summary of Lecture 23
Key points:
- Offline RL can be improved via structured imagination.
- ROMI creates safe imagined transitions by reversing dynamics.
- Reverse imagination avoids pitfalls of forward model error.
- Factored value structures provide implicit counterfactual credit assignment.
- Combining ROMI with conservative learners yields state-of-the-art performance.