π Model-based offline RL methods are a good fit for offline RL as they allow training a model on available data and using it to obtain a good policy or plan directly.
β In model-based RL, the trained model is used to answer "what if" questions about different states and actions.
βοΈ Dyna-style methods are adapted to the offline setting to simulate rollouts starting from the collected states and actions.
β One challenge in offline RL is the policy learning to exploit the model by tricking it into going into high-reward out-of-distribution states.
π§ Modifying model-based methods to penalize the policy when it tricks the model into crazy states can incentivize the policy to stay closer to the data.
π Mobile model-based offline policy optimization modifies the reward function to impose a penalty for exploiting the model.
π‘ The uncertainty penalty quantifies how wrong the model is and punishes the policy enough to discourage exploitation.
βοΈ Using model uncertainty techniques, such as training an ensemble of models, helps measure the degree of disagreement among models.
Ensemble disagreement is a common choice for obtaining error metrics in offline reinforcement learning.
Two assumptions are required for accurate estimation of the model error and value function.
The learned policy in offline reinforcement learning can be guaranteed to perform at least as well as the best policy optimized against a reward-minus-error objective.
The best policy is one that avoids states where the model may be incorrect.
The learned policy is at least as good as the behavior policy, considering the model's error.
If the model accurately represents the optimal policy, the learned policy can be close to optimal.
π Using data from the model, the critic's loss function in offline reinforcement learning is designed to balance the q values of the model and the data set.
π² Dyna-style algorithms such as CQL and MORAL aim to improve offline reinforcement learning by making the model-based states and actions look worse than the data-based ones.
π The trajectory transformer method in offline reinforcement learning trains a model over entire trajectories to estimate the distribution of state-action sequences and optimizes planning based on high-probability actions.
π Using a large and expressive model class, like a transformer, is convenient for offline reinforcement learning.
π To model multi-modal distributions, the trajectory is discretized per dimension of every state and action.
β²οΈ By modeling state and action probabilities, accurate predictions can be made for longer horizons.
Using trajectory transformer to make predictions for humanoid future steps.
Utilizing beam search to maximize reward in planning.
Generating high probability trajectories to avoid out-of-distribution states and actions.
Habits That Got Me Thru My Teens (important)
Life Lessons I Learned at 30, to Avoid Regrets When I'm 60
"okay, but I want GPT to perform 10x for my specific use case" - Here is how
"Next Level Prompts?" - 10 mins into advanced prompting
Scraping Unleashed: Bright Dataβs Scraping Browser Product Launch & Demonstration
Jurassic Park T-Rex sound design explained by Gary Rydstrom