Lazy-MDPs: Towards Interpretable RL by Learning When to Act. | 0 | 0.34 | 2022 |
Continuous Control with Action Quantization from Demonstrations. | 0 | 0.34 | 2022 |
A general class of surrogate functions for stable and efficient reinforcement learning | 0 | 0.34 | 2022 |
Generalization in Mean Field Games by Learning Master Policies. | 0 | 0.34 | 2022 |
Implicitly Regularized RL with Implicit Q-values | 0 | 0.34 | 2022 |
Scaling Mean Field Games by Online Mirror Descent. | 0 | 0.34 | 2022 |
Offline Reinforcement Learning as Anti-exploration. | 0 | 0.34 | 2022 |
Concave Utility Reinforcement Learning: The Mean-field Game Viewpoint. | 0 | 0.34 | 2022 |
Scalable Deep Reinforcement Learning Algorithms for Mean Field Games. | 0 | 0.34 | 2022 |
Offline Reinforcement Learning With Pseudometric Learning | 0 | 0.34 | 2021 |
Adversarially Guided Actor-Critic. | 0 | 0.34 | 2021 |
Show me the Way: Intrinsic Motivation from Demonstrations | 0 | 0.34 | 2021 |
How To Train Your Heron | 0 | 0.34 | 2021 |
Mean Field Games Flock! The Reinforcement Learning Way. | 0 | 0.34 | 2021 |
Adversarially Guided Actor-Critic | 0 | 0.34 | 2021 |
Self-Imitation Advantage Learning | 0 | 0.34 | 2021 |
What Matters for On-Policy Deep Actor-Critic Methods? A Large-Scale Study | 0 | 0.34 | 2021 |
Hyperparameter Selection for Imitation Learning | 0 | 0.34 | 2021 |
Evaluation Of Prioritized Deep System Identification On A Path Following Task | 0 | 0.34 | 2021 |
Learning Behaviors through Physics-driven Latent Imagination. | 0 | 0.34 | 2021 |
What Matters for Adversarial Imitation Learning? | 0 | 0.34 | 2021 |
Primal Wasserstein Imitation Learning | 0 | 0.34 | 2021 |
What Matters for On-Policy Deep Actor-Critic Methods? A Large-Scale Study. | 0 | 0.34 | 2021 |
Primal Wasserstein Imitation Learning. | 0 | 0.34 | 2021 |
Fictitious Play for Mean Field Games: Continuous Time Analysis and Applications | 0 | 0.34 | 2020 |
On The Convergence Of Model Free Learning In Mean Field Games | 0 | 0.34 | 2020 |
Self-Attentional Credit Assignment for Transfer in Reinforcement Learning | 1 | 0.36 | 2020 |
Foolproof Cooperative Learning. | 0 | 0.34 | 2020 |
Image-Based Place Recognition on Bucolic Environment Across Seasons From Semantic Edge Description. | 0 | 0.34 | 2020 |
CopyCAT:: Taking Control of Neural Policies with Constant Attacks | 0 | 0.34 | 2020 |
Leverage the Average: an Analysis of KL Regularization in Reinforcement Learning | 0 | 0.34 | 2020 |
Munchausen Reinforcement Learning | 0 | 0.34 | 2020 |
Deep Conservative Policy Iteration | 0 | 0.34 | 2019 |
Stable and Efficient Policy Evaluation. | 1 | 0.41 | 2019 |
Deep Reinforcement Learning-Based Continuous Control For Multicopter Systems | 0 | 0.34 | 2019 |
Targeted Attacks on Deep Reinforcement Learning Agents through Adversarial Observations. | 0 | 0.34 | 2019 |
A Theory of Regularized Markov Decision Processes. | 0 | 0.34 | 2019 |
Learning from a Learner | 0 | 0.34 | 2019 |
Image-Based Text Classification using 2D Convolutional Neural Networks | 0 | 0.34 | 2019 |
Foolproof Cooperative Learning. | 0 | 0.34 | 2019 |
Importance Sampling for Deep System Identification | 0 | 0.34 | 2019 |
A Deep Learning Approach For Privacy Preservation In Assisted Living | 1 | 0.35 | 2018 |
Human Activity Recognition Using Recurrent Neural Networks. | 10 | 0.62 | 2018 |
Image-based Natural Language Understanding Using 2D Convolutional Neural Networks. | 0 | 0.34 | 2018 |
Anderson Acceleration for Reinforcement Learning. | 0 | 0.34 | 2018 |
Deep Representation Learning for Domain Adaptation of Semantic Image Segmentation. | 0 | 0.34 | 2018 |
Reconstruct & Crush Network. | 0 | 0.34 | 2017 |
Is the Bellman residual a bad proxy? | 0 | 0.34 | 2017 |
Bridging the Gap Between Imitation Learning and Inverse Reinforcement Learning. | 7 | 0.51 | 2017 |
Should one minimize the expected Bellman residual or maximize the mean value? | 0 | 0.34 | 2016 |