Limits of Actor-Critic Algorithms for Decision Tree Policies Learning in IBMDPs (2309.13365v3)
Abstract: Interpretability of AI models allows for user safety checks to build trust in such AIs. In particular, Decision Trees (DTs) provide a global look at the learned model and transparently reveal which features of the input are critical for making a decision. However, interpretability is hindered if the DT is too large. To learn compact trees, a recent Reinforcement Learning (RL) framework has been proposed to explore the space of DTs using deep RL. This framework augments a decision problem (e.g. a supervised classification task) with additional actions that gather information about the features of an otherwise hidden input. By appropriately penalizing these actions, the agent learns to optimally trade-off size and performance of DTs. In practice, a reactive policy for a partially observable Markov decision process (MDP) needs to be learned, which is still an open problem. We show in this paper that deep RL can fail even on simple toy tasks of this class. However, when the underlying decision problem is a supervised classification task, we show that finding the optimal tree can be cast as a fully observable Markov decision problem and be solved efficiently, giving rise to a new family of algorithms for learning DTs that go beyond the classical greedy maximization ones.
- POLITEX: Regret bounds for policy iteration using expert prediction. In Chaudhuri, K. and Salakhutdinov, R. (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp. 3692–3702. PMLR, 09–15 Jun 2019. URL https://proceedings.mlr.press/v97/lazic19a.html.
- Optimality and approximation with policy gradient methods in markov decision processes. In Conference on Learning Theory, pp. 64–66. PMLR, 2020.
- Continuous action reinforcement learning from a mixture of interpretable experts. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 2021.
- Explainable artificial intelligence (xai): Concepts, taxonomies, opportunities and challenges toward responsible ai. Information Fusion, 58:82–115, 2020. ISSN 1566-2535. doi: https://doi.org/10.1016/j.inffus.2019.12.012. URL https://www.sciencedirect.com/science/article/pii/S1566253519308103.
- Open problem: Approximate planning of pomdps in the class of memoryless policies. In Conference on Learning Theory, pp. 1639–1642. PMLR, 2016.
- Unbiased asymmetric actor-critic for partially observable reinforcement learning. CoRR, 2021.
- Model interpretability through the lens of computational complexity. In Neural Information Processing Systems (NeurIPS), 2020.
- Verifiable reinforcement learning via policy extraction. CoRR, 2018.
- Pruning decision trees with misclassification costs. In European Conference on Machine Learning, pp. 131–136. Springer, 1998.
- Classification and Regression Trees. Taylor & Francis, 1984. ISBN 9780412048418. URL https://books.google.fr/books?id=JwQx-WOmSyQC.
- Bubeck, S. Convex optimization: Algorithms and complexity. Found. Trends Mach. Learn., 2015.
- Extracting tree-structured representations of trained networks. In Advances in Neural Information Processing Systems (NIPS), 1996.
- Tree-based batch mode reinforcement learning. Journal of Machine Learning Research, 2005.
- Online markov decision processes. Mathematics of Operations Research, 34(3):726–736, 2009. ISSN 0364765X, 15265471. URL http://www.jstor.org/stable/40538442.
- A theory of regularized markov decision processes. In International Conference on Machine Learning, pp. 2160–2169. PMLR, 2019.
- A survey of methods for explaining black box models. ACM Comput. Surv., 2018.
- Improving robot controller transparency through autonomous policy explanation. In International Conference on Human-Robot Interaction (HRI), 2017.
- Convergence of stochastic iterative dynamic programming algorithms. Advances in neural information processing systems, 6, 1993.
- Reinforcement learning algorithm for partially observable markov decision problems. Advances in neural information processing systems, 7, 1994.
- Approximately optimal approx- imate reinforcement learnin. International Conference on Machine Learning (ICML), 2002.
- Bandit based monte-carlo planning. In Proceedings of the 17th European Conference on Machine Learning, ECML’06, pp. 282–293, Berlin, Heidelberg, 2006. Springer-Verlag. ISBN 354045375X. doi: 10.1007/11871842˙29. URL https://doi.org/10.1007/11871842_29.
- Imagenet classification with deep convolutional neural networks. In Pereira, F., Burges, C., Bottou, L., and Weinberger, K. (eds.), Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc., 2012. URL https://proceedings.neurips.cc/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf.
- Discovering symbolic policies with deep reinforcement learning. In International Conference on Machine Learning, pp. 5979–5989. PMLR, 2021.
- Interpretable classifiers using rules and bayesian analysis: Building a better stroke prediction model. The Annals of Applied Statistics, 2015.
- Lipton, Z. C. The mythos of model interpretability, 2016. URL https://arxiv.org/abs/1606.03490.
- Explainable reinforcement learning through a causal lens. In Conference on Artificial Intelligence (AAAI), 2019.
- On the global convergence rates of softmax policy gradient methods, 2020. URL https://arxiv.org/abs/2005.06392.
- A survey of explainable reinforcement learning, 2022.
- Human-level control through deep reinforcement learning. nature, 518(7540):529–533, 2015.
- Definitions, methods, and applications in interpretable machine learning. Proceedings of the National Academy of Sciences, 116(44):22071–22080, oct 2019. doi: 10.1073/pnas.1900654116. URL https://doi.org/10.1073%2Fpnas.1900654116.
- Learning When-to-Treat Policies. arXiv e-prints, 2019.
- Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
- Asymmetric actor critic for image-based robot learning. In Robotics: Science and Systems XIV, Carnegie Mellon University, Pittsburgh, Pennsylvania, USA, June 26-30, 2018, 2018.
- Cost complexity-based pruning of ensemble classifiers. Knowledge and Information Systems, 3(4):449–469, 2001.
- Puterman, M. L. Markov decision processes: discrete stochastic dynamic programming. John Wiley & Sons, 2014.
- Stable-baselines3: Reliable reinforcement learning implementations. Journal of Machine Learning Research, 2021.
- ”why should i trust you?”: Explaining the predictions of any classifier. In Knowledge Discovery and Data Mining (KDD), 2016.
- Interpretable reinforcement learning via differentiable decision trees. CoRR, abs/1903.09338, 2019.
- Trust Region Policy Optimization. International Conference on Machine Learning (ICML), 2015.
- Proximal policy optimization algorithms. CoRR, 2017.
- Markov decision processes in artificial intelligence. John Wiley & Sons, 2013.
- Mastering the game of go with deep neural networks and tree search. Nature, 529:484–503, 2016. URL http://www.nature.com/nature/journal/v529/n7587/full/nature16961.html.
- An efficient explanation of individual classifications using game theory. Journal of Machine Learning Resource (JMLR), 2010.
- Reinforcement learning: An introduction. MIT press, 2018.
- Iterative bounding mdps: Learning interpretable policies via non-interpretable methods. Proceedings of the AAAI Conference on Artificial Intelligence, 2021.
- Supersparse linear integer models for optimized medical scoring systems. Machine Learning, 2015.
- Attention is all you need. In Advances in Neural Information Processing Systems (NIPS), 2017.
- Programmatically interpretable reinforcement learning. In International Conference on Machine Learning (ICML), 2018.
- Interpreting cnns via decision trees. In Conference on Computer Vision and Pattern Recognition (CVPR), 2019.