Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

InterpretCC: Intrinsic User-Centric Interpretability through Global Mixture of Experts (2402.02933v3)

Published 5 Feb 2024 in cs.LG, cs.CY, and cs.HC

Abstract: Interpretability for neural networks is a trade-off between three key requirements: 1) faithfulness of the explanation (i.e., how perfectly it explains the prediction), 2) understandability of the explanation by humans, and 3) model performance. Most existing methods compromise one or more of these requirements; e.g., post-hoc approaches provide limited faithfulness, automatically identified feature masks compromise understandability, and intrinsically interpretable methods such as decision trees limit model performance. These shortcomings are unacceptable for sensitive applications such as education and healthcare, which require trustworthy explanations, actionable interpretations, and accurate predictions. In this work, we present InterpretCC (interpretable conditional computation), a family of interpretable-by-design neural networks that guarantee human-centric interpretability, while maintaining comparable performance to state-of-the-art models by adaptively and sparsely activating features before prediction. We extend this idea into an interpretable, global mixture-of-experts (MoE) model that allows humans to specify topics of interest, discretely separates the feature space for each data point into topical subnetworks, and adaptively and sparsely activates these topical subnetworks for prediction. We apply variations of the InterpretCC architecture for text, time series and tabular data across several real-world benchmarks, demonstrating comparable performance with non-interpretable baselines, outperforming interpretable-by-design baselines, and showing higher actionability and usefulness according to a user study.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. GPT-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
  2. Ripple: Concept-based interpretation for raw time series models in education. In Proceedings of the 13th AAAI Symposium on Educational Advances in Artificial Intelligence, 2022. Accepted as a full paper at AAAI 2023: 37th AAAI Conference on Artificial Intelligence (EAAI: AI for Education Special Track), 7-14 of February 2023, Washington DC, USA.
  3. Ripple: Concept-based interpretation for raw time series models in education. In Proceedings of the AAAI Conference on Artificial Intelligence, 2023.
  4. Conditional computation in neural networks for faster models, 2015.
  5. Estimating or propagating gradients through stochastic neurons for conditional computation, 2013.
  6. How to quantify student’s regularity? In Verbert, K., Sharples, M., and Klobučar, T. (eds.), Adaptive and Adaptable Learning, pp.  277–291, Cham, 2016. Springer International Publishing. ISBN 978-3-319-45153-4.
  7. This looks like that: deep learning for interpretable image recognition. Advances in neural information processing systems, 32, 2019.
  8. Utilizing student time series behaviour in learning management systems for early prediction of course performance. Journal of Learning Analytics, 7(2):1–17, 9 2020.
  9. Low-rank approximations for conditional feedforward computation in deep neural networks, 2013.
  10. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  11. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS’18, pp.  590–601, Red Hook, NY, USA, 2018. Curran Associates Inc.
  12. A review of sparse expert models in deep learning. arXiv preprint arXiv:2209.01667, 2022a.
  13. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. The Journal of Machine Learning Research, 23(1):5232–5270, 2022b.
  14. Towards automatic concept-based explanations, 2019.
  15. Measuring student’s utilization of video resources and its effect on academic performance. In 2018 IEEE 18th International Conference on Advanced Learning Technologies (ICALT), pp.  196–198, 2018.
  16. HF Canonical Model Maintainers. distilbert-base-uncased-finetuned-sst-2-english (revision bfdd146), 2022.
  17. Interpretable mixture of experts. Transactions on Machine Learning Research, 2023.
  18. Categorical reparameterization with gumbel-softmax. In International Conference on Learning Representations, 2017.
  19. Mixtral of experts. arXiv preprint arXiv:2401.04088, 2024.
  20. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). ICML 2018, 2017.
  21. The disagreement problem in explainable machine learning: A practitioner’s perspective. arXiv preprint arXiv:2202.01602, 2022.
  22. A data-driven student model to provide adaptive support during video watching across moocs. In Bittencourt, I. I., Cukurova, M., Muldner, K., Luckin, R., and Millán, E. (eds.), Artificial Intelligence in Education, pp.  282–295, Cham, 2020. Springer International Publishing. ISBN 978-3-030-52237-7.
  23. Grade prediction of weekly assignments in moocs: mining video-viewing behavior. Education and Information Technologies, 25(2):1333–1342, 2020.
  24. Base layers: Simplifying training of large, sparse models. In International Conference on Machine Learning, pp.  6265–6274. PMLR, 2021.
  25. Branch-train-merge: Embarrassingly parallel training of expert language models. arXiv preprint arXiv:2208.03306, 2022.
  26. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, pp.  4768–4777, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964.
  27. Can feature predictive power generalize? benchmarking early predictors of student success across flipped and online courses. Proceedings of the 14th International Conference on Educational Data Mining, 2021.
  28. Early prediction of success in mooc from video interaction features. In Bittencourt, I. I., Cukurova, M., Muldner, K., Luckin, R., and Millán, E. (eds.), Artificial Intelligence in Education, pp.  191–196, Cham, 2020. Springer International Publishing. ISBN 978-3-030-52240-7.
  29. Identifying and comparing multi-dimensional student profiles across flipped classrooms. In Artificial Intelligence in Education Proceedings, volume Part I of Lecture Notes in Computer Science. 13355, pp.  90–102. Springer, 2022.
  30. Miller, T. Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence, 267:1–38, 2019.
  31. Predictive learning analytics using deep learning model in moocs’ courses videos. Education and Information Technologies, 26:371–392, 2021.
  32. Multimodal contrastive learning with limoe: the language-image mixture of experts. Advances in Neural Information Processing Systems, 35:9564–9576, 2022.
  33. Pip-net: Patch-based intuitive prototypes for interpretable image classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  2744–2753, 2023.
  34. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084, 2019.
  35. ”why should i trust you?”: Explaining the predictions of any classifier. 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp.  1135–1144, 2016.
  36. Satija, M. P. The theory and practice of the Dewey decimal classification system. Elsevier, 2013.
  37. Concept bottleneck model with additional unsupervised concepts. IEEE Access, 10:41758–41765, 2022.
  38. Dewey decimal classification. Libraries Unlimited, 1998.
  39. Grad-CAM: Visual explanations from deep networks via gradient-based localization. International Journal of Computer Vision, 128(2):336–359, 10 2019.
  40. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 conference on empirical methods in natural language processing, pp.  1631–1642, 2013.
  41. Meta transfer learning for early success prediction in moocs. In Proceedings of the Ninth ACM Conference on Learning@ Scale, pp.  121–132, 2022a.
  42. Meta transfer learning for early success prediction in moocs, 2022b.
  43. Evaluating the explainers: Black-box explainable machine learning for student success prediction in MOOCs. In Mitrovic, A. and Bosch, N. (eds.), Proceedings of the 15th International Conference on Educational Data Mining, pp.  98–109, Durham, United Kingdom, 7 2022c. International Educational Data Mining Society. ISBN 978-1-7336736-3-1.
  44. Trusting the explainers: teacher validation of explainable artificial intelligence for course design. In LAK23: 13th International Learning Analytics and Knowledge Conference, pp.  345–356, 2023a.
  45. The future of human-centric explainable artificial intelligence (xai) is not post-hoc explanations. arXiv preprint arXiv:2307.00364, 2023b.
  46. Dynamic convolutions: Exploiting spatial sparsity for faster inference, 2019.
  47. Pedagogical intervention practices: Improving learning engagement based on early prediction. IEEE Transactions on Learning Technologies, 12(2):278–289, 2019.
  48. Breast Cancer Wisconsin (Diagnostic). UCI Machine Learning Repository, 1995. DOI: https://doi.org/10.24432/C5DW2B.
  49. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems, 32, 2019.
  50. Sum-of-parts models: Faithful attributions for groups of features. arXiv preprint arXiv:2310.16316, 2023.
  51. Character-level convolutional networks for text classification. Advances in neural information processing systems, 28, 2015.
  52. Designing effective sparse expert models. arXiv preprint arXiv:2202.08906, 2, 2022.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com