MAP- and MLE-Based Teaching
Abstract: Imagine a learner L who tries to infer a hidden concept from a collection of observations. Building on the work [4] of Ferri et al., we assume the learner to be parameterized by priors P(c) and by c-conditional likelihoods P(z|c) where c ranges over all concepts in a given class C and z ranges over all observations in an observation set Z. L is called a MAP-learner (resp. an MLE-learner) if it thinks of a collection S of observations as a random sample and returns the concept with the maximum a-posteriori probability (resp. the concept which maximizes the c-conditional likelihood of S). Depending on whether L assumes that S is obtained from ordered or unordered sampling resp. from sampling with or without replacement, we can distinguish four different sampling modes. Given a target concept c in C, a teacher for a MAP-learner L aims at finding a smallest collection of observations that causes L to return c. This approach leads in a natural manner to various notions of a MAP- or MLE-teaching dimension of a concept class C. Our main results are: We show that this teaching model has some desirable monotonicity properties. We clarify how the four sampling modes are related to each other. As for the (important!) special case, where concepts are subsets of a domain and observations are 0,1-labeled examples, we obtain some additional results. First of all, we characterize the MAP- and MLE-teaching dimension associated with an optimally parameterized MAP-learner graph-theoretically. From this central result, some other ones are easy to derive. It is shown, for instance, that the MLE-teaching dimension is either equal to the MAP-teaching dimension or exceeds the latter by 1. It is shown furthermore that these dimensions can be bounded from above by the so-called antichain number, the VC-dimension and related combinatorial parameters. Moreover they can be computed in polynomial time.
- Frank Balbach. Measuring teachability using variants of the teaching dimension. Theoretical Computer Science, 397(1–3):94–113, 2008.
- Baxter S. Eaves, Jr. and Patrick Shafto. Toward a general, scaleable framework for Bayesian teaching with applications to topic models. CoRR, 2016.
- On batch teaching without collusion. Journal of Machine Learning Research, 23:1–32, 2022.
- Non-cheating teaching revisited: A new probabilistic machine teaching model. In Luc De Raedt, editor, Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, pages 2973–2979, 2022.
- Preference-based teaching. Journal of Machine Learning Research, 18(31):1–32, 2017.
- On the complexity of teaching. Journal of Computer and System Sciences, 50(1):20–31, 1995.
- Teaching a smarter learner. Journal of Computer and System Sciences, 52(2):255–267, 1996.
- XAI with machine teaching when humans are (not) informed about the irrelevant features. In ECML, 2023. To appear.
- Optimal collusion-free teaching. In Aurélien Garivier and Satyen Kale, editors, Proceedings of Machine Learning Research (ALT 2019), volume 98, pages 1–23, 2019.
- On batch teaching with sample complexity bounded by vcd. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors, Advances in Neural Information Processing Systems, volume 35, pages 15732–15742. Curran Associates, Inc., 2022.
- Stephen H. Muggleton. Learning from positive data. In Stephen H. Muggleton, editor, Inductive Logic Programming, 6th International Workshop, ILP 1996, volume 1314 of Lecture Notes in Computer Science, pages 358–376. Springer, 1996.
- A rational account of pedagogical reasoning: Teaching by, and learning from, examples. Cognitive Psychology, 71:55 – 89, 2014.
- Teachability in computational learning. New Generation Computing, 8(4):337–348, 1991.
- On the uniform convergence of relative frequencies of events to their probabilities. Theory of Probability & its Applications, 16(2):264–280, 1971.
- Scott Cheng-Hsin Yang and Patrick Shafto. Explainable artificial intelligence via bayesian teaching. In NIPS 2017 workshop on Teaching Machines, Robots, and Humans, pages 127–137, 2017.
- Jerry Zhu. Machine teaching for bayesian learners in the exponential family. In C.J. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 26. Curran Associates, Inc., 2013.
- An overview of machine teaching. CoRR, 2018.
- Models of cooperative teaching and learning. Journal of Machine Learning Research, 12:349–384, 2011.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.