Papers
Topics
Authors
Recent
Search
2000 character limit reached

Conformalized Semi-supervised Random Forest for Classification and Abnormality Detection

Published 4 Feb 2023 in cs.LG | (2302.02237v2)

Abstract: The Random Forests classifier, a widely utilized off-the-shelf classification tool, assumes training and test samples come from the same distribution as other standard classifiers. However, in safety-critical scenarios like medical diagnosis and network attack detection, discrepancies between the training and test sets, including the potential presence of novel outlier samples not appearing during training, can pose significant challenges. To address this problem, we introduce the Conformalized Semi-Supervised Random Forest (CSForest), which couples the conformalization technique Jackknife+aB with semi-supervised tree ensembles to construct a set-valued prediction $C(x)$. Instead of optimizing over the training distribution, CSForest employs unlabeled test samples to enhance accuracy and flag unseen outliers by generating an empty set. Theoretically, we establish CSForest to cover true labels for previously observed inlier classes under arbitrarily label-shift in the test data. We compare CSForest with state-of-the-art methods using synthetic examples and various real-world datasets, under different types of distribution changes in the test domain. Our results highlight CSForest's effective prediction of inliers and its ability to detect outlier samples unique to the test data. In addition, CSForest shows persistently good performance as the sizes of the training and test sets vary. Codes of CSForest are available at https://github.com/yujinhan98/CSForest.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (37)
  1. Predictive inference with the jackknife+. The Annals of Statistics, 49(1):486–507.
  2. Classification with a reject option using a hinge loss. Journal of Machine Learning Research, 9(8).
  3. Discriminative learning under covariate shift. Journal of Machine Learning Research, 10(9).
  4. Cadre, B. (2006). Kernel estimation of density level sets. Journal of multivariate analysis, 97(4):999–1023.
  5. Classification with rejection based on cost-sensitive classification. In International Conference on Machine Learning, pages 1507–1517. PMLR.
  6. Chow, C. (1970). On optimum recognition error and reject tradeoff. IEEE Transactions on information theory, 16(1):41–46.
  7. Boosting with abstention. Advances in Neural Information Processing Systems, 29.
  8. Csurka, G. (2017). Domain adaptation for visual applications: A comprehensive survey. arXiv preprint arXiv:1702.05374.
  9. Dermatologist-level classification of skin cancer with deep neural networks. nature, 542(7639):115–118.
  10. Learning by transduction. arXiv preprint arXiv:1301.7375.
  11. Covariate shift by kernel mean matching. Dataset shift in machine learning, 3(4):5.
  12. Prediction and outlier detection in classification problems. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 84(2):524–546.
  13. Cautious deep learning. arXiv preprint arXiv:1805.09460.
  14. Classification with reject option. The Canadian Journal of Statistics/La Revue Canadienne de Statistique, pages 709–721.
  15. Driving to safety: How many miles of driving would it take to demonstrate autonomous vehicle reliability? Transportation Research Part A: Policy and Practice, 94:182–193.
  16. Predictive inference is free with the jackknife+-after-bootstrap. Advances in Neural Information Processing Systems, 33:4138–4149.
  17. Second opinion needed: communicating uncertainty in medical machine learning. NPJ Digital Medicine, 4(1):1–6.
  18. Learning multiple layers of features from tiny images.
  19. MNIST handwritten digit database.
  20. Lei, J. (2014). Classification with confidence. Biometrika, 101(4):755–769.
  21. Distribution-free prediction bands for nonparametric regression. Quality control and applied statistics, 60(1):109–110.
  22. Approximations to magic: Finding unusual medical time series. In 18th IEEE Symposium on Computer-Based Medical Systems (CBMS’05), pages 329–334. IEEE.
  23. Detecting and correcting for label shift with black box predictors. In International conference on machine learning, pages 3122–3130. PMLR.
  24. Computer intrusion detection and network monitoring: a statistical viewpoint. Springer.
  25. Ix. on the problem of the most efficient tests of statistical hypotheses. Philosophical Transactions of the Royal Society of London. Series A, Containing Papers of a Mathematical or Physical Character, 231(694-706):289–337.
  26. On the calibration of multiclass classification with rejection. Advances in Neural Information Processing Systems, 32.
  27. Inductive confidence machines for regression. In Machine Learning: ECML 2002: 13th European Conference on Machine Learning Helsinki, Finland, August 19–23, 2002 Proceedings 13, pages 345–356. Springer.
  28. Securing connected & autonomous vehicles: Challenges posed by adversarial machine learning and the way forward. IEEE Communications Surveys & Tutorials, 22(2):998–1026.
  29. Classification with valid and adaptive coverage. Advances in Neural Information Processing Systems, 33:3581–3591.
  30. Least ambiguous set-valued classifiers with bounded error levels. Journal of the American Statistical Association, 114(525):223–234.
  31. On causal and anticausal learning. arXiv preprint arXiv:1206.6471.
  32. Shimodaira, H. (2000). Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of statistical planning and inference, 90(2):227–244.
  33. Storkey, A. (2009). When training and test sets are different: characterizing learning transfer. Dataset shift in machine learning, 30:3–28.
  34. Conformal prediction under covariate shift. Advances in neural information processing systems, 32.
  35. Algorithmic learning in a random world. Springer Science & Business Media.
  36. Cross-conformal predictive distributions. In Conformal and Probabilistic Prediction and Applications, pages 37–51. PMLR.
  37. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747.
Citations (2)

Summary

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.