Review Articles

Partially fixed bayesian additive regression trees

Hao Ran ,

School of Statistics and Management, Shanghai University of Finance and Economics, Shanghai, People's Republic of China

Yang Bai

School of Statistics and Management, Shanghai University of Finance and Economics, Shanghai, People's Republic of China

statbyang@mail.shufe.edu.cn

Pages 232-242 | Received 09 Jan. 2024, Accepted 05 Apr. 2024, Published online: 18 Apr. 2024,
  • Abstract
  • Full Article
  • References
  • Citations

Bayesian Additive Regression Trees (BART) is a widely popular nonparametric regression model known for its accurate prediction capabilities. In certain situations, there is knowledge suggesting the existence of certain dominant variables. However, the BART model fails to fully utilize the knowledge. To tackle this problem, the paper introduces a modification to BART known as the Partially Fixed BART model. By fixing a portion of the trees' structure, this model enables more efficient utilization of prior knowledge, resulting in enhanced estimation accuracy. Moreover, the Partially Fixed BART model can offer more precise estimates and valuable insights for future analysis even when such prior knowledge is absent. Empirical results substantiate the enhancement of the proposed model in comparison to the original BART.

References

  • Breiman, L. (2001). Random forests. Machine Learning45(1), 5–32. https://doi.org/10.1023/A:1010933404324
  • Brooks, T. F., Pope, D. S., & Marcolini, M. A. (1989). Airfoil self-noise and prediction [Tech. Rep]. NASA.
  • Chen, T., & Guestrin, C. (2016). Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 785–794). Association for Computing Machinery.
  • Chipman, H. A., George, E. I., & McCulloch, R. E. (2010). BART: Bayesian additive regression trees. The Annals of Applied Statistics4(1), 266–298. https://doi.org/10.1214/09-AOAS285
  • Dua, D., & Graff, C. (2017). UCI machine learning repository. https://archive.ics.uci.edu/ml
  • Lin, W., Shi, Z., Wang, Y., & Yan, T. H. (2023). Unfolding Beijing in a hedonic way. Computational Economics61(1), 1–24. https://doi.org/10.1007/s10614-021-10209-3
  • Linero, A. R. (2018). Bayesian regression trees for high-dimensional prediction and variable selection. Journal of the American Statistical Association113(522), 626–636. https://doi.org/10.1080/01621459.2016.1264957
  • Linero, A. R., & Yang, Y. (2018). Bayesian regression tree ensembles that adapt to smoothness and sparsity. Journal of the Royal Statistical Society Series B: Statistical Methodology80(5), 1087–1110. https://doi.org/10.1111/rssb.12293
  • Tan, Y. V., & Roy, J. (2019). Bayesian additive regression trees and the general BART model. Statistics in Medicine38(25), 5048–5069. https://doi.org/10.1002/sim.v38.25

To cite this article: Hao Ran & Yang Bai (18 Apr 2024): Partially fixed bayesian additive regression trees, Statistical Theory and Related Fields, DOI: 10.1080/24754269.2024.2341981

To link to this article: https://doi.org/10.1080/24754269.2024.2341981