Orthogonal Bootstrap: Efficient Simulation of Input Uncertainty

Kaizhao Liu*, Jose Blanchet, Lexing Ying, Yiping Lu*

*Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

Abstract

Bootstrap is a popular methodology for simulating input uncertainty. However, it can be computationally expensive when the number of samples is large. We propose a new approach called Orthogonal Bootstrap that reduces the number of required Monte Carlo replications. We decomposes the target being simulated into two parts: the non-orthogonal part which has a closed-form result known as Infinitesimal Jackknife and the orthogonal part which is easier to be simulated. We theoretically and numerically show that Orthogonal Bootstrap significantly reduces the computational cost of Bootstrap while improving empirical accuracy and maintaining the same width of the constructed interval.

Original languageEnglish (US)
Pages (from-to)30669-30701
Number of pages33
JournalProceedings of Machine Learning Research
Volume235
StatePublished - 2024
Event41st International Conference on Machine Learning, ICML 2024 - Vienna, Austria
Duration: Jul 21 2024Jul 27 2024

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Orthogonal Bootstrap: Efficient Simulation of Input Uncertainty'. Together they form a unique fingerprint.

Cite this