This document outlines the research questions and methodological guidance for a proposed Delayed Acceptance (DA) and Gaussian Process (GP) based algorithm for active learning and surrogate model construction.
1. What are the theoretical foundations for using Delayed Acceptance in the context of GP-based surrogate modeling?
- Use DA-MCMC literature (e.g., Christen & Fox, 2005) as the basis.
- Frame DA as a two-stage acceptance scheme: use GP to pre-filter proposals before evaluating the true model.
- Metrics:
- Predictive variance (mean or max).
- Leave-one-out cross-validation error.
- Stabilization of marginal likelihood.
- Thresholds on these metrics can trigger trust in the GP.
3. What criteria should be used to trigger the switch from evaluating the forward model to relying solely on the GP?
- Example criteria:
- Predictive variance below a threshold across recent samples.
- No significant change in GP posterior over several updates.
- Stagnant improvement in log marginal likelihood.
- When a new point is added: [ K_{n+1} = \begin{bmatrix} K_n & k \ k^T & k_{new} \end{bmatrix} ] Update ( L_{n+1} ) using Schur complement logic to avoid full recomputation.
- Full retraining: ( \mathcal{O}(n^3) )
- Online update: ( \mathcal{O}(n^2) )
- Enables adaptive, low-latency learning.
6. Can the DA-GP framework maintain posterior accuracy compared to full MCMC or MCMC-GP hybrid approaches?
- Use metrics such as:
- KL divergence.
- Wasserstein distance.
- Posterior moment comparison.
- Empirically validate using controlled test functions.
- Benchmark on Branin, Hartmann, Rosenbrock, etc.
- Track:
- Posterior accuracy.
- Runtime.
- GP prediction error (RMSE).
8. How does sample efficiency, acceptance rate, and computational cost compare with traditional MCMC-guided GP approaches?
- Evaluate:
- Number of model evaluations.
- Acceptance rate pre/post DA.
- Time per iteration and total convergence time.
- Increase input dimension from 2D to 10D+.
- Add Gaussian noise to the output.
- Measure effect on GP stability, DA rejection rate, and accuracy.
- Scenarios:
- High-dimensional or sparse data.
- Poor GP kernel tuning.
- Non-smooth/multimodal likelihoods.
- Replace GP with:
- Bayesian neural networks.
- Ensemble trees with uncertainty quantification.
- Maintain DA logic: cheap filter before costly evaluation.
- Use online GP updates with:
- Sliding window or forgetting factors.
- Real-time adaptation to data drift.
- Draft pseudocode of the DA-GP algorithm.
- Design experimental setup with benchmarks and metrics.
- Write a literature review comparing MCMC-GP, DA-GP, and pure GP methods.