Abstract Large‐scale Gaussian process (GP) regression is infeasible for large training data due to cubic scaling of flops and quadratic storage involved in working with covariance matrices. Remedies recent literature focus on divide‐and‐conquer, example, partitioning into subproblems inducing functional (and thus computational) independence. Such approximations can be speedy, accurate, sometime...