Paper ID | MLSP-8.5 | ||
Paper Title | H-GPR: A HYBRID STRATEGY FOR LARGE-SCALE GAUSSIAN PROCESS REGRESSION | ||
Authors | Naiqi Li, Tsinghua-Berkeley Shenzhen Institute, Tsinghua University, China; Yinghua Gao, Wenjie Li, Shenzhen International Graduate School, Tsinghua University, China; Yong Jiang, Tsinghua-Berkeley Shenzhen Institute, Tsinghua University, China; Shu-Tao Xia, Shenzhen International Graduate School, Tsinghua University, China | ||
Session | MLSP-8: Learning | ||
Location | Gather.Town | ||
Session Time: | Tuesday, 08 June, 16:30 - 17:15 | ||
Presentation Time: | Tuesday, 08 June, 16:30 - 17:15 | ||
Presentation | Poster | ||
Topic | Machine Learning for Signal Processing: [MLR-GKM] Graphical and kernel methods | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | With the massive volume of data emerging from both scientific and industrial domains, it has become a desideratum to improve the scalability of Gaussian process regression (GPR). There are two major approaches to assuage its O(n^3) training complexity: the aggregation based methods and the sparse approximation methods. This paper proposes a hybrid strategy called H-GPR to combine these two well-established approaches. We show that it is possible to improve the performance of aggregation based methods by removing some data points that severely violate its underlying assumption, and then this information loss can be recovered by a set of inducing points generated by the sparse approximation methods. A novel metric called conditional independent score is proposed, which measures to what extent the assumption made by the aggregation based methods is satisfied. A heuristic rule is developed to adjust the relative size of the local experts and the inducing subset, so that their distinction can be better reflected. Thorough experiments on synthetic and realistic datasets were performed, demonstrating that the proposed method can improve both the predictive means and variances. |