23
1

A General Theory for Kernel Packets: from state space model to compactly supported basis

Liang Ding
Abstract

It is well known that the state space (SS) model formulation of a Gaussian process (GP) can lower its training and prediction time both to O(n) for n data points. We prove that an mm-dimensional SS model formulation of GP is equivalent to a concept we introduce as the general right Kernel Packet (KP): a transformation for the GP covariance function KK such that i=0maiDt(j)K(t,ti)=0\sum_{i=0}^{m}a_iD_t^{(j)}K(t,t_i)=0 holds for any tt1t \leq t_1, 0 jm1\leq j \leq m-1, and m+1m+1 consecutive points tit_i, where Dt(j)f(t){D}_t^{(j)}f(t) denotes jj-th order derivative acting on tt. We extend this idea to the backward SS model formulation of the GP, leading to the concept of the left KP for next mm consecutive points: i=0mbiDt(j)K(t,tm+i)=0\sum_{i=0}^{m}b_i{D}_t^{(j)}K(t,t_{m+i})=0 for any tt2mt\geq t_{2m}. By combining both left and right KPs, we can prove that a suitable linear combination of these covariance functions yields mm compactly supported KP functions: ϕ(j)(t)=0\phi^{(j)}(t)=0 for any t∉(t0,t2m)t\not\in(t_0,t_{2m}) and j=0,,m1j=0,\cdots,m-1. KPs further reduce the prediction time of GP to O(log n) or even O(1), can be applied to more general problems involving the derivative of GPs, and have multi-dimensional generalization for scattered data.

View on arXiv
Comments on this paper