15
4

Fast Online Node Labeling for Very Large Graphs

Abstract

This paper studies the online node classification problem under a transductive learning setting. Current methods either invert a graph kernel matrix with O(n3)\mathcal{O}(n^3) runtime and O(n2)\mathcal{O}(n^2) space complexity or sample a large volume of random spanning trees, thus are difficult to scale to large graphs. In this work, we propose an improvement based on the \textit{online relaxation} technique introduced by a series of works (Rakhlin et al.,2012; Rakhlin and Sridharan, 2015; 2017). We first prove an effective regret O(n1+γ)\mathcal{O}(\sqrt{n^{1+\gamma}}) when suitable parameterized graph kernels are chosen, then propose an approximate algorithm FastONL enjoying O(kn1+γ)\mathcal{O}(k\sqrt{n^{1+\gamma}}) regret based on this relaxation. The key of FastONL is a \textit{generalized local push} method that effectively approximates inverse matrix columns and applies to a series of popular kernels. Furthermore, the per-prediction cost is O(vol(S)log1/ϵ)\mathcal{O}(\text{vol}({\mathcal{S}})\log 1/\epsilon) locally dependent on the graph with linear memory cost. Experiments show that our scalable method enjoys a better tradeoff between local and global consistency.

View on arXiv
Comments on this paper