Griffin: Towards a Graph-Centric Relational Database Foundation Model

We introduce Griffin, the first foundation model attemptation designed specifically for Relational Databases (RDBs). Unlike previous smaller models focused on single RDB tasks, Griffin unifies the data encoder and task decoder to handle diverse tasks. Additionally, we enhance the architecture by incorporating a cross-attention module and a novel aggregator. Griffin utilizes pretraining on both single-table and RDB datasets, employing advanced encoders for categorical, numerical, and metadata features, along with innovative components such as cross-attention modules and enhanced message-passing neural networks (MPNNs) to capture the complexities of relational data. Evaluated on large-scale, heterogeneous, and temporal graphs extracted from RDBs across various domains (spanning over 150 million nodes), Griffin demonstrates superior or comparable performance to individually trained models, excels in low-data scenarios, and shows strong transferability with similarity and diversity in pretraining across new datasets and tasks, highlighting its potential as a universally applicable foundation model for RDBs. Code available atthis https URL.
View on arXiv@article{wang2025_2505.05568, title={ Griffin: Towards a Graph-Centric Relational Database Foundation Model }, author={ Yanbo Wang and Xiyuan Wang and Quan Gan and Minjie Wang and Qibin Yang and David Wipf and Muhan Zhang }, journal={arXiv preprint arXiv:2505.05568}, year={ 2025 } }