Use FT-Transformer architecture to learn dataset-specific embeddings and a shared backbone in a federated setting. Pre-training a transformer-based model shows superior performance over not. However, the proposed method is still beaten by CatBoost.
Approach
Overview of the XTab framework
Main Idea
Train a shared backbone (transformer stack) that can handle arbitrary embedding representation of tables.