Revisiting Deep Learning Models for Tabular Data
1 minute read ∼ Filed in : A paper notelayout: post
Introduction
Contributions
- Evaluate the main models.
- It shows that the ResNet-like model is effective
- Introduce FT-Transformer, a simple adaptation of transformer for tabular data
- No universally superior solution among GBDT and deep models.
Existing work
- Tree-based models
- XGBoost, LightGBM, CatBoost.
- Deep learning models
- Differentiable trees (traditional tree is not differentiable).
- Attention-based models ().
- Explicit modeling of multiplicative interactions between features.
- MLP is unsuitable for modeling the multiplicative interactions between features.
FT-Transformer
Feature Tokenizer
It transforms the input features x to embedding T.
- Numerical Features: multiple by a W in element-wise.
- Categorical Features: Lookup Table.
Transformer
asdf