Skip to content

layer6ai-labs/TabDPT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

44 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

TabDPT: Scaling Tabular Foundation Models

Installation

pip install tabdpt

or

git clone [email protected]:layer6ai-labs/TabDPT.git
cd TabDPT
pip install -e .

Example Usage

Please take a look at tests/cls_example.py and tests/reg_example.py For better performance, please increase context_size or increase n_ensembles to trade off speed and accuracy

Updates

Update January 2025

Weights are now stored on Git LFS, at the path checkpoints/tabdpt_76M.ckpt, in addition to Google drive. Please do git lfs pull in order to get the latest weights inside checkpoints folder.

Update December 2024

Added support for flash attention (with bf16 precision) and compile flag. Both are enabled to True by default and should lead to a significant speed-up.

Citation

@article{ma2024tabdpt,
  title={TabDPT: Scaling Tabular Foundation Models},
  author={Ma, Junwei and Thomas, Valentin and Hosseinzadeh, Rasa and Kamkari, Hamidreza and Labach, Alex and Cresswell, Jesse C and Golestan, Keyvan and Yu, Guangwei and Volkovs, Maksims and Caterini, Anthony L},
  journal={arXiv preprint arXiv:2410.18164},
  year={2024}
}

Roadmap

  • Release other model sizes
  • Release training code

About

TabDPT: Scaling Tabular Foundation Models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •  

Languages