BiTimelyGPT is a bidirectional generative pre-training Transformer designed for representation learning using healthcare time-series data, including continuously monitored biosignals and irregularly sampled time series from longitudinal clinical records.
This figure indicates an overview of BiTimelyGPT architecture.
Panel a. BiTimelyGPT architecture is stacked with
Panel b. Bidirectional Alternating AutoRegressive Modeling (BAAR) framework alternately models left-to-right and right-to-left information across layers, thereby pre-training deep bidirectional contextualized representations.
It pre-trains by a Next-Previous-Token Prediction strategy on the final two layers and fine-tunes the [SOS] token in the final layer for discriminative tasks.
Panel c. Bidirectional Retention Layer alternates between forward and backward Retention across layers.
This figure depicts the process of BiTimelyGPT's Next-Previous-Token Prediction pre-training.
The input of time-series sequence with
This published code is referenced from:
Ziyang Song, Qincheng Lu, Mike He Zhu, David Buckeridge, and Yue Li. (2024). Bidirectional generative pre-training for improving healthcare time-series representation learning. Machine Learning for HealthCare (MLHC), Proceedings of Machine Learning Research (JMLR Proceedings track).