Towards Event Sequence Foundation Models
Ladda ner
Publicerad
Författare
Typ
Examensarbete för masterexamen
Master's Thesis
Master's Thesis
Modellbyggare
Tidskriftstitel
ISSN
Volymtitel
Utgivare
Sammanfattning
Interruptions in electric power systems cause significant disruptions and economic
losses. The growing availability of fault data from electric power grids, combined
with advances in machine learning, particularly deep learning, is opening up new
opportunities for fault prediction. Transformer architectures are especially well-suited
to model sequential data, like the temporal sequence of historical disturbances in the
power grid. This thesis proposes a foundation model approach to fault prediction,
in which a transformer-based model is pretrained on large-scale event sequence
datasets from diverse domains. We then fine-tune the model on downstream fault
prediction tasks in a continual learning manner, using historical fault data segmented
into 3-day sequences. While an equivalent specialized fault model outperforms this
smaller foundation model by achieving 21.8 % higher average precision, fine-tuning on
multiple proxy tasks - common in foundation model training - significantly improved
fault prediction performance of both models. These results suggest that while
foundation models for event sequences are still emerging, the idea of proxy task
fine-tuning is already beneficial to existing models. Notably, the foundation model
exhibited less forgetting and higher forward transfer than the specialized model,
indicating superior retention of knowledge and ability to adapt to new tasks. Scaling
up foundation models remains a promising path towards reliable and scalable fault
prediction systems for the power grid.
Beskrivning
Ämne/nyckelord
Machine learning, foundation models, temporal point process, transformers, event sequence, transfer learning, power grid fault prediction.