Detecting out-of-distribution text using topological features of transformer-based language models
Abstract
To safeguard machine learning systems that operate on textual data against out-of-distribution (OOD) inputs that could cause unpredictable behaviour, we explore the use of topological features of self-attention maps from transformer-based language models to detect when input text is out of distribution. Self-attention forms the core of transformer-based language models, dynamically assigning vectors to words based on context, thus in theory our methodology is applicable to any transformer-based language model with multihead self-attention. We evaluate our approach on BERT and compare it to a traditional OOD approach using CLS embeddings. Our results show that our approach outperforms CLS embeddings in distinguishing in-distribution samples from far-out-of-domain samples, but struggles with near or same-domain datasets.
- Publication:
-
arXiv e-prints
- Pub Date:
- November 2023
- DOI:
- arXiv:
- arXiv:2311.13102
- Bibcode:
- 2023arXiv231113102P
- Keywords:
-
- Computer Science - Computation and Language;
- Computer Science - Machine Learning;
- Mathematics - Algebraic Topology
- E-Print:
- 8 pages, 6 figures, 3 tables, to be published in proceedings of the IJCAI-2024 AISafety Workshop