Communication Dans Un Congrès Année : 2024

Using Locally Learnt Word Representations for better Textual Anomaly Detection

Résumé

The literature on general purpose textual Anomaly Detection is quite sparse, as most textual anomaly detection methods are implemented as out of domain detection in the context of pre-established classification tasks. Notably, in a field where pre-trained representations and models are of common use, the impact of the pre-training data on a task that lacks supervision has not been studied. In this paper, we use the simple setting of k-classes out anomaly detection and search for the best pairing of representation and classifier. We show that well-chosen embeddings allow a simple anomaly detection baseline such as OC-SVM to achieve similar results and even outperform deep state-of-the-art models.

Fichier principal
Vignette du fichier
2024.insights-1.11.pdf (342.57 Ko) Télécharger le fichier
Origine Fichiers éditeurs autorisés sur une archive ouverte
licence

Dates et versions

hal-05526854 , version 1 (25-02-2026)

Licence

Identifiants

Citer

Alicia Breidenstein, Matthieu Labeau. Using Locally Learnt Word Representations for better Textual Anomaly Detection. Proceedings of the Fifth Workshop on Insights from Negative Results in NLP, Jun 2026, Mexico City, Mexico. pp.82-91, ⟨10.18653/v1/2024.insights-1.11⟩. ⟨hal-05526854⟩
79 Consultations
14 Téléchargements

Altmetric

Partager

  • More