The Locality and Symmetry of Positional Encodings - Equipe Data, Intelligence and Graphs Accéder directement au contenu
Communication Dans Un Congrès Année : 2023

The Locality and Symmetry of Positional Encodings

Résumé

Positional Encodings (PEs) are used to inject word-order information into transformer-based language models. While they can significantly enhance the quality of sentence representations, their specific contribution to language models is not fully understood, especially given recent findings that various positional encodings are insensitive to word order. In this work, we conduct a systematic study of positional encodings in Bidirectional Masked Language Models (BERT-style) , which complements existing work in three aspects: (1) We uncover the core function of PEs by identifying two common properties, Locality and Symmetry; (2) We show that the two properties are closely correlated with the performances of downstream tasks; (3) We quantify the weakness of current PEs by introducing two new probing tasks, on which current PEs perform poorly. We believe that these results are the basis for developing better PEs for transformer-based language models. The code is available at https://github.
Fichier principal
Vignette du fichier
The_Locality_and_Symmetry_of_Positional_Encodings.pdf (2.92 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-04330367 , version 1 (08-12-2023)

Identifiants

  • HAL Id : hal-04330367 , version 1

Citer

Lihu Chen, Gaël Varoquaux, Fabian M. Suchanek. The Locality and Symmetry of Positional Encodings. EMNLP 2023 - Conference on Empirical Methods in Natural Language Processing, Dec 2023, Singapore, Singapore. ⟨hal-04330367⟩
85 Consultations
19 Téléchargements

Partager

Gmail Facebook X LinkedIn More