This paper explores the correlations between attention layers of Audio Speech Recognition (ASR) and Natural Language Processing (NLP) models, with the aim of improving ASR performance and gaining insights into the existence of hidden language models in ASR. We build upon previous research on attention mechanisms and use BERT for NLP and Wav2vec2 for ASR. We define a metric for correlation, compare attention layers, address alignment challenges, and investigate the relationship across the English language. Our user-friendly interface visually presents aligned attention matrices, aiding in model analysis and investigating those correlations. The paper concludes with potential linguistic and semantic realtions between those models, and with a method and metrics that can be applied for feature research.
-
Notifications
You must be signed in to change notification settings - Fork 0
Final Project in Advanced Natural Language Processing 67664@ HUJI
License
eliyahabba/SoundOfAttention
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
Final Project in Advanced Natural Language Processing 67664@ HUJI
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published