Towards Cross-Lingual Audio Abuse Detection in Low-Resource Settings with Few-Shot Learning

Published in Proceedings of the 31st International Conference on Computational Linguistics (COLING 2025), 2025

Online abusive content detection, particularly in low-resource settings and within the audio modality, remains underexplored. We investigate the potential of pre-trained audio representations for detecting abusive language in low-resource languages, in this case, in Indian languages using Few Shot Learning (FSL). Leveraging powerful representations from models such as Wav2Vec and Whisper, we explore cross-lingual abuse detection using the ADIMA dataset with FSL. Our approach integrates these representations within the Model-Agnostic Meta-Learning (MAML) framework to classify abusive language in 10 languages. We experiment with various shot sizes (50-200) evaluating the impact of limited data on performance. Additionally, a feature visualization study was conducted to better understand model behaviour. This study highlights the generalization ability of pre-trained models in low-resource scenarios and offers valuable insights into detecting abusive language in multilingual contexts.

Recommended citation: Aditya Narayan Sankaran, Reza Farahbakhsh, and Noel Crespi. 2025. Towards Cross-Lingual Audio Abuse Detection in Low-Resource Settings with Few-Shot Learning. In Proceedings of the 31st International Conference on Computational Linguistics, pages 5558–5569, Abu Dhabi, UAE. Association for Computational Linguistics.
Download Paper | Download Slides