AfriHuBERT: A self-supervised speech representation model for African languages
Paper
•
2409.20201
•
Published
This is a compact multilingual self-supervised speech encoder based on mHuBERT-147. We performed continued pretraining through multilingual adaptive finetuning (MAFT) on over 10,000 hours of African language data aggregated from various sources. This model can be considered the fourth iteration of mHuBERT-147, specifically trained on African languages. According to the paper, this is the AfriHuBERT-n model. You can click here for the AfriHuBERT-o model.

AfriHuBERT covers 1,230 languages in total including 1,226 indigenous African languages
@misc{alabi2024afrihubertselfsupervisedspeechrepresentation,
title={AfriHuBERT: A self-supervised speech representation model for African languages},
author={Jesujoba O. Alabi and Xuechen Liu and Dietrich Klakow and Junichi Yamagishi},
year={2024},
eprint={2409.20201},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2409.20201},
}