Automatic Speech Recognition for Kikuyu

License

Model Description 🍍

This model is a fine-tuned version of Wav2Vec2-BERT 2.0 for Kikuyu automatic speech recognition (ASR). It was trained on the 100+ hours of transcribed speech, covering Health, Government, Finance, Education, and Agriculture domains. The in-domain WER for this ASR model is below 25.0%.

  • Developed by: Badr al-Absi
  • Model type: Speech Recognition (ASR)
  • Language: Kikuyu (kik)
  • License: CC-BY-4.0
  • Finetuned from: facebook/w2v-bert-2.0

Model Architecture

  • Base model: Wav2Vec2-BERT 2.0
  • Architecture: transformer-based with convolutional feature extractor
  • Parameters: ~600M (inherited from base model)
  • Objective: connectionist temporal classification (CTC)

Funding

The development of this model was supported by CLEAR Global and Gates Foundation.

Citation

@misc{w2v_bert_kikuyu_asr,
  author = {Badr M. Abdullah},
  title = {Adapting Wav2Vec2-BERT 2.0 for Kikuyu ASR},
  year = {2025},
  publisher = {Hugging Face},
  url = {https://huggingface.co/badrex/w2v-bert-2.0-swahili-asr}
}

Model Card Contact

For questions or issues, please contact via the Hugging Face model repository in the community discussion section.

Downloads last month
17
Safetensors
Model size
0.6B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for badrex/w2v-bert-2.0-kikuyu-asr

Finetuned
(368)
this model

Collection including badrex/w2v-bert-2.0-kikuyu-asr