Self-Supervised Models for Phoneme Recognition: Applications in Children's Speech for Reading Learning
Child speech recognition is still an underdeveloped area of research due to the lack of data (especially on non-English languages) and the specific difficulties of this task. Having explored various architectures for child speech recognition in previous work, in this article we tackle recent self-supervised models. We first compare wav2vec 2.0, HuBERT and WavLM models adapted to phoneme recognition in French child speech, and continue our experiments with the best of them, WavLM base+. We then further adapt it by unfreezing its transformer blocks during fine-tuning on child speech, which greatly improves its performance and makes it significantly outperform our base model, a Transformer+CTC. Finally, we study in detail the behaviour of these two models under the real conditions of our application, and show that WavLM base+ is more robust to various reading tasks and noise levels. Index Terms: speech recognition, child speech, self-supervised learning
View on arXiv@article{medin2025_2503.04710, title={ Self-Supervised Models for Phoneme Recognition: Applications in Children's Speech for Reading Learning }, author={ Lucas Block Medin and Thomas Pellegrini and Lucile Gelin }, journal={arXiv preprint arXiv:2503.04710}, year={ 2025 } }