ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.07024
14
0

Data Augmentation and Hyperparameter Tuning for Low-Resource MFA

9 April 2025
Alessio Tosolini
Claire Bowern
ArXivPDFHTML
Abstract

A continued issue for those working with computational tools and endangered and under-resourced languages is the lower accuracy of results for languages with smaller amounts of data. We attempt to ameliorate this issue by using data augmentation methods to increase corpus size, comparing augmentation to hyperparameter tuning for multilingual forced alignment. Unlike text augmentation methods, audio augmentation does not lead to substantially increased performance. Hyperparameter tuning, on the other hand, results in substantial improvement without (for this amount of data) infeasible additional training time. For languages with small to medium amounts of training data, this is a workable alternative to adapting models from high-resource languages.

View on arXiv
@article{tosolini2025_2504.07024,
  title={ Data Augmentation and Hyperparameter Tuning for Low-Resource MFA },
  author={ Alessio Tosolini and Claire Bowern },
  journal={arXiv preprint arXiv:2504.07024},
  year={ 2025 }
}
Comments on this paper